<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[AI Policy Perspectives : AI Policy Primer]]></title><description><![CDATA[A monthly run-down of 3 papers policymakers should read ]]></description><link>https://www.aipolicyperspectives.com/s/ai-policy-primer</link><image><url>https://substackcdn.com/image/fetch/$s_!XGVU!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa24053ba-9bcb-4c21-a969-fe02656ce349_585x585.png</url><title>AI Policy Perspectives : AI Policy Primer</title><link>https://www.aipolicyperspectives.com/s/ai-policy-primer</link></image><generator>Substack</generator><lastBuildDate>Tue, 05 May 2026 19:29:19 GMT</lastBuildDate><atom:link href="https://www.aipolicyperspectives.com/feed" rel="self" type="application/rss+xml"/><language><![CDATA[en]]></language><webMaster><![CDATA[aipolicyperspectives@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[aipolicyperspectives@substack.com]]></itunes:email><itunes:name><![CDATA[AI Policy Perspectives]]></itunes:name></itunes:owner><itunes:author><![CDATA[AI Policy Perspectives]]></itunes:author><googleplay:owner><![CDATA[aipolicyperspectives@substack.com]]></googleplay:owner><googleplay:email><![CDATA[aipolicyperspectives@substack.com]]></googleplay:email><googleplay:author><![CDATA[AI Policy Perspectives]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[AI Policy Primer (#24)]]></title><description><![CDATA[Identifying agents, self-improvement, and artificial clouds]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-24</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-24</guid><dc:creator><![CDATA[Conor Griffin]]></dc:creator><pubDate>Thu, 09 Apr 2026 14:50:28 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!CLkj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Every six weeks, we round up three papers that we think AI policy folks should be reading. In this edition, we look at a <a href="https://arxiv.org/abs/2603.10028">proposal</a> for how to identify the agents that will soon fill the economy; <a href="https://cset.georgetown.edu/publication/when-ai-builds-ai/">research</a> on the prospect of self-improving AI; and<a href="https://arxiv.org/pdf/2603.06909"> new insights</a> about how to use AI to prevent contrails, or artificial clouds, from warming the planet. </em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CLkj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CLkj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png 424w, https://substackcdn.com/image/fetch/$s_!CLkj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png 848w, https://substackcdn.com/image/fetch/$s_!CLkj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png 1272w, https://substackcdn.com/image/fetch/$s_!CLkj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CLkj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2365898,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.aipolicyperspectives.com/i/193691288?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CLkj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png 424w, https://substackcdn.com/image/fetch/$s_!CLkj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png 848w, https://substackcdn.com/image/fetch/$s_!CLkj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png 1272w, https://substackcdn.com/image/fetch/$s_!CLkj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879e6456-0c9d-4813-a7b9-1fdc297b6a23_8000x4500.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>1. Identifying (and incentivising) AI agents</h2><ul><li><p><strong>What happened: </strong>A trio of law and philosophy professors considered how to identify who (or what) is responsible for AI agents&#8217; actions in the world, and came up with a two-part <a href="https://arxiv.org/abs/2603.10028">proposal</a>: that the disparate and evolving agents within a system should exist legally as a new form of corporation; and that each corporation should link to accountable humans.</p></li><li><p><strong>What&#8217;s interesting: </strong>The paper by <a href="https://law.ua.edu/faculty_staff/yonathan-arbel/">Yonathan Arbel,</a> <a href="https://law.ua.edu/faculty_staff/yonathan-arbel/">Simon Goldstein</a>, and <a href="https://www.law.uh.edu/faculty/main.asp?PID=6428">Peter N. Salib</a> starts with a thought experiment. It&#8217;s 2030, and your AI assistant suggests that it optimizes your slow WiFi connection. After you agree, it spawns a swarm of agents. Some are copies, while others are cheaper agents running on open-source models. Some start to interface with AI agents from other companies. Three months later, two FBI agents knock on your door and explain that your network has been piggybacking on a local defense contractor&#8217;s WiFi network.</p></li><li><p>Before determining who is responsible and what the repercussions should be, there are more basic questions: Who are the AI actors in this story? How many are there?</p></li><li><p><a href="https://www.aipolicyperspectives.com/p/an-agents-economy">The economy will soon be filled with capable AI agents</a>. To deter and respond to such harms, the authors argue that we need to be able to identify these agents, at two levels.</p><ul><li><p>To prevent human misuse or negligence, we need &#8216;<strong>thin identity&#8217;</strong>. This would connect AI agents to the humans most able to control them, similar to how &#8216;know-your-customer&#8217; rules tie banking transactions to humans.</p></li><li><p>Humans will be unable to monitor and control every AI decision, so we also need to be able to identify agents themselves, hold them accountable and incentivize them to behave well. To do so, we need &#8216;<strong>thick identity&#8217; </strong>that can distinguish AI agents as stable, coherent entities, with persistent goals. This goal is pragmatic and does not require viewing AIs as conscious in any sense.</p></li></ul></li><li><p><em>Thickly </em>identifying agents is harder and more novel, as AI agents need not be attached to a physical body. Multiple agents can also work together on a single task. Any single agent can be copied, spun up, spun down, or be continually updated.</p></li><li><p>To address such challenges, the authors propose creating algorithmic corporations, or &#8216;A-corps&#8217;. These would have two key elements:</p><ul><li><p><strong>Legal personhood: </strong>Like a traditional corporation, an A-corp would be a single legal entity that persists over time. It could hold property, make contracts, and be sued. But it would be run by a collection of AI agents. As such, the proposal runs contrary to scholars who have argued <a href="https://arxiv.org/pdf/2502.18359">against</a> granting legal personhood to AI agents, or called for <a href="https://openscholarship.wustl.edu/law_lawreview/vol95/iss4/7/#:~:text=This%20Article%20argues%20that%20algorithmic,which%20have%20non%2Dhuman%20controllers.">bans</a> on algorithms running companies because of concerns about crime and companies using them to avoid liability.</p></li><li><p><strong>Computationally-secure governance: </strong>Each A-corp would have a unique digital certificate and a secure private key to authorise transactions. The humans that own each A-corp could grant the key to an AI &#8216;manager&#8217; agent who in turn could grant more limited permissions to sub-agents within the A-corp, or to other A-corps, such as permissions to spend up to $100 or to read a batch of emails.</p></li></ul></li><li><p>The proposal addresses thin identity by reducing the vast number of AI agents down to a smaller number of A-corps, whose actions are traceable back to their human owners. As with limited liability companies (LLCs), the human owners would not be responsible for <em>all </em>harm their A-corps cause, but could lose all funds they invest and possibly face further liability, for example in cases of fraud or negligence.</p></li><li><p>The proposal addresses thick identity via its &#8216;resource constraint thesis&#8217;. All AI agents need resources, like money and compute. A-corps provide AIs with a way to access these resources and an incentive to manage them well. For example, A-corps that tightly monitor and audit their sub-agents&#8217; performance would get more resources, while A-corps that allow fraud or waste will lose resources. This encourages A-corps to self-organise<em>, </em>into stable, coherent, multi-agent systems.</p></li><li><p>The authors argue that A-corps could also address alignment concerns, for example by reducing the incentive for an AI agent to exfiltrate its own weights, because that new AI instance would lose access to resources and permissions from the A-corp.</p></li><li><p>To make it happen, the authors call for a public registry of A-corps. This would list each A-corp&#8217;s human owners, the certificates to authenticate it against, as well as (potentially) the differing permissions enjoyed by its agents. Ultimately, the authors argue that A-corps should become mandatory for any AI agent taking &#8220;economically significant actions&#8221;, and to guard against criminals using AI agents anonymously.</p></li><li><p>The authors respond to some expected pushback. They do not see A-corps as anthropomorphising AI because the proposal does not require anybody to view agents as having deeper desires or wants. They also think A-corps can prevent the risk that AI agents might slowly build up resources before deploying them for harm, by encouraging inter-agent trade that penalises rogue behavior. Could A-corps disempower humans? The authors argue that they provide a pathway to tax and redistribution, and enable humans to better steer agents, for example by designating the parts of the economy that A-corps are permitted to operate in.</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for free. Lots more in the pipeline. </p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>2. When AI builds AI</h2><ul><li><p><strong>What happened: </strong>The Centre for Security and Emerging Technology, CSET, released <a href="https://cset.georgetown.edu/publication/when-ai-builds-ai/">a report</a> on the prospects for AI improving itself, known as automated R&amp;D or recursive self-improvement, based on an expert workshop in July 2025.</p></li><li><p><strong>What&#8217;s interesting: </strong>In 1964, the computer scientist I.J. Good wrote about the possibility of an &#8220;intelligence explosion&#8221; that would leave &#8220;the intelligence of man.&#8230;far behind&#8221;. Researchers have also long automated aspects of writing code and AI model design.</p></li><li><p>However, the speed of AI coding advances suggests that something qualitatively different may soon occur. This makes two questions salient: 1. Could AI automate the <em>entire </em>AI R&amp;D process? 2. Will this R&amp;D automation extend across all scientific disciplines? The CSET report focuses on the first question.</p></li><li><p>CSET defines AI R&amp;D by distinguishing between <em>research scientists, </em>who generate hypotheses, design experiments and interpret results; and <em>research engineers, </em>who write code, fix bugs and generate data. They also note the inputs that AI R&amp;D relies on, such as raising funds and acquiring compute.</p></li><li><p>They sketch out four overlapping scenarios for how AI R&amp;D may play out:</p><ul><li><p><strong>1. Explosion: </strong>AI systems automate a growing share of AI R&amp;D. Initially, this leads to modest productivity gains, but as the length and complexity of tasks that AI performs grows, productivity soars. AI systems become far more capable than humans, whose involvement in AI R&amp;D falls to zero.</p></li><li><p><strong>2. Fizzle: </strong>The share of R&amp;D tasks done by AI rises, but rather than leading to compounding improvements, capabilities start to plateau.</p></li><li><p><strong>3. Amdahl&#8217;s Law: </strong>AI automates certain activities, like writing code and running experiments, but not others, like research strategy.</p></li><li><p><strong>4. The expanding pie: </strong>As AI automation grows, humans realise that new ideas and breakthroughs are needed that AI systems cannot yet provide.</p></li></ul></li><li><p>The experts in CSET&#8217;s workshop held widely diverging views on which scenario was most likely. Most importantly, new empirical data is unlikely to resolve these conflicts, because participants may view the same data as confirming their own assumptions.</p><ul><li><p>For example, an AI system&#8217;s inability to reliably use a keyboard or mouse may look like a bottleneck to one expert, but a source of explosive growth to another&#8212;if they expect this human-focussed tooling to get adapted for the AI era. Similarly, different experts may view AI automating a growing share of R&amp;D tasks as progress towards a fast takeoff, or as low-hanging fruit being picked off, accelerating progress only as far as the upcoming wall.</p></li></ul></li><li><p>These differing views are also visible in more recent commentary on the topic.</p><ul><li><p>The prominent AI researcher and writer Nathan Lambert recently <a href="https://www.interconnects.ai/p/lossy-self-improvement">cited</a> Paul Allen concept of a &#8216;complexity break&#8217; to argue that as we understand intelligence better, further progress becomes exponentially harder. In addition to incurring financial costs, Lambert argued that running suites of AI agents won&#8217;t necessarily lead to exponential progress, because those agents will perform best on narrow, verifiable tasks, will be hard to manage in large numbers, and will sample from similar parts of the distribution of AI research ideas, inhibiting more novel breakthroughs.</p></li><li><p>Conversely, Ajeya Cotra at METR, the Model Evaluation and Threat Research organisation, recently wrote about how she &#8220;<a href="https://www.planned-obsolescence.org/p/i-underestimated-ai-capabilities?utm_source=substack&amp;utm_medium=email">underestimated AI capabilities (again)</a>&#8221;.  She argued that AIs may, counterintuitively, find it easier to decompose <a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">longer projects</a> into sub-components that multiple agents can run in parallel, than for shorter tasks. AIs will also produce good documentation for their fellow AIs, which could accelerate progress.</p></li></ul></li><li><p>If faster automation and progress does occur, the CSET authors see two main risks: Less time to prepare for safety risks from AI, and lower human understanding of AI systems. To address these risks, their recommendations have a strong focus on improving access to evidence, including:</p><ul><li><p><strong>New evaluations of AI R&amp;D, </strong>including for &#8216;<a href="https://arxiv.org/pdf/2503.14499">messy</a>&#8217; tasks such as research strategy, which lack clear specifications and success criteria and take place in a dynamic environment with various real-world interactions.</p></li><li><p><strong>New approaches to evaluation</strong> to better distinguish &#8216;degrees of accomplishment&#8217; from a simple success/failure binary.</p></li><li><p><strong>Better insights into how automated R&amp;D is progressing within AI labs,</strong> such as data on how funding is allocated and qualitative impressions of progress from leading AI researchers and engineers.</p><p></p></li></ul></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for free. Lots more in the pipeline. </p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>3. Planes and global warming</h2><ul><li><p><strong>What happened: </strong>A team of researchers, including from Google and American Airlines, published <a href="https://arxiv.org/pdf/2603.06909">results</a> from their latest experiment to use AI to reduce condensation trails from planes&#8212;a key contributor to global warming.</p></li><li><p><strong>What&#8217;s interesting: </strong>When pilots fly, particles from the plane&#8217;s exhaust can mix with low-pressure air to form <em>contrails</em>&#8212;white, artificial clouds, made up of ice crystals. These contrails are a net contributor to global warming, because they trap heat that would otherwise escape. Debates continue over exactly how much they contribute, but one <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC7468346/">estimate</a> suggests that they contribute a lot, causing around 2% of &#8216;radiative forcing&#8217;, which measures how different factors, like CO<sup>2</sup>, heat or cool the planet.</p></li><li><p>As the environmental writer Hannah Ritchie <a href="https://hannahritchie.substack.com/p/contrails-google-ai">explains</a>, more important than the absolute figure is the fact that contrails offer a rare opportunity to reduce global warming almost immediately, at relatively low cost. This is because a small share of flights cause most of the warming-inducing contrails&#8212;generally those that fly through parts of the atmosphere that are both very cold and very humid. If planes take short detours to avoid these patches of air, contrails (and warming) should drop.</p></li><li><p>A few years ago, Google researchers<a href="https://blog.google/innovation-and-ai/technology/ai/ai-airlines-contrails-climate-change/"> partnered with</a> American Airlines on a proof of concept. Using satellite imagery and AI, they were able to predict where contrails would emerge and guide planes to avoid them, reducing contrails by &gt;50%, across 70 test flights.</p></li><li><p>In the latest <a href="https://arxiv.org/pdf/2603.06909">study,</a> they expanded the experiment to 2,400 American Airlines flights from the US to Europe. They placed ~50% of planes in a treatment group, where flight dispatchers were given two choices: a standard flight plan and an alternative contrail-avoidance one. Their decision for which to recommend was voluntary.</p></li><li><p>For flights in this intervention group, contrails fell by 12% compared to a control group with no contrail-avoidance plan. Importantly, the contrail-avoidance routes also did not lead to a significant increase in fuel use. At first glance, these results seem positive, but modest. Digging into the results highlights the challenge of getting useful AI deployed at scale.</p></li><li><p>In particular, dispatchers who received contrail avoidance plans only recommended them to pilots 15% of the time. Even then, the avoidance plan was only <em>successfully</em> flown in 60% of flights. For planes that did successfully follow the avoidance plan, contrails fell by more than 60%, a much larger reduction. So the tech worked, but was often not used.</p></li><li><p>Why? Dispatchers are busy<strong> </strong>and must often deal with other priorities, like bad weather and turbulence. To avoid contrails, planes also need to climb and descend mid-flight. This is safe, but creates more work for pilots and air traffic controllers. As it was voluntary, the incentive to change to a contrail-avoidance plan was weak.</p></li><li><p>The way that the dispatchers received the information also meant that they didn&#8217;t fully understand <em>why</em> the suggested up and down changes were necessary. Happily, the authors feel that most of these obstacles are addressable, with a combination of a better user interface, some automation, and more incentives.</p></li><li><p>In addition to its immediate usefulness, the study is a rare real-world attempt to quantify the benefits of AI to tackling global warming. At the moment, the AI and climate change policy discussion is often negative and focuses on the emissions that may result from building and operating data centres (and other devices) to train and run AI models. This is important, but there are reasons to think that these emissions will be <a href="https://blog.andymasley.com/p/individual-ai-use-is-not-bad-for?open=false#%C2%A7emissions">relatively low</a>, or at least lower than many assume. In contrast, AI could potentially reduce emissions and warming by far larger amounts, for example by accelerating research on solar and fusion power, or making buildings and energy grids more efficient. But these benefits are typically more speculative, harder to quantify, or in the case of contrails, more <em>contingent </em>on human behaviour.</p></li><li><p>This experiment demonstrates that the benefits of AI to tackling global warming are real, but also points to the interventions that will be needed to push them to their full potential.  The study is also timely, given that governments <a href="https://assets.publishing.service.gov.uk/media/69b83baacf4af9cad362b4e7/jet-zero-taskforce-contrail-impact-mitigation-task-and-finish-group-a-strategic-framework-for-uk-contrail-impact-mitigation.pdf">are focussing</a> on contrail avoidance and some policy action may be required, for example to help standardise and mandate contrail prediction software or to generate high-resolution humidity data.</p><p></p></li></ul><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/p/ai-policy-primer-24?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.aipolicyperspectives.com/p/ai-policy-primer-24?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#23)]]></title><description><![CDATA[Science, safety & doctors]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-23</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-23</guid><dc:creator><![CDATA[Conor Griffin]]></dc:creator><pubDate>Thu, 22 Jan 2026 16:57:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!uSFA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uSFA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uSFA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png 424w, https://substackcdn.com/image/fetch/$s_!uSFA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png 848w, https://substackcdn.com/image/fetch/$s_!uSFA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png 1272w, https://substackcdn.com/image/fetch/$s_!uSFA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uSFA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2366393,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.aipolicyperspectives.com/i/185431142?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!uSFA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png 424w, https://substackcdn.com/image/fetch/$s_!uSFA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png 848w, https://substackcdn.com/image/fetch/$s_!uSFA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png 1272w, https://substackcdn.com/image/fetch/$s_!uSFA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56ad262d-3689-4391-9a22-c73d7bbeca94_8000x4500.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: Venus Krier </figcaption></figure></div><h1>1. LLMs are making it easier for scientists to write papers, for better or worse</h1><ul><li><p><strong>What happened: </strong>A team at Cornell and Berkeley<a href="https://www.science.org/doi/epdf/10.1126/science.adw3000"> investigated</a> how scientists are using LLMs to help write papers, and what this means for the future volume, quality and fairness of research.</p></li><li><p><strong>What&#8217;s interesting: </strong>The authors built a dataset of ~2.1 million preprints from arXiv, bioRxiv and SSRN, between 2018-2024. To detect whether scientists had used AI to help write a paper, the team compared the distribution of words in the abstract against human- and LLM-written baselines. When an author&#8217;s paper hit a threshold on this &#8220;AI detection&#8221; metric, they were labelled as an &#8220;AI adopter&#8221;. According to the study, LLM adopters subsequently enjoyed a major productivity boost, compared with non-adopters with similar profiles, publishing 36-60% more frequently. The gains were particularly large for researchers with Asian names at Asian institutions.</p></li><li><p>The team also assessed the complexity of the writing, using measures like<a href="https://readable.com/readability/flesch-reading-ease-flesch-kincaid-grade-level/"> Flesch Reading Ease</a>, which evaluates sentence length and the number of syllables per word. They found that human-written papers with more complex language were more likely to be subsequently accepted by peer-reviewed journals or conferences&#8212;suggesting that, for humans, writing complexity is an (imperfect) signal of research effort and quality. For LLM-assisted papers, the relationship was inverted, with the authors concluding that the polished text of LLMs is helping to disguise lower-quality work. (They validated the findings against a separate dataset).</p></li><li><p>The authors also used the launch of Bing Chat, an LLM-based search engine, in 2023 to conduct a natural experiment. They compared views and downloads on arXiv that Bing Chat had referred, to those that Google Search referred. Bing Chat was more likely to refer scientists to newer and less-cited literature, as well as to books, possibly because LLMs are better able to parse long documents or a larger number of documents. (They also validated this finding with a separate dataset, although we don&#8217;t know how <em>good </em>the new sources cited by Bing were).</p></li><li><p>As the authors note, their study has a number of limitations. Their AI detection method is imperfect, only looks at abstracts, and doesn&#8217;t capture authors who may have edited LLM-generated text. There are also various potential confounders: maybe less experienced researchers are more likely to use LLMs?  That said, the findings highlight (at least) three major questions posed by the growing integration of AI into science:</p><ul><li><p>First, AI is leading to a big increase in the supply of papers (and grant applications). This poses a challenge for preprint repositories, which don&#8217;t want to host slop. ArXiv, whose founder<a href="https://en.wikipedia.org/wiki/Paul_Ginsparg"> Paul Ginsparg</a> is a co-author of this study, recently<a href="https://www.nature.com/articles/d41586-025-03664-7"> banned</a> computer science review and position papers, citing a surge in low-quality AI papers. LLM-assisted papers also pose a challenge for peer reviewers, who are already<a href="https://worksinprogress.co/issue/real-peer-review/"> under strain</a>, and are typically prohibited from using AI, although<a href="https://www.nature.com/articles/d41586-025-04066-5"> many do so anyway</a>. This seems unsustainable. As the authors of this study suggest, it is likely time to consider how to integrate AI into at least some aspects of the peer-review process.</p></li><li><p>Second, the findings illustrate how LLMs may both mitigate and exacerbate fairness issues in science. For some scientists, the complexity of their writing may be a reliable indicator of their thinking and effort. For others, particularly non-native English speakers, writing may be more of an obstacle that has previously penalised them. A hopeful outcome is that LLMs may ease that burden. But a more worrying outcome is that, if reviewers and readers can no longer rely on writing complexity as an (albeit unfair) signal of good work, they may fall back on (even more unfair) signals, such as the institution that a person works at. This challenge is not limited to science, and may also occur in other areas where writing serves this purpose, like with cover letters.</p></li><li><p>Finally, the finding that LLM-based search engines may <em>increase</em> the diversity of sources that researchers review is the opposite of what some suggested would happen: that AI models would continually cite the same high-profile studies, exacerbating the &#8220;<a href="https://en.wikipedia.org/wiki/Matthew_effect">Matthew effect</a>&#8221;.</p></li></ul></li><li><p>Collectively, the study serves as a reminder that for every concerning scenario about the integration of AI into science, there are plausible counter-scenarios. Will AI lessen scientific reliability because of hallucinations? Or will AI &#8220;<a href="https://www.refine.ink/">review agents</a>&#8221; and AI-supported evidence reviews reduce (the many) inaccuracies that are already in the evidence base? Will AI remove the intuitive and serendipitous ideas that humans come up with? Or will AI enable scientists to pursue more novel hypotheses? Ultimately, AI could well upend the standard processes and traditions of science but do so in a way that delivers fresh benefits. To know if and how that is occurring, we need more empirical evidence about how AI is changing science.</p></li></ul><h1>2.  Lessons from two years of AI safety evaluations</h1><ul><li><p><strong>What Happened:</strong> In December, the UK AI Security Institute<a href="https://www.aisi.gov.uk/frontier-ai-trends-report"> shared</a> a set of trends observed since they started to evaluate frontier AI systems in November 2023.</p></li><li><p><strong>What&#8217;s Interesting:</strong></p><ul><li><p>The report features more than 60 authors, a testament to the deep expertise that AISI has built up. Their trends are based on their evaluations of more than 30 frontier AI systems, with methodologies ranging from asking those AI systems questions to adversarially red-teaming them.</p></li><li><p>Their headline finding is striking, if unsurprising: AI capabilities have rapidly improved across all the domains that AISI tests. In the cyber domain, AI models and agents can now successfully complete more than 40% of the 1-hour software tasks they are tested on, up from &lt;5% in 2023. Last year, a model completed an &#8220;expert-level&#8221; cyber task for the first time. In biology and chemistry, AI has gone from significantly underperforming PhD-level human experts at troubleshooting experiments, to significantly outperforming them, including for requests about images.</p></li><li><p>On the risk that AI models may &#8220;self-replicate&#8221; in a way that subverts human control, AISI&#8217;s evaluations suggest that AI agents have gotten better at simplified versions of<a href="https://arxiv.org/html/2504.18565v2"> some tasks</a> that could be instrumental to self-replication, such as passing know-your-customer checks to access financial services, but less so at others, like retaining access to compute and deploying successor agents. AISI&#8217;s evaluations also suggest that models are capable of deliberately obstructing attempts to measure their true capabilities (&#8220;sandbagging&#8221;), but only when explicitly prompted to do so.</p></li><li><p>The report also sheds light on AI systems&#8217; limitations.  In the cyber domain, AISI notes that AI systems still struggle in open-ended environments where they must complete long sequences of actions autonomously. Similarly, regarding chembio threats, biologists and chemists, and potential threat actors, need &#8220;tacit&#8221; knowledge and expertise, such as how to pipette. AISI&#8217;s evaluations to date have focussed more on explicit knowledge although they plan to share more on wet lab tasks.</p></li><li><p>When it comes to mitigations, the report provides both reassurance and concern. On one hand, the safeguards that leading labs have introduced <em>have</em> made their models safer, in one instance increasing the amount of expert effort needed to jailbreak a model by 40x. On the other hand, AISI says that it was still able to find a vulnerability in every AI system it tested.  Worryingly, AISI also found no notable correlation between how capable a model is, and the strength of safeguards it has in place.</p></li><li><p>AISI also sheds light on two other sources of AI risk: open source and scaffolding. They argue that the performance gap between open source and proprietary AI models has narrowed. This introduces risks as safeguards for open models (where they exist) can be removed, and jailbreaks are hard to patch. AISI also found that scaffolding can make AI agents more capable than the underlying base AI models, even if those gaps later narrow when the base models are updated. Some complex scaffolds are in proprietary products, such as coding agents, but others are in<a href="https://poetiq.ai/posts/arcagi_verified/"> open-source</a> efforts.</p></li><li><p>The report also touches on AISI&#8217;s evaluations of the broader societal impacts of AI, such as the degree to which people are using AI to access political information, or the risks of harmful manipulation. One striking statistic, picked up in<a href="https://www.theguardian.com/technology/2025/dec/18/artificial-intelligence-uk-emotional-support-research"> media coverage</a> of the report, was that one-third of UK respondents to a recent AISI survey had used AI for emotional support or social interaction in the preceding year, although just 4% do so daily. In a separate effort, AISI found that some dedicated AI companion users reported signs of &#8220;withdrawal&#8221; during outages.</p></li><li><p>Overall, AISI argues that AI labs are taking an uneven approach to safety, focussing more on safeguards for biosecurity risks, for example, than for other threats. This is arguably true of AISI as well, given their strong focus on biology and chemical risks rather than radiological or nuclear risks. This raises a question: Given finite resources, what evaluations of frontier AI systems are most lacking in the current landscape?</p></li></ul></li></ul><blockquote></blockquote><h1>3. One in four UK doctors are using AI in their clinical practice</h1><ul><li><p><strong>What happened: </strong>The Nuffield Trust and the Royal College of General Practitioners<a href="https://www.nuffieldtrust.org.uk/research/how-are-gps-using-ai-insights-from-the-front-line"> surveyed</a> more than 2,000 UK GPs to understand how they view and use AI, in what the authors called the largest and most up-to-date survey on the topic.</p></li><li><p><strong>What&#8217;s interesting:</strong></p><ul><li><p>28% of UK GPs now use AI. This is up from ~<a href="https://pubmed.ncbi.nlm.nih.gov/30892270/">10% in 2018</a>, but below the rates seen in some other UK professions. According to the survey, the GPs most likely to use AI are younger, male, and work in more affluent areas. This is similar to disparities in the wider public&#8217;s use of LLMs, although there, the early gender gap may have<a href="https://openai.com/index/how-people-are-using-chatgpt/"> narrowed</a>.</p></li><li><p>Just over half of AI-using GPs procure AI tools themselves rather than relying on those that their practices select. This kind of &#8220;shadow AI use&#8221; is not unique to GPs, but a Nuffield focus group sheds light on why UK GPs feel compelled to do it: some GP practices or<a href="https://www.england.nhs.uk/integratedcare/what-is-integrated-care/"> Integrated Care Boards</a> ban AI tools, while others are slow to respond to GPs&#8217; requests and instead prefer to stick with legacy digital tools.</p></li><li><p>UK GPs mainly use AI for clinical documentation and note-taking. Some say that AI note-taking allows them to look at, and speak more, with their patients, a non-trivial benefit given that the UK public<a href="https://www.health.org.uk/reports-and-analysis/analysis/ai-in-health-care-what-do-the-public-and-nhs-staff-think"> worries</a> about AI making healthcare staff more distant.</p></li><li><p>GPs also use LLMs to produce documents, from translations of patient communications to referral letters; and to stay abreast of new research, with some younger practitioners turning to LLM &#8220;study modes&#8221; to help with their mandatory professional development.</p></li><li><p>GPs cite &#8220;saving time&#8221; as the primary benefit of AI, and mainly use this to reduce overtime, rest, and engage in professional development, rather than to see more patients. This is notable as<a href="https://www.gov.uk/government/publications/10-year-health-plan-for-england-fit-for-the-future/fit-for-the-future-10-year-health-plan-for-england-executive-summary"> the UK government wants AI to reduce the wait time</a> to get a GP appointment, which is a top concern for the public. These findings suggest that more nuanced evaluations of AI&#8217;s impact on GP services will be needed.</p></li><li><p>GPs worry about errors and liability issues with AI. As a result, the authors call on tech suppliers to do better evaluations of hallucinations. Ideally, such evaluations would compare the accuracy of AI, human and hybrid outputs in real-world settings, and all the nuances that might entail. For example, when explaining the benefits of AI note-taking, some GPs pointed out that certain colleagues can&#8217;t touch type and so, without AI, struggle to capture all the details in a patient consultation ( this is, presumably, a form of inaccuracy).</p></li><li><p>Use of AI for more complex &#8220;clinical support&#8221; tasks remains relatively low, owing to GPs&#8217; concerns about errors, their desire to retain control over clinical judgement, and a lack of regulatory approval. However, some GPs did report using AI, or wanting to use future systems, to help check diagnoses, formulate care plans, and analyse lab results.</p></li><li><p>This suggests that more GPs may start to use AI to enhance their own clinical judgement, spurred by a growing body of<a href="https://arxiv.org/pdf/2510.22414"> evidence</a> that LLM-based systems may be useful in this area, and by the public&#8217;s own<a href="https://cdn.openai.com/pdf/2cb29276-68cd-4ec6-a5f4-c01c5e7a36e9/OpenAI-AI-as-a-Healthcare-Ally-Jan-2026.pdf"> growing use</a> of LLMs for answering medical questions.</p></li><li><p>In their recommendations, the Nuffield authors call for clearer guidelines and regulatory frameworks for GPs, including as part of the UK&#8217;s new<a href="https://www.gov.uk/government/groups/national-commission-into-the-regulation-of-ai-in-healthcare"> National Commission into the Regulation of AI in Healthcare</a>. However, the report also acknowledges that much guidance already exists, such as the<a href="https://www.bma.org.uk/advice-and-support/nhs-delivery-and-workforce/technology/principles-for-artificial-intelligence-ai-and-its-application-in-healthcare"> British Medical Association&#8217;s AI principles</a> and the<a href="https://www.england.nhs.uk/long-read/guidance-on-the-use-of-ai-enabled-ambient-scribing-products-in-health-and-care-settings/"> NHS guidance</a> on AI note-taking (which some GPs appear to be breaking by procuring their own tools). This raises a question: what exactly should any new guidance stipulate? How to get the burden on GPs right? And how to ensure that they are actually following it?</p></li></ul></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for free to read future pieces. Lots in the pipeline for 2026! </p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#22) ]]></title><description><![CDATA[Agent economies, science, political misinformation & fellowships]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-22</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-22</guid><dc:creator><![CDATA[Conor Griffin]]></dc:creator><pubDate>Fri, 31 Oct 2025 12:25:50 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Ps3v!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Every six weeks, we look at three AI policy developments that caught our eye. As always, we include a &#8216;view from the field&#8217; from an interesting thinker on each item.  Thanks to <a href="https://substack.com/@empiricrafting?utm_source=about-page">Andrey Fradkin</a>, <a href="https://substack.com/@primeposterior?utm_source=about-page">Seth Benzell</a>, and <a href="https://substack.com/@stuartbuck?utm_source=about-page">Stuart Buck</a> for taking part. </em> </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ps3v!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ps3v!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png 424w, https://substackcdn.com/image/fetch/$s_!Ps3v!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png 848w, https://substackcdn.com/image/fetch/$s_!Ps3v!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png 1272w, https://substackcdn.com/image/fetch/$s_!Ps3v!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ps3v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2366457,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.aipolicyperspectives.com/i/177647482?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ps3v!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png 424w, https://substackcdn.com/image/fetch/$s_!Ps3v!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png 848w, https://substackcdn.com/image/fetch/$s_!Ps3v!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png 1272w, https://substackcdn.com/image/fetch/$s_!Ps3v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8da474c-aadb-41f1-b8b0-af9b39059fc6_8000x4500.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: Venus Krier </figcaption></figure></div><h2>1. The AI agent economy</h2><ul><li><p><strong>What happened: </strong>In September, researchers at Google DeepMind published a <a href="https://arxiv.org/html/2509.10147v1">paper</a> examining how AI agents might be integrated into the economy. Gillian Hadfield and Andrew Koh also published a <a href="https://arxiv.org/pdf/2509.01063">paper</a> on the implications of AI agent economies. The papers were timely, with <a href="https://cloud.google.com/blog/products/ai-machine-learning/announcing-agents-to-payments-ap2-protocol">Google</a> and <a href="https://stripe.com/en-fr/newsroom/news/stripe-openai-instant-checkout">OpenAI</a> recently launching new protocols to enable agents to make payments online.</p></li><li><p><strong>What&#8217;s interesting: </strong>AI economic impact debates often position the technology as a &#8216;tool&#8217; and focus on how it may affect employees&#8217; productivity or job prospects. This overlooks the more radical ways that AI agents may change what we even mean by the &#8216;economy&#8217; or &#8216;economic actors&#8217;.</p></li><li><p>Deploying AI agents at scale will be hard. It will require capability improvements and <a href="https://www.aipolicyperspectives.com/p/an-agents-economy">overcoming barriers</a> from legacy infrastructure to the tacit &#8216;grey knowledge&#8217; that humans use to navigate organisations. But there are routes to doing so and the GDM paper argues that as AI agents become more capable and interconnected they will begin to transact with each other, at scale and speeds beyond direct human oversight.</p></li><li><p>One way to analyse the potential effects of this is to compare AI agents with humans. LLMs are trained on economic textbooks and <a href="https://arxiv.org/abs/2502.13119">early research</a> suggests that some AI agent behaviour may be consistent with that of humans, whether in terms of maximising expected utility or displaying common <a href="https://arxiv.org/abs/2301.07543">behavioural biases</a>. However, Hadfield and Koh argue that evaluations of AI agent behaviour are weak and that agents may lead to novel behaviours and impacts, particularly via <em>multi</em>-agent systems:</p><ul><li><p>For example, when it comes to customer welfare, AI agents could be effective personal shoppers, searching widely and continually checking prices. This could lead to better outcomes for consumers, but only if agents correctly infer human preferences and avoid biases towards certain marketplaces. Less positively, agents may develop exploitative strategies, such as seller agents generating a large number of fake reviews to mislead buyer agents, that exacerbate fraud.</p></li><li><p>When it comes to market power, AI agents could help buyers seek out novel competitors, but may also reduce the <a href="https://www.aipolicyperspectives.com/p/coasean-bargaining-at-scale">transaction costs</a> and communication challenges that normally prevent any one firm from becoming too large.</p></li><li><p>AI agents may also exacerbate inequality, as superior agents - equipped with better compute, data, and models - could engage in &#8216;high frequency negotiation&#8217; on behalf of their higher-income users. AI agents may also pose a <a href="http://google.com/url?q=https://papers.ssrn.com/sol3/papers.cfm?abstract_id%3D4452704&amp;sa=D&amp;source=docs&amp;ust=1760966703509053&amp;usg=AOvVaw2a1GQ55lcAAPEsLq5fmaDo">greater risk of collusion</a> than their human counterparts, as the reinforcement learning algorithms used to train them can cause a bias that leads agents to insufficiently explore &#8220;off-path&#8221; strategies.</p></li></ul></li><li><p>Given these challenges, how should we respond? The GDM paper argues that the default response will be to allow the agents to fully <em>permeate</em> the human-led economy, in an <em>emergent</em> or spontaneous manner, with limited safeguards. The authors argue that we should instead aim to prescriptively demarcate (or &#8216;sandbox&#8217;) agents in a controlled sector or section of the economy. This would give policymakers and researchers the opportunity to test them before they are deployed more widely.</p></li><li><p>They also propose various other policy ideas:</p><ul><li><p>Inspired by Ronald Dworkin&#8217;s principles of <a href="https://plato.stanford.edu/entries/justice-distributive/">distributive justice</a>, they propose granting every human user an equal initial endowment of a virtual currency to bid for compute, tools, or priority execution slots on behalf of their agent. They also propose using incentives to steer agents towards socially useful &#8216;missions&#8217;, such as accelerating scientific discovery or tackling climate change.</p></li><li><p>The authors also lay out various technical ideas, such as identifiers for each agent, verifiable credentials that allow agents to build a &#8216;tamper-proof&#8217; reputation, a &#8216;proof-of-personhood&#8217; that links digital accounts to unique human beings, and standards that encourage interoperability between agents.</p></li><li><p>They also propose a hybrid oversight infrastructure that uses AI for real-time monitoring before escalating cases to human experts.</p></li></ul></li><li><p><strong>A view from the field: </strong><em>What are you excited or worried about with respect to AI agents and the economy</em>? <a href="https://substack.com/@empiricrafting?utm_source=about-page">Andrey Fradkin</a> &amp; <a href="https://substack.com/@primeposterior?utm_source=about-page">Seth Benzell</a>, from the <a href="https://empiricrafting.substack.com/podcast">Justified Priors</a> podcast:</p><ul><li><p><strong>Andrey: </strong>&#8220;<em>I am excited by the ways in which markets can be redesigned for AI agents in a way that makes people better off. For example, in the car market, can we create the infrastructure so that a buyer AI agent can find and negotiate a good deal on a car with a lot less human effort?&#8221;</em></p></li><li><p><strong>Seth: </strong>&#8220;<em>In a world where agents and robots can do anything, output will be determined by the level of capital investment. In such a world, the most important growth policy will be national savings policy. High consumption for Boomers and Gen X would require investing less in the future, at exponentially compounding cost to their children. Intergenerational conflict will become more salient.&#8221;</em></p></li></ul></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives.  Subscribe for free to receive new posts. Lots more in the pipeline! </p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>2. Nine ideas to accelerate science with AI</h2><ul><li><p><strong>What Happened:</strong> In August, the Institute for Progress published <a href="https://ifp.org/the-launch-sequence/">nine policy ideas</a> to accelerate the use of AI in science.</p></li><li><p><strong>What&#8217;s Interesting: </strong>The evidence is growing that science will be one of the domains where AI will yield <a href="https://www.aipolicyperspectives.com/p/a-new-golden-age-of-discovery">its greatest benefit to society</a>. Governments are paying more attention. The EU just released an <a href="http://research-and-innovation.ec.europa.eu/document/download/c1afd7d0-ff65-4f84-be48-b0e0949596c5_en?filename=COM_2025_724_1_EN_ACT_part1_v8.pdf">AI for Science strategy</a>, the UK is working on their own strategy, and the US is prioritising science in their new <a href="https://www.whitehouse.gov/wp-content/uploads/2025/07/Americas-AI-Action-Plan.pdf">AI Action Plan</a>. But the policies to pursue are not obvious. How is<em> AI for Science</em> policy different from standard science policy? Or from AI policy? What might an ambitious role for government look like?</p></li><li><p>The IFP provides nine ideas, with a focus on the US. Some aim to improve how science functions, such as <a href="https://ifp.org/the-replication-engine/">using AI agents to replicate scientific papers</a>. Others propose new kinds of organisations, such as <a href="https://ifp.org/scaling-materials-discovery-with-self-driving-labs/">self-driving labs to validate new AI-designed materials</a>; <a href="https://ifp.org/how-x-labs-can-unleash-ai-driven-scientific-breakthroughs/">&#8216;X-Labs&#8217;</a> to work on more ambitious AI projects than grant funding normally allows; and <a href="https://ifp.org/benchmarking-for-breakthroughs/">a new office to commission AI for science &#8216;grand challenges&#8217; and evaluations</a>.</p></li><li><p>The success of AI for Science efforts will hinge on the availability of <a href="https://www.aipolicyperspectives.com/p/a-new-golden-age-of-discovery">a core set of ingredients</a>, the most important of which is <em>data. </em>This is where most of the IFP ideas focus. Adam Marblestone and Andrew Payne propose creating <a href="https://ifp.org/mapping-the-brain-for-alignment/">maps of five small mammal brains,</a> such as laboratory mice, to better understand behaviours that we would like AI systems to learn, such as cooperation. Maxwell Tabarrok proposes <a href="https://ifp.org/a-million-peptide-database-to-defeat-antibiotic-resistance/">a public database of one million antimicrobial peptides</a>, to train AI models to tackle antibiotic resistance.</p></li><li><p>Three other ideas focus on better leveraging the data that already exists, but is inaccessible to most scientists.</p><ul><li><p>Andrew Trask and Lacey Strahm <a href="https://ifp.org/unlocking-a-million-times-more-data-for-ai/">lay out</a> an &#8216;Attribution-Based Control&#8217; system that would allow owners of healthcare, financial, and industrial sensor data to specify AI models that could access it.</p></li><li><p>Ruxandra Teslo <a href="https://ifp.org/biotechs-lost-archive/">argues</a> that LLMs grant an advantage to large pharmaceutical firms that can draw on their historical archives of new drug applications to create &#8216;AI copilots&#8217;, an opportunity that is unavailable to most startups. In response, she proposes a new entity to monitor biotech bankruptcy cases and buy up &#8216;orphaned&#8217; regulatory dossiers and clinical trial data, before anonymising and open-sourcing it.</p></li><li><p>Ben Reinhardt <a href="https://ifp.org/teaching-ai-how-science-actually-works/">argues</a> that most of today&#8217;s AI for Science models are trained on &#8216;clean&#8217; curated datasets and scientific papers. This privileges the final <em>outcome</em> of science research and overlooks the messy <em>process</em> of doing it. In response, he proposes creating &#8216;Unstructured Data Generation Labs&#8217; where scientists would carry out research in fields like biotech and materials science, record themselves using everything from bodycams to equipment sensors, and then use that data to train AI models.</p></li></ul></li><li><p><strong>A view from the field: </strong><em>What AI for Science policy idea are you passionate about?</em> <a href="https://substack.com/@stuartbuck?utm_source=about-page">Stuart Buck</a>, <a href="https://goodscience.substack.com/">The Good Science Project</a>:</p><ul><li><p><strong>Stuart: </strong>&#8220;<em>One policy that could accelerate AI in science isn&#8217;t about AI per se: Funders should sponsor many more direct replications, including in collaboration with the original labs. The reason: so much about science involves both tacit knowledge (which can&#8217;t be articulated) and unwritten knowledge (which can be articulated, but is so routine that no one even thinks to mention it). Most of that knowledge isn&#8217;t accessible to AI currently, but if we carried out more direct replications in tandem with AI tools, we could make quicker progress towards figuring out all of the unseen and unwritten factors that explain why an experiment reached the results it did. See this <a href="https://goodscience.substack.com/p/hot-dogs-cancer-cells-replication">essay</a> from the Good Science Project.&#8221;</em></p></li></ul></li></ul><h2>3. Using LLMs for political guidance</h2><ul><li><p><strong>What happened: </strong>Researchers at the UK AI Security Institute <a href="https://arxiv.org/pdf/2509.05219">published results</a> from a randomised controlled trial which found that individuals who used LLMs to research political information before the 2024 UK election were subsequently less likely to believe false information.</p></li><li><p><strong>What&#8217;s interesting: </strong>The authors first ran a survey of ~2,500 UK adults and found that 9% of voters, or ~1/3rd of chatbot users, used LLMs to get political information in the week before the July 2024 election. Most LLM users found the models useful and accurate.</p></li><li><p>The share of chatbot users who used the models to get political information is quite high, and is likely higher again in October 2025. But LLMs are still well behind other sources of political information, such as television, social media, and search engines.</p></li><li><p>The authors then ran an RCT of UK residents. The first group was given access to an LLM and asked to research issues of concern to the UK election, such as climate change, immigration, criminal justice and Covid-19 policy. The control group was given access to a search engine. The study found that both groups subsequently showed similar declines in belief in false information and similar increases in belief in true facts. The main difference was <em>efficiency</em> - the LLM group completed their task 6-10% quicker. The results held across different AI models (GPT-4o, Claude, Mistral) and also held when the models were prompted to be more sycophantic.</p></li><li><p>The results suggest that widespread concerns about LLMs exacerbating political misinformation may be misplaced, which in turn may reflect how hallucination rates have dropped over the past two years. The speed of LLMs also means that they could potentially help to debunk fast-spreading misinformation more quickly, enabling what the authors describe as &#8220;<em>rapid, reliable public learning during high-stakes events.</em>&#8221;</p></li><li><p>The results also challenge the common survey finding that the &#8216;public doesn&#8217;t trust&#8217; AI. Rather, the authors find that &#8216;information seeking&#8217;, including for political information, is one of the public&#8217;s main AI use cases. This highlights the need to judge public attitudes to AI based on &#8216;revealed&#8217; as well as &#8216;stated&#8217; preferences.</p></li><li><p>The study comes with caveats and limitations. It focussed on one country and only tested a small number of models - others may be more likely to generate political misinformation. </p></li><li><p>The study also evaluates LLMs by comparing them against an antecedent technology: a non-AI search engine. But as the authors note, such comparisons are increasingly difficult as LLMs are now integrated more directly into the search experience. This will make it harder to know what baseline to compare future LLMs against.</p></li><li><p><strong>A view from the field: </strong><em>How do you see AI changing access to political information?</em> <a href="https://substack.com/@tomrachman">Tom Rachman</a>, Google DeepMind:</p><ul><li><p><strong>Tom: </strong>&#8220;<em>How AIs will remake the news ecosystem is a matter of vast import to democracies. Each person&#8217;s evaluation of information is mediated by their trust in its source, particularly for political content. One could envisage a future in which different AI models gain specific political reputations, affecting their influence as sources. Another plausible future could see personalised AI agents as everyone&#8217;s fundamental font. This study, among <a href="https://arxiv.org/pdf/2507.13919">other ambitious experiments</a> led by researchers at the AI Security Institute, helps establish a baseline effect as we wait for new information paradigms to crystallise.&#8221;</em></p></li></ul></li></ul><h2>Bonus: AI Policy Fellows of the World, Unite!</h2><p>Every year, AI fellowships send fresh outstanding minds into the world of policy. But fellowships are more than points of transit; they are sources of valuable research. To highlight this, we scanned recent projects from leading programs at the <a href="https://www.governance.ai/">Centre for the Governance of AI</a>, <a href="https://erafellowship.org/">ERA Cambridge</a>, <a href="https://www.pivotal-research.org/fellowship">Pivotal</a>, <a href="https://www.matsprogram.org/">MATS</a> and <a href="https://pibbss.ai/fellowship/">PIBBSS</a>. What struck us was the sheer range of insightful work. While we cannot list all the excellent contributions, here are three that caught our attention:</p><ul><li><p><strong>Jacob Schaal,</strong> a Cambridge ERA fellow, extracted economic insight from Moravec&#8217;s Paradox that what is easy for AI is hard for humans, and vice versa, conceiving a new way to judge which jobs are most exposed to automation. Management and STEM occupations, he found, face the highest automation exposure. <strong>Interested in more?</strong> Ask for details from Jacob at jacobvschaal@gmail.com</p></li></ul><ul><li><p><strong>Said Saillant</strong>, a GovAI summer fellow who recently joined UNIDO&#8217;s Innovation Lab, developed the concept of AI-ready special economic zones, or AI-SEZ, for Latin America and regulatory sandbox models for the UK. His work focused on adaptive regulation to accelerate safe AI diffusion. <strong>Interested in more?</strong> Ask for details from Said at ssaillant@societassapiens.org</p></li></ul><ul><li><p><strong>Jo&#235;l Naoki Christoph,</strong> a fellow at GovAI, argued that middle powers chasing &#8220;sovereign compute&#8221; are walking into a costly trap because such projects fail to achieve real autonomy. Instead, he advocates for &#8220;managed dependency,&#8221; allowing nations to avoid fruitless expenditure and reduce foreign leverage. <strong>Interested in more?</strong> Ask for details from Jo&#235;l at <a href="mailto:jchristoph@hks.harvard.edu">jchristoph@hks.harvard.edu</a></p></li></ul><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives.  Subscribe for free to receive new posts. Lots more in the pipeline!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#21)]]></title><description><![CDATA[Expertise, metascience & cognitive debt]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-21</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-21</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Fri, 15 Aug 2025 09:08:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!naiZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Every month, we look at three AI policy developments that caught our eye. Today, we cover how AI may affect demand for expertise, metascience, and human cognitive abilities. We once again include a &#8216;view from the field&#8217; from an interesting researcher on each topic. Thanks to <a href="https://mariadelriochanona.info/">Maria del Rio Chanona</a>, <a href="https://eamonduede.com/">Eamon Kenneth Duede</a>, and <a href="https://www.empiricallykev.com/">Kevin Mckee </a>for lending their time &amp; expertise.</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!naiZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!naiZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp 424w, https://substackcdn.com/image/fetch/$s_!naiZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp 848w, https://substackcdn.com/image/fetch/$s_!naiZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp 1272w, https://substackcdn.com/image/fetch/$s_!naiZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!naiZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp" width="800" height="503" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:503,&quot;width&quot;:800,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:66542,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.aipolicyperspectives.com/i/171013845?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!naiZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp 424w, https://substackcdn.com/image/fetch/$s_!naiZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp 848w, https://substackcdn.com/image/fetch/$s_!naiZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp 1272w, https://substackcdn.com/image/fetch/$s_!naiZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c48f88d-d91c-4d04-a77d-7f7aa357b72c_800x503.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives ! Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1><strong>Study Watch</strong></h1><h2>AI&#8217;s impact on jobs will depend on how it affects &#8216;expertise&#8217;</h2><ul><li><p><strong>What happened: </strong>The economists David Autor and Neil Thompson published a <a href="https://www.nber.org/papers/w33941">paper</a> in which they argue that the concept of &#8216;expertise&#8217; explains why, over the past 40 years, new technologies have led to higher wages and lower employment for certain US roles, and the inverse for others. They expect similar dynamics in the AI era.</p></li><li><p><strong>What&#8217;s interesting: </strong>Autor and others previously coined the idea of &#8216;<a href="https://id.elsevier.com/as/authorization.oauth2?platSite=SD%2Fscience&amp;additionalPlatSites=GH%2Fgeneralhospital%2CMDY%2Fmendeley%2CSC%2Fscopus%2CRX%2Freaxys&amp;scope=openid%20email%20profile%20els_auth_info%20els_idp_info%20els_idp_analytics_attrs%20els_sa_discover%20urn%3Acom%3Aelsevier%3Aidp%3Apolicy%3Aproduct%3Ainst_assoc&amp;response_type=code&amp;redirect_uri=https%3A%2F%2Fwww.sciencedirect.com%2Fuser%2Fidentity%2Flanding&amp;authType=SINGLE_SIGN_IN&amp;prompt=login&amp;client_id=SDFE-v4&amp;state=retryCounter%3D0%26csrfToken%3D341cf1b4-dc1d-4207-8807-49240a1bd9a8%26idpPolicy%3Durn%253Acom%253Aelsevier%253Aidp%253Apolicy%253Aproduct%253Ainst_assoc%26returnUrl%3D%252Fscience%252Farticle%252Fabs%252Fpii%252FS0169721811024105%26prompt%3Dlogin%26cid%3Darp-b24a3c44-db99-4ad0-ac22-48ee7cecd85c">skills-based technological change</a>&#8217; to describe how new technologies complement some workers, while displacing others. What causes this? Autor and Thompson provide a partial answer via the concept of &#8216;expertise&#8217;, which is also the full title of their new <a href="https://www.nber.org/papers/w33941">paper</a>.</p></li><li><p>In Autor and Thompson&#8217;s model, expertise describes a worker&#8217;s ability to perform tasks that others cannot. They view expertise as <em>hierarchical </em>&#8212; a senior surgeon can do their own job, but they could also perform the tasks of a junior nurse (e.g. taking blood pressure). However, the junior nurse cannot perform complex surgery.</p></li><li><p>They turn this definition into a statistical assessment of the relevant expertise of different tasks and jobs, by assessing the frequency and entropy of words used to describe them. Words like &#8216;elasticity&#8217;, which tend to be used rarely and in specific contexts, like economics, are more likely to describe high-expertise tasks.</p></li><li><p>Armed with this definition and statistical measure, Autor and Thompson assess ~40 years of US job data to understand how new technologies have affected different jobs. They find that when technology automates a task, the effects on workers hinge on whether that task requires lower or higher expertise.</p><ul><li><p>When technology is<strong> </strong><em><strong>complementary</strong></em><strong>,</strong> it <em>augments </em>experts by automating <em>inexpert tasks. </em>For example, computers automated data entry tasks, allowing accounting clerks to shift to more complex, specialised analysis. Wages rose, from an average of ~$13 in 1980 to ~$18 in 2018, but overall employment fell, from 1.6m to 1.1m, as the pool of qualified workers shrank.</p></li><li><p>Conversely, when technology removes the most complex expert tasks from roles, it diminishes or<strong> </strong><em><strong>displaces</strong></em> the value of experts.<strong> </strong>For inventory clerks, computers automated more specialised pricing tasks, reducing the expertise of the job. Wages fell, from ~$14 in 1980 to ~$12 in 2018, but overall employment rose, from 0.5m workers to 1.5m.</p></li></ul></li><li><p>What does this mean for AI&#8217;s future effects? The authors argue that we should focus less on wondering whether AI will automate certain<em> jobs</em>, and more on the relative expertise of the tasks that AI will automate, within jobs.</p></li><li><p>Think of any job (perhaps the one that you do!). An AI policy researcher may read papers, build relationships with experts, and write analysis and recommendations for internal and external audiences. Today, that researcher can use an LLM to assist with literature reviews, preparing expert interviews, or reviewing their writing, freeing them up to focus on the more expert tasks that LLMs cannot yet do. Per Autor and Thompson&#8217;s theory, the LLM is likely making the AI policy researcher more productive, and their expertise more valuable. But it may also impose barriers to entry for new entrants by automating some of the traditional tasks they might have worked on. As the models continue to improve, the calculus could change. If they begin to encroach on tasks that currently require more expertise, organisations may find it more efficient to hire more 'inexpert&#8217; employees to use the more powerful LLMs, or rely more on AI agents in lieu of human employees.</p></li><li><p>Practitioners are exploring various evaluations to understand AI&#8217;s potential impact on employment. This paper suggests that one goal should be assessing the relative &#8216;expertise&#8217; of the tasks that AI systems can do, in an economic context, and what that means for workers with that expertise. The paper also raises questions about the underlying concept of expertise that might merit future study, including:</p><ul><li><p><strong>Is expertise always hierarchical? </strong>Or are there tasks that a more junior employee may excel at, such as serving customers, where a more senior executive would struggle?</p></li><li><p><strong>Do job task descriptions capture what makes employees valuable?</strong> Or are there other forms of expertise and value that are less visible, such as organisational knowledge, trust, and judgement?</p></li><li><p><strong>How to account for wider economic trends?</strong> Autor and Thompson acknowledge that some of the trends they highlight are &#8216;<em>almost surely explained by a combination of automation and international trade</em>.&#8217; Beyond trade, demands for expertise in a given location will also depend on factors like outsourcing, policy shifts, new business models, and wider demand trends in the economy.</p></li><li><p><strong>How to best develop expertise?</strong> Autor and Thompson&#8217;s analysis also raises the question of how people can develop the right expertise to capture the complementary benefits of AI. As we <a href="https://www.aipolicyperspectives.com/p/ai-and-the-retraining-challenge">recently analysed</a>, this is where further research on the role of worker retraining would be valuable - as well as a sober understanding of where retraining may be insufficient.</p></li></ul></li><li><p><strong>View from the field:</strong> <strong>How do you see AI affecting demand for expertise?</strong> <a href="https://profiles.ucl.ac.uk/97689-maria-del-riochanonahttps://profiles.ucl.ac.uk/97689-maria-del-riochanona">Maria del Rio Chanona, University College London</a></p><ul><li><p><em>&#8220;Whether AI expands or contracts expertise depends on the type of tasks it affects. <a href="https://www.sciencedirect.com/science/article/pii/S0167268124004591">Our research</a> shows that for substitutable tasks like content writing and translation, AI decreases demand across all levels of expertise - i.e. the technology may be good enough to replace what previously required top-tier human expertise. For most complementary work, like JavaScript coding, HTML development, or general programming projects, we see the opposite pattern: AI raises the expertise bar by eliminating demand for novice workers while experienced developers remain sought after. It's worth noting that our findings focus on labor demand patterns rather than wages or employment outcomes - we're observing how employers' willingness to hire for different types of expertise is shifting, which may precede but doesn't necessarily translate directly to changes in worker compensation or employment levels.&#8221;</em></p></li></ul></li></ul><h1><strong>Policymakers taking action</strong></h1><h2>The UK Metascience Unit shares early results and AI plans</h2><ul><li><p><strong>What Happened:</strong> The UK Metascience Unit <a href="https://assets.publishing.service.gov.uk/media/685a83af72588f418862071d/a-year-in-metascience-2025.pdf">shared results</a> from their early experiments to reform UK science, and spotlights upcoming work, including a strong focus on AI.</p></li><li><p><strong>What&#8217;s Interesting: </strong>Metascience, or &#8216;<em>the science of science itself&#8217;, </em>uses research, data and experiments to improve how science works.</p><ul><li><p>As detailed in <a href="https://assets.publishing.service.gov.uk/media/685bcd40c07c71e5a87097d1/the-past-present-future-of-uk-metascience.pdf">an appendix</a>, this desire to understand and shape science has a long history. In 1939, the pioneering X-ray crystallographer and Marxist J.D. Bernal published &#8216;<a href="https://www.faber.co.uk/product/9780571272723-the-social-function-of-science/?srsltid=AfmBOoplLi2TIGOke9sYpOBC71wglvCfTSwIpjdbkFmSAMVOL2kwRnxR">The Social Function of Science</a><em>&#8217;. </em>In it, Bernal analysed British science &#8220;<em>as if viewing it through his microscope</em>&#8221;, assessing everything from its funding and organisation to its role in industry and war.</p></li><li><p>In subsequent decades, the UK continued to advance the foundation of modern metascience via a &#8216;golden triangle&#8217; of universities in Sussex, Manchester and Edinburgh. More recently, metascience has served as a meeting place for individuals with varied objectives, from supporting open science to improving replication - the latter goal electrified by John Ioannidis&#8217;s 2005 paper, &#8216;<a href="https://journals.plos.org/plosmedicine/article?id=10.1371/journal.pmed.0020124">Why Most Published Research Findings Are False</a>&#8217;.</p></li><li><p>The greatest metascience spur is the fact that science is now extremely large - ~$2.5 trillion per year - and creaking. Observers warn of <a href="https://scienceplusplus.org/trouble_with_rcts/">slow and conservative funding processes</a>, a <a href="https://onlinelibrary.wiley.com/doi/10.1002/leap.1544">broken peer review system</a>, and a <a href="https://rori.figshare.com/articles/preprint/_b_A_new_typology_of_national_research_assessment_systems_b_b_continuity_and_change_in_thirteen_countries_b_b_b_b_RoRI_Working_Paper_b_b_No_15_b_b_b_/29041787/4?file=54670646">continued over-reliance</a> on metrics - such as publications and citations - that create perverse incentives.</p></li></ul></li><li><p>Various actors hope to use metascience to address these issues, while taking care to ensure that their work is not used to justify ill-founded cuts in scientific research. Among these actors, the UK Metascience Unit is the first of its kind to be embedded within both a central government department (DSIT) and the country&#8217;s largest research funder (UKRI). This gives it the potential to directly translate experimental findings into national policy. Their ~&#163;3-4m budget is small, accounting for just ~0.03% of UKRI&#8217;s, but they have also secured funding from third-party sources, such as Open Philanthropy.</p></li><li><p>How are they spending this money? In their first year, they focussed on improving the processes used to allocate research funding and ran a successful trial of &#8216;<em>distributed peer review</em>&#8217; &#8212; a format where funding applicants must also agree to review other applications, and which gained prominence in the scientific world after a canonical experiment to allocate &#8216;<a href="https://ui.adsabs.harvard.edu/abs/2009A%26G....50d..16M/abstract">telescope time without tears</a>&#8217;. In their experiment, the Unit used their own AI Metascience Fellowship programme as a test case and took steps to address potential concerns, such as gaming issues. They found that, overall, distributed peer review shortened the assessment process, reduced the admin burden, and improved participants' knowledge of the field.</p></li><li><p>The Unit also ran simulations and trials of &#8216;<em>partial randomisation</em>&#8217; - a process that subjects &#8216;middle-ranking&#8217; grant applications to a lottery process, in the hope of encouraging more novelty, risk-taking, and efficiency. However, the Unit found that the evidence for such randomisation is not yet sufficient.</p></li><li><p>In the year ahead, AI will be a major focus for the Unit. In particular:</p><ul><li><p>Their<a href="https://www.ukri.org/opportunity/ukri-metascience-ai-early-career-fellowships/"> 18 early career fellows</a> will study how AI is affecting science. As we touched on <a href="https://www.aipolicyperspectives.com/p/a-new-golden-age-of-discovery">in a past essay</a>, which the Unit cites, the questions here are vast - from how AI is affecting the methods that scientists use, and the pace of scientific progress, to AI&#8217;s effects on scientific creativity and understanding.</p></li><li><p>The Unit has also allocated grants to researchers working on specific AI-related questions, such as to assess if LLMs can reliably review academic research, and if AI could prevent problematic randomised controlled trials from being included in systematic reviews, where they can hurt patients.</p></li><li><p>The Unit is also co-running a global competition to find and validate AI-driven indicators of &#8216;scientific novelty&#8217;, to support efforts to understand if novelty truly is lacking in scientific research or whether the opposite may be true e.g. we need a stronger push to coalesce, deepen and replicate <em>existing </em>research.</p></li></ul></li><li><p><strong>View from the field: </strong><em>What effects from AI on science should metascientists be exploring?</em> <a href="https://eamonduede.com/">Eamon Duede, Purdue University</a></p><ul><li><p><em>&#8220;There is a tendency in metascience research to treat science as something rather monolithic and to ask questions, the answers to which generalize to all of science. This approach has been enormously illuminating. But when it comes to grappling with the impact of AI on science, it is likely to limit what we can learn. In contrast to prior domain-specific innovations, contemporary AI systems, particularly LLMs, will impact every discipline, yet do so in ways that differ profoundly from one field to another.</em></p></li><li><p><em>So rather than only asking how AI affects science in aggregate, we should ask how it differentially transforms the distinctive epistemic aims, methodological norms, and evaluative standards of physics versus history, or philosophy versus biology. Grappling with this question promises more than just antiquarian insights into AI&#8217;s role in research. Rather, it offers a powerful new lens through which to understand the very nature of science itself.&#8221;</em></p></li></ul></li></ul><h1><strong>Study Watch</strong></h1><h2>AI and cognitive debt</h2><ul><li><p><strong>What happened: </strong>Authors from MIT Media Lab published a <a href="https://arxiv.org/pdf/2506.08872">widely-discussed preprint</a>, in which students who used LLMs for essay writing reported lower levels of brain activity and later struggled to recall quotes from their essays.</p></li><li><p><strong>What&#8217;s interesting: </strong>Observers have long fretted that new technologies will hurt students&#8217; ability to learn. In Plato&#8217;s dialogue <em>Phaedrus, </em>Socrates <a href="https://talesoftimesforgotten.com/2018/08/18/misunderstood-ancient-quotes/">worried</a> that writing would create forgetfulness in learners&#8217; souls. The towering Renaissance figure Conrad Gessner helped to establish the fields of bibliography and zoology, partly <a href="https://oxfordre.com/politics/display/10.1093/acrefore/9780190228637.001.0001/acrefore-9780190228637-e-1360?d=%2F10.1093%2Facrefore%2F9780190228637.001.0001%2Facrefore-9780190228637-e-1360&amp;p=emailAsJyxUpfEDl5U">out of fear</a> that the printing press would overwhelm learners with &#8216;information overload&#8217;. Observers have raised similar concerns about calculators, television, the internet, and now AI.</p></li><li><p>In this study, the authors randomly assigned 54 university students in Boston into one of three groups. Across three sessions, each group had to write an essay on an SAT topic in 20 minutes. The first group used an LLM, the second used a search engine, and the third had to rely solely on their brains. An EEG headset monitored the students&#8217; brain activity. At the end, the LLM and Brain-only students were given the option to swap groups and participate in an optional fourth session.</p></li><li><p>The authors reported three main effects:</p><ul><li><p><strong>Reduced neural connectivity: </strong>According to EEG data, the LLM-only group showed the weakest neural connectivity - a proxy for cognitive effort - while the Brain-only group showed the strongest. When the LLM-only users had to rely only on their brains in the voluntary fourth session, their neural connectivity did not rebound to the level of the Brain-only group.</p></li><li><p><strong>Worse memory: </strong>15 of the participants in the LLM group failed to provide a correct quote from the essay they had just written (83%), while only two students in each of the other groups had the same difficulty. In interviews, LLM group participants reported a weaker sense of ownership over their work.</p></li><li><p><strong>Homogenised language: </strong>Students using LLMs also produced essays that were more linguistically similar to one another - tying into a broader concern about AI and homogeneity, that Kalim Ahmed <a href="https://www.aipolicyperspectives.com/p/the-internet-is-a-place-where-no">recently explored</a> in an essay on this site.</p></li></ul></li><li><p>The study has limitations, as <a href="https://thebsdetector.substack.com/p/the-cognitive-debt-of-digging-through">one review</a> points out. It relies on a very small sample of elite students that drops further for the optional fourth session. The authors also ran a very large number of tests over the EEG data, a type of data that can be challenging to interpret, raising concerns about p-hacking. More fundamentally, the study converts relatively unsurprising results &#8212; if you use an LLM to help you write an essay in 20 minutes you will struggle to remember quotes from that essay &#8212; into a very strong claim: that LLM use will lead to &#8216;cognitive debt&#8217; that impedes students&#8217; future learning.</p></li><li><p>The reality will likely be more nuanced. As John Sweller&#8217;s <a href="https://cognitiveloadtheory.wordpress.com/">Cognitive Load Theory</a> has illustrated, more cognitive effort is not always good for learning. Some cognitive load is good, because you are thinking hard about what matters. But some of it is <em>extraneous, </em>and a barrier to learning, such as the &#8216;<a href="https://en.wikipedia.org/wiki/Split_attention_effect">split-attention</a><em>&#8217; and</em> <a href="https://www.researchgate.net/profile/Richard-Mayer-4/publication/228698670_Cognitive_Principles_of_Multimedia_Learning_The_Role_of_Modality_and_Contiguity/links/57799c7608aead7ba0764344/Cognitive-Principles-of-Multimedia-Learning-The-Role-of-Modality-and-Contiguity.pdf">&#8216;modality</a>&#8217; effects that arise when students are presented with a confusing jumble of text and images.</p></li><li><p>In some scenarios, LLMs could reduce cognitive load in a way that allows students to go deeper on a topic of interest, such as by providing a more compelling, integrated learning experience. Another positive scenario might see students using LLMs as a sort of &#8216;<a href="https://en.wikipedia.org/wiki/Extended_mind_thesis">extended mind</a>&#8217; to automate certain tasks, in pursuit of higher-order thinking.</p></li><li><p>However, these scenarios all require that students be motivated to learn in the first instance. Some worry that the ready-availability of LLMs may reduce such motivation, particularly among younger students developing foundational skills. This experiment doesn&#8217;t shed light on whether that scenario is happening. Rather, we would need other kinds of evaluations to assess how different kinds of students are using LLMs in the real world.</p></li><li><p><strong>View from the field: </strong><em>How should practitioners study the effects of AI on cognitive load? </em><a href="https://www.linkedin.com/in/kevin-mckee-b729b092/?originalSubdomain=uk">Kevin McKee, Google DeepMind</a></p><ul><li><p><em>&#8220;Randomised controlled trials are particularly helpful for questions like this because they force us to think about what specific skills we care about and what measurements we can take to know if they've actually changed. And of course, if we want to understand their effects on students' independent cognitive abilities &#8211; how well they're able to function when they can't rely on LLMs &#8211; we'll have to specifically design RCTs in ways where we're confident students aren't accessing LLMs at test time.</em></p></li><li><p><em>As a complement to that, we should also think about in-depth studies that can examine how students try to solve problems or tackle self-study lessons. A well-designed &#8216;narrow&#8217; study would help by shedding light on the mechanisms at play &#8211; like how students might be replacing some of their cognitive work with LLMs &#8211; while also giving us a better qualitative understanding of students' experiences, including how they feel after working with an LLM.&#8221;</em></p></li></ul></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives ! Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#20)]]></title><description><![CDATA[The economy, the environment, and where Dean Ball thinks AI is headed]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-18</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-18</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Thu, 12 Jun 2025 09:57:14 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!-rsJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Every month, we look at three AI policy developments that caught our eye. Today, we cover how AI may affect the economy, the environment, and Dean Ball&#8217;s views on AI liability and governance. In response to a reader suggestion (thanks!), we also include a &#8216;view from the field&#8217; from an interesting thinker on each topic. Thanks to <a href="https://law-ai.org/team/gabriel-weil-2/">Gabe Weil</a>, <a href="https://www.governance.ai/team/sam-manning">Sam Manning</a> &amp; <a href="https://andymasley.substack.com/">Andy Masley</a> for lending their time &amp; expertise.</em></p><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-rsJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-rsJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg 424w, https://substackcdn.com/image/fetch/$s_!-rsJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg 848w, https://substackcdn.com/image/fetch/$s_!-rsJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!-rsJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-rsJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg" width="1000" height="563" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:563,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:148936,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.aipolicyperspectives.com/i/165705652?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-rsJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg 424w, https://substackcdn.com/image/fetch/$s_!-rsJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg 848w, https://substackcdn.com/image/fetch/$s_!-rsJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!-rsJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96f393f2-8d52-40fe-8fd6-dbc9d6388f31_1000x563.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em> </em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives. Please share your take &amp; subscribe for free to receive new posts. Lots more in the pipeline! </p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Influential views</h1><h2>Where Dean Ball thinks AI is headed &amp; how to govern it</h2><ul><li><p><strong>What happened: </strong>In April, prior to joining the White House Office of Science and Technology Policy as an AI policy advisor, Dean Ball published a two-part essay series outlining his <a href="https://www.hyperdimensional.co/p/where-we-are-headed">expectations for AI</a> and <a href="https://www.hyperdimensional.co/p/where-we-are-headed-part-ii">how policymakers should respond</a>.</p></li><li><p><strong>What&#8217;s interesting: </strong>In the <a href="https://www.hyperdimensional.co/p/where-we-are-headed">first essay</a>, Dean lays out his core thesis: we are on the brink of powerful AI agents that will be able to source information, use software tools, and communicate: &#8220;<em>These abstract tasks do not constitute everything a knowledge worker does, but they constitute a very large fraction&#8230;</em>&#8221;.</p><ul><li><p>Early AI agents can be glitchy, but as AI labs put them into reinforcement learning systems, they will get better. As Dean notes, this will be easier in domains like maths, where outputs can be more easily verified. But even in more subjective areas - like writing a newsletter! - AI systems can increasingly review each other&#8217;s outputs, which will accelerate progress.</p></li><li><p>As they deploy growing fleets of AI agents, Dean expects organisations that rely on knowledge workers to become more efficient and profitable. They may also become <em>stranger</em> <em>-</em> heavier at the top and so more variable in character, with leaders able to rely on agents for better information flow and control.</p></li><li><p>Widespread job loss may happen, particularly if prompted by a recession, but this may also be offset, or delayed, by the &#8220;in-person&#8221; aspects that many jobs have, or by regulations requiring &#8220;human alternatives&#8221; to AI decisions. In the near-term, young people entering the labour market may be the most affected.</p></li><li><p>Dean also expects transformative progress from the use of AI in science, from new cancer cures to room-temperature semiconductors, but these will take longer, as data still needs to be gathered and real-world experiments run. The prospects for the parts of the economy that are largely offline - like social care or the construction industry - are not analysed.</p></li></ul></li><li><p>In a <a href="https://www.hyperdimensional.co/p/where-we-are-headed-part-ii">2nd essay</a>, and an <a href="https://arxiv.org/pdf/2504.11501">accompanying paper,</a> Dean examines how the US government should respond. Dean&#8217;s starting point is that AI will become a &#8220;foundational technology&#8221; - closer to a natural resource like energy than to, say, social media. Past foundational technologies, like railroads, telecoms networks, electricity, and the Internet all differ, in form and function, so commonalities in how we govern them could map to AI.</p><ul><li><p>The main commonality Dean sees is that the US eliminated or severely limited the exposure of providers of these foundational technologies to tort liability for downstream misuse of their products. Dean does not think AI developers should face no liability - if a data centre explodes due to mismanagement or an agent exfiltrates itself and defrauds people - companies should face different types of statutory liability, much like power providers do today.</p></li><li><p>But Dean argues that attempts to design AI liability schemes that go beyond this, to impose a &#8220;reasonable care&#8221; standard on AI model developers to foresee and prevent a wider range of downstream harms could be weaponised. Building his previous <a href="https://www.hyperdimensional.co/p/how-should-ai-liability-work-part">two-part essay series</a> on liability, he notes how third parties, including anonymous investors, can fund and extend US liability cases and how the target is often those with the most resources, rather than those who are most directly responsible for the harm. Relying on liability can also mean that judges and juries are effectively determining how to govern frontier AI systems and what good safety practices look like.</p></li><li><p>What is Dean&#8217;s proposed solution? Building on Gillian Hadfield&#8217;s work on &#8216;<a href="https://arxiv.org/abs/2304.04914">regulatory markets</a>&#8217;, he proposes a framework where governments would authorise private bodies to develop safety standards that AI companies could voluntarily opt to be certified and audited against. The AI companies that opt in would receive a &#8216;safe harbour&#8217; from tort liability stemming from third parties&#8217; misuse of their models. The goal would be to support AI innovation, while providing incentives for safety and encouraging a marketplace of ideas for how to best achieve it.</p></li></ul></li><li><p><strong>View from the field:<a href="https://law-ai.org/team/gabriel-weil-2/"> Prof Gabriel Weil, LawAI</a>:</strong></p><ul><li><p><em><strong>&#8220;</strong>Tort law is especially useful for mitigating risks from AI (over which there is substantial disagreement and uncertainty) because (unlike ex-ante regulation) it scales automatically with the risk, and shifts the onus to AI companies, where the relevant expertise is concentrated, to figure out how to make their systems safe. Voluntary private certification is poorly situated to protect third parties, since there is no market feedback to induce certifiers to craft rules that protect non-users and prevent a race to the bottom.</em></p></li><li><p><em>To read more, see my<a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5283275"> recent working paper</a> which argues that liability should be the governance tool of first resort for AI risk, my shorter<a href="https://www.lawfaremedia.org/article/tort-law-should-be-the-centerpiece-of-ai-governance#:~:text=The%20key%20advantage%20of%20liability,policy%20problem%20of%20frontier%20AI"> LawFare piece</a>, and my<a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4694006"> earlier paper</a> on using tort law to mitigate catastrophic risk from AI. On the mixed relationship between liability and innovation, the risk of excessive litigation, and the nuances that apply here, see my<a href="https://law-ai.org/draft-report-of-the-joint-california-policy-working-group/"> recent piece</a> with Mackenzie Arnold.&#8221; </em></p></li></ul></li></ul><h1>Study watch</h1><h2>In Denmark, chatbots aren&#8217;t turbocharging productivity growth</h2><ul><li><p><strong>What Happened:</strong> In May, Anders Humlum and Emilie Vestergaard from the universities of Chicago and Copenhagen published <a href="https://www.nber.org/papers/w33777">a new analysis</a> of chatbot adoption among ~25,000 Danish workers in 11 occupations that are &#8216;exposed&#8217; to AI, such as journalists, customer service employees and software developers. They found that while chatbot use is high, and growing, this has not yet resulted in a statistically significant impact on productivity growth.</p></li><li><p><strong>What&#8217;s Interesting: </strong>The analysis is based on two large surveys that the authors conducted in 2023-24 to understand how Danish employees were using chatbots and what the perceived impacts were. The authors then used social security numbers to match the survey data with government data on employment and earnings. Using a <em><a href="https://dimewiki.worldbank.org/Difference-in-Differences#:~:text=The%20difference%2Din%2Ddifferences%20method,useful%20tool%20for%20data%20analysis.">difference-in-differences</a> </em>approach - which attempts to mimic a randomised controlled trial using real-world data - they then analysed the chatbots&#8217; impacts on productivity.</p><ul><li><p>What did they find? If chatbots were making individuals more productive, we might expect to see an increase in their wages and/or a reduction in work hours. However, the survey finds zero statistically significant effects on these variables or on firms&#8217; profits.</p></li><li><p>At first glance, this is disappointing. For many economists, growth in productivity is the most important determinant of long-term economic growth and all that rests on it. For the past two decades, productivity growth has been <a href="https://www.worldbank.org/en/research/publication/global-productivity">stagnant in much of the world</a>, rich and poor alike, hurting public services and living standards. Optimists hope that AI will now super-charge it, while pessimists worry about a repeat of the &#8216;<a href="https://en.wikipedia.org/wiki/Robert_Solow">Solow paradox</a>&#8217; - in 1987, the Nobel Laureate famously quipped that &#8220;<em>you can see the computer era everywhere but in the productivity statistics</em>&#8221;.</p></li><li><p>The Danish analysis does hint at some productivity gains from AI. Survey respondents who used chatbots reported an average saving of 2.8% in their work hours, but only a very small fraction of this resulted in higher wages. When added to the growing literature on AI&#8217;s productivity effects, this suggests a pattern where: <a href="https://academic.oup.com/qje/article/140/2/889/7990658">academic experiments</a> that give individuals access to chatbots or AI tools to complete a certain task often demonstrate quite high self-reported productivity effects, of 15-<a href="https://www.sciencedirect.com/science/article/pii/S0167268124004591">50</a>%, in a short time frame (e.g. a week). Yet once we turn to early real-world outcomes from AI use at organisations, these effects shrink. And when we look at aggregate productivity growth at the level of the entire economy, evidence for AI&#8217;s benefits is even more scant.</p></li><li><p>What might explain this? First, the Danish survey focuses on chatbots from 2023-24, and so the AI capabilities may have been too nascent to have had much effect. Second, the <em><a href="https://www.nber.org/system/files/working_papers/w25148/w25148.pdf">J curve </a></em><a href="https://www.nber.org/system/files/working_papers/w25148/w25148.pdf">theory</a> put forward by Erik Brynjolfsson et. al. argues that AI can increase productivity growth, but only after employees and organisations work out how to best use it, which takes time and resources. Humlum and Vestergaard find some evidence for this - employees report that introducing chatbots creates new tasks for employees who have to integrate them into workflows and ensure compliance. Finally, wage growth may also be a limited way to measure productivity growth, not least since effects on wages take time to materialise. Past <a href="https://academic.oup.com/qje/article-abstract/135/2/645/5721266">research</a> also demonstrates that new technologies may benefit a relatively small number of firms and workers, and so will not always clearly manifest in the data.</p></li><li><p>This suggests that as AI capabilities improve, and organisations and individuals become better at using it, productivity growth could start to increase, perhaps rapidly. However, this is not guaranteed. For one, AI will need to be usefully deployed across<em> all</em> or most consequential sectors, and catalyse new sectors, if countries are to avoid a &#8216;<a href="https://medium.com/@arnoldkling/what-gets-expensive-and-why-33bf4b891be2">Baumol cost disease&#8217;</a> scenario where new productivity gains in some sectors, such as the technology sector, lead to increased demand in other sectors, such as education, where productivity gains are not materialising. Such a scenario, which likely played a role in <a href="https://www.sciencedirect.com/science/article/pii/S0954349X22001394">the original Solow Paradox</a>, could blunt macro-level productivity growth. </p></li><li><p>AI may also introduce complexities that make it harder to measure productivity growth - for example, if people start using &#8216;free&#8217; or low-cost LLMs to perform tasks where they previously hired a company, this could cause output to nominally &#8216;decline&#8217;, at least under current measurement approaches. So not only do the potential effects of AI on productivity growth remain unclear, so does the best approach to measuring it.</p></li></ul></li><li><p><strong>View from the field: <a href="https://www.governance.ai/team/sam-manning">Sam Manning, GovAI</a>:</strong></p><ul><li><p><em><strong>&#8220;</strong>Outside the headline result, this paper includes several notable findings. For example, on days when they use AI, marketing professionals and software developers report higher time savings than teachers (~7-11% vs ~4.5%). These numbers don&#8217;t strike me as negligible and suggest that AI&#8217;s productivity effects are likely to vary quite a bit across occupations. If some roles are already saving 7&#8211;11% of their workday thanks to AI, firms will eventually begin adjusting workflows to better capture those time savings and competitive pressures will result in broader productivity gains that are measurable at the firm level. It&#8217;s also interesting that when employers actively encourage the use of chatbots, the reported effects on time savings, work quality, task expansion, creativity, and job satisfaction rise by 10&#8211;40%. That points to an important role for firms in promoting more widespread and effective use of LLMs in the workplace.&#8221;</em></p></li></ul></li></ul><h1>Topic deepdive</h1><h2>Will AI exacerbate or mitigate climate change?</h2><ul><li><p><strong>What happened: </strong>In April and May, <a href="https://www.economist.com/leaders/2025/04/10/how-ai-could-help-the-climate">The Economist</a> and <a href="https://www.technologyreview.com/2025/05/20/1116327/ai-energy-usage-climate-footprint-big-tech/">MIT Tech Review</a> published special reports examining how AI may affect climate change, with The Economist&#8217;s Alex Hern <a href="https://www.linkedin.com/posts/activity-7316415215242272768-nskI/?utm_source=share&amp;utm_medium=member_ios&amp;rcm=ACoAAAFKUQwBWnPeHDQoWtF6yYTxDJGBeu8oKoY">noting</a> that he has been trying to &#8220;nail down&#8221; this question ever since AI took off.</p></li><li><p><strong>What&#8217;s interesting: </strong>To understand how AI will affect the climate, we need to answer two different questions, which the reports shed light on, in different ways.</p></li><li><p>The first question is: how will training and inferencing AI models <em>directly</em> affect greenhouse gas emissions, via the power they consume and the &#8216;embodied&#8217; emissions required to build, maintain and recycle the data centres, devices and networks?</p><ul><li><p>Researchers like <a href="https://scholar.google.com/citations?hl=en&amp;user=UCDMtM0AAAAJ&amp;view_op=list_works&amp;sortby=pubdate">Emma Strubell</a>, <a href="https://scholar.google.co.uk/citations?hl=fr&amp;user=nP8cwkIAAAAJ&amp;view_op=list_works&amp;sortby=pubdate">Alexandra Sasha Luccioni</a> and <a href="https://arxiv.org/abs/2505.06371">Jae-Won Chung </a>have devised <a href="https://codecarbon.io/">methods</a> to help address these questions. The MIT report draws on these methods to provide new estimates. For example, they find that asking Llama 3.1 8b to make a travel itinerary requires ~114 joules of energy, once cooling and other factors are accounted for - a tiny amount, equivalent to riding six feet on an e-bike. At the other end of the spectrum, generating a five second video using a ZhiPuAI model, uses about 3.4 <em>million</em> joules, equivalent to ~38 miles on an e-bike.</p></li><li><p>It is challenging to both compile and interpret these estimates. First, researchers typically have to focus on open-source models, as they argue that companies who develop leading proprietary models do not release the necessary data, although the EU AI Act may soon require estimates for the largest AI models. A second issue is that some past estimates have been <a href="https://x.com/fiiiiiist/status/1836471413198459331">miscalculated</a>, or misleadingly reported, in a way that makes them sound a lot larger - <a href="https://www.sciencedirect.com/science/article/pii/S2542435121002117">echoing past panics about energy use from technology</a>, such as <a href="https://www.iea.org/commentaries/the-carbon-footprint-of-streaming-video-fact-checking-the-headlines">around video streaming during Covid-19</a>. Finally, there is no clear way to tally these individual estimates into an <em>aggregate </em>estimate for the overall emissions from training and running <em>all </em>AI models to understand how relevant it is, from a global emissions perspective.</p></li><li><p>Instead, the best macro estimates come from looking at data centres&#8217; power consumption. Today, data centres <a href="https://www.sustainabilitybynumbers.com/p/ai-energy-demand">account</a> for just ~1.5% of global power consumption, or ~2% if crypto is included. Most of this comes from activities like streaming, rather than AI. In April, the IEA shared its latest forecast for how this may change in the AI era. In its base case scenario, data centres&#8217; power consumption would rise to 945 terawatt-hours by 2030, up from 415 in 2024. If this proves accurate, it would be a non-trivial increase and could put pressure on energy grids in certain locations, as data centres are geographically concentrated - in Ireland they <a href="https://www.iea.org/reports/electricity-2024">accounted for ~17% of power consumed</a> in 2022. However, data centres would still account for just ~3% of global power consumption and the increase in their power consumption would be smaller than that of other sectors, such as electric vehicles and air-conditioning.</p></li><li><p>From a climate change perspective, the precise amount of power that AI consumes will also be less important than <em>the source</em> of that power. Optimists hope that AI acts as a forcing function to dramatically accelerate the roll-out of nuclear and renewable energy in the near-term. Skeptics worry that the AI race will compel companies to use fossil fuels that they may have otherwise eschewed.</p></li></ul></li><li><p>When it comes to determining how AI will affect climate change, a second question is arguably more important than AI&#8217;s future power use: what applications will AI be used for, at what scale, and how will these applications affect emissions?</p><ul><li><p>AI supporters argue that it will accelerate renewable energy and make the  economy - including energy-intensive sectors - dramatically more efficient.</p></li><li><p>The Economist&#8217;s report provides some grounds for optimism on this front. They note how companies such as <a href="https://octopus.energy/">Octopus Energy</a> and <a href="https://x.company/projects/tapestry/">Tapestry</a> are using AI to make it easier to deploy renewable energy and to optimise the grid, for example by helping to locate green-energy projects and enable <a href="https://octopus.energy/blog/100k-zero-bills-homes-by-2030/">smart homes</a> and vehicles to autonomously draw power during fallow periods. Other case studies document how energy-intensive industries are using AI, for example to optimise heating and cooling in buildings, reduce waiting times for ships in ports, or to enable new kinds of steel manufacturing processes.</p></li><li><p>Estimating how these AI use cases may affect emissions is even more challenging than estimating AI&#8217;s power use. There is no formal stocktaking of beneficial AI climate applications and few efforts to estimate how they will affect emissions or the <em>additionality</em> that AI brings. In theory, these AI applications could reduce the future emissions that would have otherwise occurred by much more than AI&#8217;s future power use increases them, but the level of uncertainty, and the timeline to impact, are greater. This is even more true when we consider efforts to use AI in science, which could lead to new materials for solar panels, batteries, and direct air capture, or even accelerate fusion - an effectively limitless, clean energy. Could AI make breakthroughs in these areas 30% more likely, or accelerate them by 30 years? Given their complexity, the temptation to skip such questions and focus on what we can measure - AI&#8217;s power use - is high.</p></li><li><p>A final complication comes from the fact that most AI applications are not obviously good, or bad, from an emissions perspective but may still shift individual or organisational behaviour in consequential ways. For example, what might happen if people start to shift more of their economic activities to AI agents? History tells us that the impact of technologies often depends on context and the activities they replace. For example, the Internet enabled music streaming, ecommerce, and home working, but whether these shifts in behaviour increase or reduce emissions can vary depending on the individual case, such as the size of a person&#8217;s home or whether ecommerce transport is electrified. At the aggregate level, there are reasons to think that digitisation helps to make economies less carbon intensive. But it&#8217;s hard to reliably &#8216;prove&#8217; this and much depends on context and efficiency gains - which so far have been remarkably high for <a href="https://www.science.org/doi/10.1126/science.aba3758">data centres</a> and <a href="https://x.com/karpathy/status/1811467135279104217">AI</a>.</p></li></ul></li><li><p>Given these complexities, how should policymakers ensure that AI benefits the climate? The Economist argues that the best policy would be a strong global carbon tax to enable the market to incentivise and penalise different AI applications and uses. However, it views this as politically intractable and so calls on governments to instead undertake permitting reforms to allow AI companies to fund and build more clean energy, and to build more flexible data centres that can match workloads to intermittent wind and solar.</p></li><li><p>The Economist also calls on other geographies to emulate the EU AI Act and impose obligations on AI developers to share estimates for the power used by their leading models. The reliance on open-source models to estimate AI power use does seem inadequate but the usefulness of this recommendation could also be challenged, given the <a href="https://energy.ec.europa.eu/topics/energy-efficiency/energy-efficiency-targets-directive-and-rules/energy-efficiency-directive_en">overlapping energy reporting requirements</a> that already exist and the risk of creating the kind of arduous &#8216;environmental impact assessments&#8217; seen in other sectors that can stymie innovation at little benefit to the environment.</p></li><li><p><strong>View from the field: <a href="https://x.com/AndyMasley">Andy Masley,</a> Author of <a href="https://andymasley.substack.com/">Weird Turn Pro</a> and <a href="https://substack.com/@andymasley/p-162196004">Why ChatGPT is not bad for the environment</a></strong></p><ul><li><p><em><strong>&#8220;Excited: </strong>AI seems ecologically costly if we only look at its total energy use without considering the value we get out of it. But, hospitals use more energy than yachts, and if we look at value per unit of energy, AI seems very likely to become one of the most energy efficient sectors. See for example <a href="https://fly.io/blog/youre-all-nuts/">this</a>,<a href="https://steveklabnik.com/writing/i-am-disappointed-in-the-ai-discourse/"> this</a>, and<a href="https://simonwillison.net/2025/Mar/11/using-llms-for-code/"> this</a> about how LLMs are adding value to programming, at relatively little cost. And that&#8217;s before we consider the more direct ways that AI can be useful to the climate, such as by optimising energy and transportation.</em></p></li><li><p><em><strong>Worried: </strong>In line with <a href="https://en.wikipedia.org/wiki/Jevons_paradox">Jevon&#8217;s Paradox</a>, I worry that even though AI might make processes more efficient, if it&#8217;s not paired with a switch to renewable energy, we may emit more in total and miss key climate targets. I'm also concerned that AI-enabled weapons or widespread job automation could threaten political stability, eroding the trust needed for international climate cooperation.&#8221; </em></p></li></ul></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives. Please share your take &amp; subscribe for free to receive new posts. Lots more in the pipeline! </p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>We are exploring ways to make this Substack more useful. If you have ideas, please reach out to <a href="mailto:aipolicyperspectives@google.com">aipolicyperspectives@google.com</a>. </em></p>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#19)]]></title><description><![CDATA[Jobs, scientific reliability, and the free society]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-issue-17</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-issue-17</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Thu, 17 Apr 2025 14:00:36 GMT</pubDate><enclosure url="https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Every month, our AI Policy Primer looks at three external developments that caught our eye. Today, we look at new AI tools for detecting errors in scientific papers; an exploration into whether AGI might upset the delicate balance underpinning liberal societies; and at a study assessing how AI is affecting employment in the US. Please leave a comment to let us know your thoughts. </em></p><p><em>We are exploring ways to make this newsletter more useful - if you have ideas, please reach out to <a href="mailto:aipolicyperspectives@google.com">aipolicyperspectives@google.com</a>. Thanks for reading!</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw"><img src="https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" width="3000" height="1688" data-attrs="{&quot;src&quot;:&quot;https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1688,&quot;width&quot;:3000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;a close up of a blue and green structure&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="a close up of a blue and green structure" title="a close up of a blue and green structure" srcset="https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives. Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Will AI help or hurt scientific reliability?</h2><ul><li><p><strong>What happened?: </strong>An <a href="https://www.nature.com/articles/d41586-025-00648-5?utm_source=Live+Audience&amp;utm_campaign=30e726b684-nature-briefing-daily-20250307&amp;utm_medium=email&amp;utm_term=0_b27a691814-30e726b684-50458408#correction-0">article</a> in Nature News highlighted two early efforts to use LLMs to detect errors in scientific papers. If successful, they could provide a much-needed boost to scientific reliability, but many scientists remain skeptical about their usefulness.</p></li><li><p><strong>What&#8217;s interesting?: </strong>&#8216;Reliability&#8217; refers to scientists&#8217; ability to depend upon each other&#8217;s findings and trust that they are not due to chance or error. A series of interrelated <a href="https://www.sciencefictions.org/p/book?open=false#%C2%A7corrections">challenges</a> currently undermine scientific reliability, including the p-hacking and <a href="https://en.wikipedia.org/wiki/Publication_bias">publication bias </a>that lead <a href="https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0010068">researchers to underreport negative results</a>; <a href="https://www.nature.com/articles/d41586-024-02280-1">a lack of standardisation</a> in how scientists carry out routine scientific tasks; <a href="https://worksinprogress.co/issue/real-peer-review/">challenges with the peer review process</a>, and <a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2685008/">scientific fraud</a>. Another issue is that scientists can make mistakes, for example in <a href="https://mbnuijten.com/wp-content/uploads/2013/08/nuijtenetal_2016_reportingerrorspsychology.pdf">how they apply statistical methods</a>. At the aggregate level, such mistakes are non-trivial - a 2013 study claimed that<a href="https://mbnuijten.com/wp-content/uploads/2013/08/nuijtenetal_2016_reportingerrorspsychology.pdf"> 13% of psychology papers</a> include a mistake that, if corrected, would alter the interpretation of their results.</p></li><li><p>Some scientists worry that the growing use of AI in research will further undermine scientific reliability, not least because AI models are prone to &#8216;hallucinate&#8217; outputs, including scientific citations. In response, AI practitioners are working on mitigations to these risks, such as techniques to better <a href="https://blog.google/technology/ai/google-d">ground model outputs to trusted sources</a>.</p></li><li><p>Other practitioners, including those behind two new AI-based error detection efforts, hope to go further and use AI to improve the reliability of the wider research base. The first effort is the <a href="https://the-black-spatula-project.github.io/">&#8216;Black Spatula Project</a>,&#8217; which was named after scientists used AI to detect a mathematical error in a <a href="https://www.sciencedirect.com/science/article/abs/pii/S0045653524022173">widely-covered study </a>that had incorrectly claimed that black plastic cooking utensils contained worrying levels of cancer-linked flame retardants. Staffed by volunteers, the open-source project has so far used AI to review ~500 papers. It has not yet made the errors that it has found public and is instead sharing them with the papers&#8217; authors. The second effort,<a href="https://yesnoerror.com/"> YesNoError</a>, uses an AI agent to scan papers for errors and aspires to check the entire scientific literature.</p></li><li><p>As the article notes, some scientific integrity practitioners cautiously support the efforts, but not everybody is a fan. The researcher Nick Brown claims the false positive rate is high and that many of the &#8216;errors&#8217; are minor typos or writing issues. The practitioners behind YesNoError also aim to work with the <a href="https://www.researchhub.com/">ResearchHub</a> platform - which pays researchers cryptocurrency to do peer review. They want to let holders of the cryptocurrency suggest which papers get scrutinized first, which some worry could lead to people targeting research they don&#8217;t like.</p></li><li><p>Such skepticism is also evident in the EU&#8217;s <a href="https://research-and-innovation.ec.europa.eu/document/download/2b6cf7e5-36ac-41cb-aab5-0d32050143dc_en?filename=ec_rtd_ai-guidelines.pdf">recently-updated guidelines</a> for researchers and funders on how to use AI in research, which focus almost exclusively on the risks that AI poses, and the responsibility of researchers and funders to mitigate them. It is also visible in the bans that many journals and conferences have imposed on the use of AI in peer review, even if many individual peer reviewers <a href="https://www.nature.com/articles/d41586-025-00894-7">appear to be using it</a>.</p></li><li><p><strong>What&#8217;s the takeaway?: </strong>Many of the concerns stem from a desire for AI not to replace the role of human reviewers, particularly for consequential decisions, such as approving publication decisions. However, if fast-improving AI reasoning models were instead framed as aids to human researchers or peer reviewers, to sense-check or enhance their own review processes, particularly for error detection, they may become more popular. To do so, independent evaluations that can reliably demonstrate the ability of AI to detect consequential errors that humans overlook, will likely be critical.</p></li></ul><h2>AGI and the free society</h2><ul><li><p><strong>What Happened?: </strong>A new <a href="https://arxiv.org/abs/2503.05710">working paper</a> from Justin B. Bullock, Samuel Hammond, and S&#233;b Krier explores how AGI might affect the balance in power between state and society, upsetting the delicate equilibrium that underpins liberal democracies.</p></li><li><p><strong>What&#8217;s Interesting?: </strong>The paper builds on a <a href="https://www.amazon.fr/Narrow-Corridor-States-Societies-Liberty/dp/0735224382">framework</a> developed by Daron Acemoglu and James A. Robinson, which argues that true liberty exists in a &#8216;narrow corridor&#8217; between an overly-powerful, despotic state on one hand, and a chaotic absent state that is too weak to govern on the other.</p></li><li><p>AGI may upset this narrow corridor if it enables states or non-state actors to engage in new kinds of harmful monitoring, coordination, or decision-making. For example, governments are using AI to detect tax fraud and manage traffic flows. But AGI could go much further, potentially automating entire public sector roles. Positively, this could help governments to better understand and shape behaviours across society, similar to how digitisation has helped to visualise and suppress black markets in <a href="http://india.it">India.</a></p></li><li><p>It could also promote more narrowly tailored rule enforcement and curb blanket policies that produce inefficiencies - for instance, the type of policies that led to the recall of Tesla&#8217;s Full Self&#8209;Driving for carrying out harmless <a href="https://www.progressive.com/lifelanes/what-is-a-rolling-stop/">rolling stops</a> could be replaced with AGI systems that better weigh real&#8209;time context, such as the presence of pedestrians, visibility, and thus the actual level of risks.</p></li><li><p>However, this improved legibility and enforcement also brings risks of illegitimate surveillance and loss of liberty, such as AI systems that might ensure that every vehicle perfectly obeys every traffic rule, or AI-enabled CCTV that might detect and punish every minor infraction, undermining the useful personal judgement and empathy that today&#8217;s officials often show in such situations.</p></li><li><p>AGI could also empower <em>non-state</em> actors. In more positive scenarios, this could enable citizens to better understand and advocate for policy positions, fact-check officials, and usher in new kinds of public deliberation. However, individuals/groups could also use AI agents to orchestrate harmful actions or create opaque financial communication methods that make the economy <em>less </em>legible to governments, rather than more, similar to how cryptocurrency can be used to launder money, despite its legibility.</p></li></ul><ul><li><p><strong>What&#8217;s the takeaway?: </strong>To stay liberal and democratic, the authors call for investment in robust technical safeguards - like privacy-preserving AI - and intentional policies - like identity verification protocols and advanced encryption. In short, states must neither blindly hand off power to AI systems nor clamp down on them in ways that stifle innovation.</p></li></ul><h2>Grappling with the economic impact of AI</h2><ul><li><p><strong>What happened?: </strong>A <a href="https://www.nber.org/system/files/working_papers/w33509/w33509.pdf">new paper</a> from Yale&#8217;s <a href="http://menakahampole.com/">Menaka Hampole</a> and colleagues found that jobs that are highly-exposed to AI are experiencing lower labor market demand, compared to less exposed occupations. However, this is partly offset by boosts to productivity and profits in firms that adopt AI, which increases their ability to hire.</p></li><li><p><strong>What&#8217;s interesting?: </strong>Measuring and predicting how fast-improving AI systems will affect employment and other economic variables, like growth or inequality, is a mounting <a href="https://www.governance.ai/analysis/predicting-ais-impact-on-work">challenge</a>. Economists are pursuing different approaches - <a href="https://www.governance.ai/analysis/predicting-ais-impact-on-work">none of them perfect</a>.</p></li><li><p>In this study, the authors treat occupations as bundles of tasks. They posit that AI&#8217;s effects on demand for a particular occupation will depend on how many tasks within that occupation AI can substitute for, and how many other tasks remain, including new tasks that an employee can pivot into - such as an automated expense system that enables an accountant to pursue more complex financial analysis.</p></li><li><p>To measure exposure to AI, the researchers reviewed a large volume of online job postings from US publicly-traded companies between 2010-2023. They used LLMs to identify specific AI applications described within these postings, such as &#8220;analysing customer reviews&#8221; or &#8220;forecasting risk and fraud&#8221;. They then matched these AI applications to the tasks that humans perform, by drawing on the US government&#8217;s <a href="https://www.onetonline.org/">O*Net database</a>. They also drew on firm-level data on sales, profits, productivity, and hiring to assess how AI adoption affects them.</p></li><li><p>The authors find that some broad occupation groups that were highly exposed to AI, like 'Business and Financial' and 'Architecture and Engineering,' experienced the largest AI-related declines in their relative employment shares during the study period, estimated at between 2%-2.5%. Examples of highly-exposed occupations in these categories include market researchers, credit analysts, and financial specialists. However, the paper finds that, on average, AI adoption increases sales, profit, productivity growth and <em>overall </em>hiring in firms.</p></li><li><p>The authors also found that more highly-paid roles tend to be more exposed to AI, but that this tails off above the 90th percentile level, perhaps because the most highly-paid jobs require strong interpersonal and management skills, which AI cannot currently automate.</p></li><li><p>As the authors acknowledge, their approach faces various challenges and uncertainties;</p><ul><li><p><strong>Bias: </strong>The firms that embrace AI may have other characteristics that contribute to their higher growth trajectories - a potential source of bias that the authors try to address using an <em>instrumental variable </em>approach, which they acknowledge is imperfect - see more on this mechanism <a href="https://www.broadstreet.blog/p/a-good-instrument-is-hard-to-find">here</a>. The reliance on an online jobs dataset may also lead to biases, including in the types of AI use that it includes/excludes - e.g. organizations may not necessarily describe AI applications that are more likely to displace employees in their job postings.</p></li><li><p><strong>Time period: </strong>The authors&#8217; analysis ends in 2023, which means that it overlooks more recent GenAI tools, whose adoption is still nascent. There is also a question about how representative the trends will be of longer-term effects as AI improves, diffuses across the economy, and individuals and firms respond to its impacts. Previous <a href="https://www.sciencedirect.com/science/article/abs/pii/S0169721811024105">technological shocks</a> during the two industrial revolutions and the computerization of the late 20th century often led to an <a href="https://economics.mit.edu/sites/default/files/inline-files/Why%20Are%20there%20Still%20So%20Many%20Jobs_0.pdf">initial boost </a>to employment in technology-exposed occupations, followed by eventual displacement. They also led to new jobs, but many of these jobs were either not available to those who were displaced, or were less satisfying, leading to rising labour market polarization and inequality.</p></li></ul></li><li><p><strong>What&#8217;s the takeaway?: </strong>The effects that AI will have on aggregate (un)employment, across different time periods, and the degree to which different employees benefit/suffer, remain open questions. But it seems plausible that, over the next two years, in high-income countries, AI could have a moderate positive impact on productivity and economic growth. There may be no major increase in aggregate unemployment, yet, but we will likely see early signs of increased inequality between those employees who are able to benefit from AI and those who cannot.</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives. Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#18) ]]></title><description><![CDATA[European AI investments; a market for AI safety; and autonomous vehicles]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-february-2025</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-february-2025</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Tue, 04 Mar 2025 08:01:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Nedd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Every month, our AI Policy Primer looks at three external developments from the world of AI policy that caught our eye. In this edition, we spotlight recent French and European investments in AI, a study exploring the concept of an AI safety marketplace, and recent developments in autonomous vehicle deployment. </em></p><p><em>As always, please leave a comment below to let us know your thoughts, or send any feedback to <a href="mailto:aipolicyperspectives@google.com">aipolicyperspectives@google.com</a>. Thanks for reading!</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nedd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nedd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp 424w, https://substackcdn.com/image/fetch/$s_!Nedd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp 848w, https://substackcdn.com/image/fetch/$s_!Nedd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp 1272w, https://substackcdn.com/image/fetch/$s_!Nedd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nedd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp" width="800" height="503" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:503,&quot;width&quot;:800,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:59886,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.aipolicyperspectives.com/i/158324269?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3adc855-5f59-4ab5-a0e9-1a21356c99ca_800x503.webp&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Nedd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp 424w, https://substackcdn.com/image/fetch/$s_!Nedd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp 848w, https://substackcdn.com/image/fetch/$s_!Nedd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp 1272w, https://substackcdn.com/image/fetch/$s_!Nedd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F768dfea1-da41-4bb1-bba3-a84717ba4e67_800x503.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><h1><strong>Policymakers taking action</strong></h1><h2>France and the EU announce large AI investments</h2><ul><li><p><strong>What happened: </strong>At the recent AI Action Summit in Paris, President Emmanuel Macron unveiled &#8364;109bn in private sector funding for AI infrastructure. This <a href="https://humangeneralintelligence.substack.com/p/france-and-the-eu-announced-big-ai?utm_source=post-banner&amp;utm_medium=web&amp;utm_campaign=posts-open-in-app&amp;triedRedirect=true">includes</a> a new &#8364;50bn AI campus of data centres, led by the UAE&#8217;s <a href="https://www.mgx.ae/en">MGX</a>, which is also involved in the US &#8216;Stargate&#8217; project; a &#8364;10bn AI &#8220;supercomputer&#8221; from <a href="https://www.fluidstack.io/">Fluidstack</a>, a British AI Cloud platform; and &#8364;5bn from the US firm Apollo to invest in AI energy infrastructure.</p></li><li><p>The European Commission also unveiled a &#8364;200bn <a href="https://ec.europa.eu/commission/presscorner/detail/en/ip_25_467">&#8220;InvestAI&#8221; initiative</a>, which includes plans to create four &#8220;AI gigafactories&#8221; with &#8220;100,000 last-generation AI chips&#8221;, complementing the smaller &#8216;<a href="https://digital-strategy.ec.europa.eu/en/policies/ai-factories">AI factories</a>&#8217; the EU is already developing.</p></li><li><p><strong>What&#8217;s interesting</strong>: One narrative that emerged from the Paris Summit, following <a href="https://www.presidency.ucsb.edu/documents/remarks-the-vice-president-the-artificial-intelligence-action-summit-paris-france">a speech</a> by US Vice President JD Vance, was that the EU&#8217;s AI efforts are mired in excessive regulation, while the US is powering ahead with a more supportive regulatory environment and strong financing. The reality is more nuanced. While the US has not yet passed any federal AI regulation akin to the EU&#8217;s AI Act, <a href="https://substack.com/@deanwball/p-157005561">many US states are advancing AI bills that could impose similar obligations</a> - if passed. Unlike the EU&#8217;s harmonised approach, some of these state-level bills also take inconsistent approaches, potentially increasing regulatory complexity.</p></li><li><p>The new French and EU funding announcements, alongside Macron&#8217;s repeated promotion of Mistral at the Summit, also underscored that EU member states do want future frontier AI models to be trained in Europe. Commission President Ursula von der Leyen also pledged that the new gigafactories would prioritise these efforts. While the EU is advancing the AI Act, the Commission also made the rare decision, shortly after the Summit, <a href="https://www.euractiv.com/section/tech/news/commission-withdraws-ai-liability-directive-after-vance-attack-on-regulation/">to withdraw its proposed AI liability directive</a> - as part of <a href="https://commission.europa.eu/news/commission-proposes-cut-red-tape-and-simplify-business-environment-2025-02-26_en">broader efforts to streamline the EU&#8217;s regulations</a> and boost competitiveness.</p></li><li><p>Still, many remain sceptical about whether France and the EU can rapidly secure and deploy the new AI funding, much of which remains to be mobilised. There are also doubts about whether they can narrow the wider gap with the US AI ecosystem, particularly when it comes to training the most advanced AI foundation models. With these challenges in mind, von der Leyen <a href="https://ec.europa.eu/commission/presscorner/detail/en/ip_25_467">emphasised</a> the need to prioritise &#8216;industry-specific&#8217; AI applications. This could include sectors like green energy, where the EU has strong expertise but faces intense competition from China and others.</p></li><li><p>This suggests that, while a small number of EU AI startups will continue developing frontier foundation models, the most promising efforts may emerge in areas that draw on local economic strengths, such as in finance, tourism, or healthcare.</p></li></ul><blockquote></blockquote><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.aipolicyperspectives.com/subscribe?"><span>Subscribe now</span></a></p><h1><strong>Study watch</strong></h1><h2>Building a market for AI safety</h2><ul><li><p><strong>What happened: </strong>Philip Moreira Tomei and colleagues at the <a href="https://ai.objectives.institute/">AI Objectives Institute</a> published a <a href="https://arxiv.org/pdf/2501.17755">paper</a> arguing that market-based mechanisms could help reduce AI safety risks, by complementing regulatory efforts.</p></li><li><p><strong>What&#8217;s interesting: </strong>Discussions about AI risks often quickly pivot to how to pass or adapt new regulations. However, uncertainty about specific AI risk scenarios makes it difficult to craft rules that are targeted and effective. When regulation does arrive, it is often vague, hard to implement, and can create uncertainty that inhibits wider AI adoption. The rapid pace of AI development also makes it difficult to design regulation that is resilient and adaptable.</p></li><li><p>The authors argue that market-based mechanisms could help to complement AI regulation by providing AI developers and deployers with financial incentives to identify, evaluate, and mitigate AI risks, while distributing risk management across a broader range of actors. The paper outlines four market-based approaches, citing examples from other high-risk industries:</p><ul><li><p><strong>Insurance: </strong>Firms could take out liability insurance against AI risks, for example building on existing <a href="https://www.abi.org.uk/products-and-issues/choosing-the-right-insurance/cyber-insurance/">cybersecurity insurance</a> or <a href="https://getindemnity.co.uk/business-insurance/professional/what-is-technology-errors-and-omissions-insurance">technology errors and omissions insurance</a>, which have encouraged firms to invest in risk mitigation.</p></li><li><p><strong>Auditing &amp; certification:</strong> Firms could hire third-party auditors to assess AI safety practices, leading to certifications for meeting certain standards. For example, after facing scrutiny over their cybersecurity, Zoom engaged <a href="https://www.trailofbits.com/">Trail of Bits</a> and<a href="https://www.nccgroup.com/us/"> NCC Group</a> for an audit, which led to enhanced end-to-end encryption.</p></li><li><p><strong>Procurement:</strong> Large purchasers of AI could demand performance on safety benchmarks or require specific disclosures - similar to how governments <a href="https://www.nber.org/papers/w32831">use procurement</a> to shape markets or how corporations push suppliers to improve working conditions and environmental standards.</p></li><li><p><strong>Investor due diligence:</strong> Investors could also demand safety and transparency measures from AI companies, similar to how investors pressured BP to share more about their risk management processes, following the 2010 Deepwater Horizon oil spill, which also accelerated BP&#8217;s transition to renewable energy.</p></li></ul></li><li><p>Similar ideas have been explored by other organisations and sectors in the past. A related promising angle is supply-side interventions: governments or philanthropic organisations could act as '<a href="https://worksinprogress.co/issue/buyers-of-first-resort/">buyers of first resort</a>' by establishing <a href="https://worksinprogress.co/issue/how-to-start-an-advance-market-commitment/">Advance Market Commitments</a> that guarantee future purchases of innovative products, like AI safety tools, to incentivise their development. <a href="https://arxiv.org/pdf/2304.04914">Regulatory markets</a>' - where governments license private regulators to compete to provide AI safety oversight services to companies - could also address gaps left by traditional regulatory mechanisms.</p></li><li><p>Although they hold promise, market-based mechanisms also face a range of challenges, including how to encourage bottom-up action from a diverse range of organisations (from insurance providers to investors); how to prioritise and price different AI risks; how to ensure sufficient independence and skills among auditors and certification agencies; and how to balance the goal of AI safety against others. For example, after BP&#8217;s strategic redirection towards renewables, its <a href="https://www.economist.com/business/2025/02/11/bp-is-underperforming-and-under-pressure">financial performance slumped</a>, and the company recently <a href="https://www.bbc.co.uk/news/articles/c3374ekd11po">reversed course</a>, saying it had gone "too far, too fast" in the transition away from fossil fuels.</p></li></ul><h1><strong>Sector deep dive</strong></h1><h2>The deployment of autonomous vehicles slowly accelerates</h2><ul><li><p><strong>What happened: </strong>In January,<strong> </strong><a href="https://kodiak.ai/">Kodiak Robotics</a> announced that its client <a href="https://kodiak.ai/news/kodiak-delivers-customer-owned-autonomous-robotrucks-to-atlas">Atlas Energy Solutions</a> - which serves oil and gas companies in the Permian Basin (West Texas and New Mexico) - had successfully delivered 100 loads of material using driverless trucks.</p></li><li><p><strong>What&#8217;s interesting: </strong>Excitement and skepticism about autonomous vehicles has fluctuated over the past decade, but parts of the sector are now seeing renewed momentum. For example, Waymo <a href="https://techcrunch.com/2025/02/27/waymo-has-doubled-its-weekly-robotaxi-rides-in-less-than-a-year/">now logs </a>200,000 paid robotaxi rides every week, a 20x increase in two years, and will soon begin <a href="https://waymo.com/blog/2024/12/partnering-with-nihon-kotsu-and-go-on-our-first-international-road-trip">testing in Japan,</a> following a recent $5.6B funding round.</p></li><li><p>The stop-start development of autonomous vehicles highlights a classic challenge in technology development: the mismatch between a technology&#8217;s <em>capabilities</em> and its practical <em>deployment</em>. Widespread adoption of autonomous vehicles has faced multiple barriers, including: a <a href="https://www.sciencedirect.com/science/article/pii/S2590198224000198">complex and evolving</a> regulatory landscape that varies by country and (in the US) by state; the far higher safety standards <a href="https://waymo.com/blog/2024/12/new-swiss-re-study-waymo">expected of autonomous vehicles compared to human drivers</a>; low levels of <a href="https://yougov.co.uk/travel/articles/35562-car-manufacturers-still-some-way-convincing-brits-">public trust</a>; and the sheer complexity of real-world roads, particularly in dense urban centres.</p></li><li><p>In response, some companies, like Kodiak, are prioritising industry-specific use cases in more controlled, remote environments, such as mines, seaports, large industrial farms, and military domains. These settings can also expose autonomous vehicles to harsh conditions - like dust, uneven terrain, and strict local site regulations - which could help improve the technology.</p></li><li><p>These deployment decisions are also influenced by labour market trends. While concerns persist about AI replacing drivers and supply chain workers, a shortage of personnel is arguably a greater challenge. In the US alone, there are about <a href="https://www.trucking.org/news-insights/ata-american-trucking-trends-2024#:~:text=Preliminary%20figures%20indicate%20that%20the,million%20professional%20drivers%20in%202023.">3.5m truckers</a>, but companies <a href="https://www.mckinsey.com/industries/automotive-and-assembly/our-insights/will-autonomy-usher-in-the-future-of-truck-freight-transportation">struggle with </a>an ageing workforce and high turnover rates, which exceed 90% in some sectors, partly owing to poor working conditions. This in turn is leading to supply chain disruptions and higher consumer costs.</p></li><li><p>A broader question is whether the deployment of autonomous vehicles over the past five years offers any insights into how other types of AI-enabled robots might be deployed across the economy in the coming years. Traditional industrial robots are <a href="https://ifr.org/ifr-press-releases/news/record-of-4-million-robots-working-in-factories-worldwide">already well established</a> in manufacturing and warehousing, but most are limited to a narrow range of repetitive tasks in structured environments. Companies are now using foundation models to develop more general-purpose robots that could learn novel tasks and adapt to real-world environments. If technical challenges can be overcome, these robotics could be particularly valuable in sectors like agriculture, healthcare or social care, where labour shortages are mounting. That said, they may face similar deployment obstacles to autonomous vehicles, which could lead them to seek out use cases where safety risks are lower, environments are easier to control and cost pressures are high.</p></li></ul>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#17)]]></title><description><![CDATA[Infrastructure, journalism, and critical thinking skills]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-january-2025</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-january-2025</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Fri, 31 Jan 2025 16:10:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!X1kA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Every month, our AI Policy Primer looks at three external developments from the world of AI policy that caught our eye. In this edition, we compare and contrast the recent UK and US AI infrastructure announcements, spotlight a study on how AI may affect critical thinking, and explore the future of AI-enabled journalism. Please leave a comment below to let us know your thoughts, or send any feedback to <a href="mailto:aipolicyperspectives@google.com">aipolicyperspectives@google.com</a>. Thanks for reading!</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!X1kA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!X1kA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp 424w, https://substackcdn.com/image/fetch/$s_!X1kA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp 848w, https://substackcdn.com/image/fetch/$s_!X1kA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp 1272w, https://substackcdn.com/image/fetch/$s_!X1kA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!X1kA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp" width="800" height="503" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:503,&quot;width&quot;:800,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:28310,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!X1kA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp 424w, https://substackcdn.com/image/fetch/$s_!X1kA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp 848w, https://substackcdn.com/image/fetch/$s_!X1kA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp 1272w, https://substackcdn.com/image/fetch/$s_!X1kA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2a368b6-1033-45fc-87d7-5d97a3aabc77_800x503.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h1><strong>Policymakers taking action</strong></h1><h2>US and UK government announce flagship AI infrastructure efforts</h2><ul><li><p><strong>What happened:</strong> The UK Government announced a new <a href="https://www.gov.uk/government/publications/ai-opportunities-action-plan/ai-opportunities-action-plan">AI Opportunities Action Plan</a>, while the Trump administration announced &#8216;Project Stargate&#8217; - a $500bn initiative to expand AI training infrastructure, led by Softbank, OpenAI, Oracle and others.</p></li><li><p><strong>What&#8217;s interesting: </strong>Both efforts aim to attract investment for AI infrastructure, but they also highlight the starkly different approaches on either side of the Atlantic. The UK&#8217;s Plan, praised for its ambition, includes 50 recommendations, ranging from deregulating data centre planning to promoting AI adoption in the public sector. The goal is to &#8216;on-shore&#8217; AI activity by leveraging the UK&#8217;s existing strengths - such as its strong university research base (Oxford, Cambridge, Imperial, etc) and a robust AI safety community - while addressing planning and energy constraints that have <a href="https://ukfoundations.co/">hindered broader economic growth</a> .</p><ul><li><p>The plan also signals a potential shift in UK economic policymaking. By establishing a &#8220;UK Sovereign AI unit&#8221; inside No.10, and &#8220;AI Growth Zones&#8221; with assumed planning approval for AI infrastructure and energy projects, the Plan explicitly states that &#8220;the invisible hand of the market&#8221; alone will not suffice. Instead, the UK must take a proactive role to remain competitive in AI.</p></li><li><p>The US, with lower energy costs, a stronger industrial base, and a more liberal approach to infrastructure, is in a fundamentally different position. Project Stargate is private-sector-led and operates on a scale that the UK and other nations cannot realistically match.</p></li><li><p>However, both initiatives face questions about funding and execution. For Stargate, the US government&#8217;s role remains unclear, raising concerns about bottlenecks in energy, land, and resource allocation. In the UK, much will hinge on the forthcoming Spending Review.</p></li><li><p>Ultimately, these announcements underscore how AI is becoming central to economic strategy and geopolitical influence. The &#8216;race for compute&#8217; is set to intensify, despite ongoing breakthroughs in AI efficiency, leading to more public-private partnerships for large-scale AI investment.</p></li><li><p>Whether a significant share of this spending happens outside of the US will largely depend on how effectively other nations, including the UK, execute on their AI strategies, but will also be impacted by policies like the new US <a href="https://www.rand.org/content/dam/rand/pubs/perspectives/PEA3700/PEA3776-1/RAND_PEA3776-1.pdf">AI Diffusion Framework</a>.</p></li></ul></li></ul><h1><strong>Study watch</strong></h1><h2>Will AI use hurt critical thinking skills?</h2><ul><li><p><strong>What Happened:<a href="https://www.mdpi.com/2075-4698/15/1/6"> </a></strong><a href="https://www.mdpi.com/2075-4698/15/1/6">A new study</a> of 666 UK participants, spanning diverse age groups and educational backgrounds, found a strong self-reported negative correlation between AI tool use and self-reported critical thinking skills.</p></li><li><p><strong>What&#8217;s Interesting: </strong>Over the past 20 years, the advent of the Internet has led educators to focus on equipping students with &#8216;hard&#8217; STEM skills, alongside &#8216;soft&#8217; skills like collaboration and critical thinking, to prepare them for an increasingly digital society.</p><ul><li><p>Critical thinking involves analysing, evaluating, and synthesising information to make reasoned decisions. It encompasses problem-solving, decision-making, and reflective thinking, but as a concept it remains somewhat vague and it is difficult to assess how teaching programmes or technology affect it.</p></li><li><p>In particular, a long-running debate exists over whether technologies that automate routine tasks - from calculators to personal computers to AI - support or hinder critical thinking. There is also concern over whether such technologies may erode foundational knowledge or skills - like reading or numeracy - that may be essential for critical thinking and which appear to be <a href="https://www.ft.com/content/e2ddd496-4f07-4dc8-a47c-314354da8d46">stagnating</a> or even declining in many countries.</p></li></ul></li></ul><ul><li><p>In this study, the authors surveyed participants on how frequently and in what ways they used AI to retrieve information and make decisions. Participants were then asked about AI&#8217;s impact on their ability to think critically and solve problems independently, as well as their concerns regarding AI bias and transparency.</p></li></ul><ul><li><p>The findings suggest that while AI enhances efficiency for some individuals, it may come at the cost of a decline in independent problem-solving and critical analysis - or at least a perception of one. This could be due to &#8216;cognitive offloading&#8217;, where users delegate tasks to AI without redirecting their efforts to more complex, higher-order thinking.</p></li><li><p>The study also found that younger participants (17&#8211;25 years old) and those with lower educational attainment reported a greater dependence on AI, potentially reducing their ability to critically evaluate information and identify biases.</p></li><li><p>These findings mirror a recent <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4895486">experiment</a> where students given temporary access to LLMs for learning performed worse on exams once access was revoked, compared to students who had never used the tools in the first place.</p></li></ul><ul><li><p>While this study raises concerns about AI&#8217;s impact on critical thinking, it also presents an opportunity. The research suggests that AI&#8217;s effects are malleable, and education could provide &#8216;cognitive scaffolding&#8217; to help users engage with AI in more productive ways. For example, in their recent <a href="https://blog.google/outreach-initiatives/education/google-learnlm-gemini-generative-ai/">LearnLM</a> work, our colleagues developed pedagogy-inspired LLMs that encourage students to reflect on questions rather than simply answering them. More broadly, there is a need for <a href="https://experience-ai.org/en/">new kinds of AI literacy</a> that does not just teach students, or educators, what AI is, but how to use it most effectively.</p></li></ul><h1><strong>Sector spotlight</strong></h1><ul><li><p><strong>What Happened: </strong>John Micklethwait, editor-in-chief of Bloomberg News, and former editor of <em>The Economist,</em> shared <a href="https://www.bloomberg.com/news/articles/2025-01-10/8-ways-ai-will-transform-journalism?embedded-checkout=true">predictions</a> on AI&#8217;s impact on journalism in a talk at the James Cameron Memorial Lecture. He outlines a positive vision for AI&#8217;s integration into journalism and journalist jobs, yet predicts a decline for traditional Search and ad-based revenue models.</p></li><li><p><strong>What&#8217;s Interesting: </strong>Micklethwait compares AI&#8217;s disruption to the early 2000s, when the Internet was blamed for &#8216;<a href="https://www.economist.com/leaders/2006/08/24/who-killed-the-newspaper">killing&#8217; newspapers</a>. He argues that media outlets were too quick to accept tech companies&#8217; claims that content should be free, leading to a race to the bottom in pursuit of clicks. More recently, publications like The <em>New York Times</em> and <em>The Information </em>have reversed course, building high-priced subscription businesses. With AI, he expects a faster shift to high-quality AI-enhanced content, as media outlets try to avoid repeating past mistakes.</p><ul><li><p>To illustrate AI&#8217;s role in journalism, Micklethwait highlights that one-third of Bloomberg&#8217;s 5,000 daily articles now incorporate some form of automation. In one example, investigative journalists used AI to analyse satellite imagery of ship movements to uncover oil smuggling from Iran. On the other end of the journalism spectrum, Bloomberg also uses AI to generate bullet-point article summaries - a feature disliked by journalists but valued by readers.</p></li><li><p>In his predictions, Micklethwait envisions AI reshaping journalism tasks but not eliminating jobs. He cites Bloomberg&#8217;s continued employment of roughly the same number of company earnings reporters, despite years of automation in that area. Similarly, he expects AI to play a growing role in editing and formatting articles, but to remain less capable of other editorial actions - such as commissioning stories or persuading a cabinet minister to reveal a resignation.</p></li><li><p>Micklethwait also warns of AI-enabled misinformation, particularly in image and video content, which he sees as more harmful than text-based misinformation - especially for fast-moving news stories, where social media plays a key role in verification. Following licensing deals, he expects the decline of traditional search engines and the media outlets dependent on search-driven revenue. After many false dawns, he also expects the long-awaited emergence of truly <em>personalised </em>AI news offerings.</p></li></ul></li><li><p>Looking backwards, over the past 15 years, the number of working journalists in the <a href="https://pressgazette.co.uk/media-audience-and-business-data/number-of-journalists-prs-uk/">UK</a> and <a href="https://www.washingtonpost.com/business/2024/07/12/news-reporters-journalism-jobs-census/">US</a> appears to have grown slowly, and with increasing diversification of Internet-enabled roles. While uncertainty is high, we expect this trend to continue over the next five years, as AI-driven analysis becomes standard, though it remains to be seen whether traditional or newer media outlets will lead the shift.</p></li></ul><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives ! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#16)]]></title><description><![CDATA[AI & material science, AI Safety Institutes, and Central Banks]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-december-2024</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-december-2024</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Tue, 17 Dec 2024 15:41:50 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!s2E4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Every month, our AI Policy Primer looks at 3 external developments from the world of AI policy that caught our eye. In our final edition of the year, we look at a study into the effects of AI on material scientists, the recent meeting of the AI Safety Institutes, and how Central Banks want to use AI to promote financial stability. Please leave a comment below to let us know your thoughts, or send any feedback to <a href="mailto:aipolicyperspectives@google.com">aipolicyperspectives@google.com</a>. Thanks for reading!</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!s2E4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!s2E4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp 424w, https://substackcdn.com/image/fetch/$s_!s2E4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp 848w, https://substackcdn.com/image/fetch/$s_!s2E4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp 1272w, https://substackcdn.com/image/fetch/$s_!s2E4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!s2E4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp" width="800" height="503" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:503,&quot;width&quot;:800,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:27672,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.aipolicyperspectives.com/i/153265353?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!s2E4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp 424w, https://substackcdn.com/image/fetch/$s_!s2E4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp 848w, https://substackcdn.com/image/fetch/$s_!s2E4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp 1272w, https://substackcdn.com/image/fetch/$s_!s2E4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ea776c3-2cfd-4da8-9899-44077b55ac10_800x503.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><h1><strong>Study watch</strong></h1><h2>AI helps scientists to discover new materials, but it may make them enjoy their job less</h2><ul><li><p><strong>What Happened: </strong>Aidan Toner-Rodgers at MIT <a href="https://aidantr.github.io/files/AI_innovation.pdf">published</a> the results of an experiment in which more than 1,000 scientists at a US private sector R&amp;D lab were given access to an AI material design model, to assess how it changed the rate at which they discovered new materials.</p></li></ul><ul><li><p><strong>What&#8217;s Interesting: </strong>To design materials, scientists often rely on iteration or trial and error to explore the huge search space of potential compounds. This study highlights how AI could help improve this process. In 2022, the scientists received access to an unnamed AI model that outputted compounds that were predicted to possess desired characteristics. On average, the AI-assisted scientists subsequently discovered 44% more materials, filed 39% more patents, and produced 17% more product prototypes, with the new materials scoring high on both &#8216;novelty&#8217; and &#8216;quality&#8217; - although the experiment did not capture longer-term commercial impact.</p></li></ul><ul><li><p>In the past 1-2 years, various experiments have studied the effects of giving AI tools to professionals, including <a href="https://arxiv.org/pdf/2302.06590">programmers</a>, <a href="https://economics.mit.edu/sites/default/files/inline-files/Noy_Zhang_1.pdf">writers</a> and <a href="https://www.nber.org/system/files/working_papers/w31161/w31161.pdf">customer service agents</a>. Many of these studies - <a href="https://www.researchgate.net/publication/369589868_When_and_How_Artificial_Intelligence_Augments_Employee_Creativity">though not all </a>- found evidence that lower-skilled employees benefit most from AI. This MIT study finds the inverse, with experienced scientists enjoying the biggest gains. This is because the task of material design is not simply about finding a novel compound, but rather about being able to identify compounds that are most likely to be <em>viable </em>and <em>useful</em>. As the AI model predicts new compounds, it shifts the focus of scientists to<em> </em>evaluating<em> </em>the viability and usefulness of those predictions - a task that those with deep domain expertise are best-suited to.</p></li></ul><ul><li><p>AI will not just affect science, but also <em>scientists. </em>To understand these effects, the MIT study surveyed scientists who received access to the AI model and found that most reported a decline in their work satisfaction. This was true even for those scientists who benefited from the AI tool, owing to concerns that their skills were being under-utilised and the creativity of their role reduced. Wellbeing is hard to measure, and attitudes to technology can change with time, but this finding highlights the need to better understand how AI may affect scientists, a topic that we also explored in our recent <a href="https://www.aipolicyperspectives.com/p/a-new-golden-age-of-discovery">essay</a> about AI and science.</p></li><li><p>In the next 1-2 years, we hope to see an increase in evaluations focussed on empirically assessing how AI is affecting <em>science </em>and scientists, in a similar vein to the recent suite of new evaluations that focus on AI safety.</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives ! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1><strong>Policymakers taking action</strong></h1><h2>US &amp; UK AI Safety Institutes convene in San Francisco</h2><ul><li><p><strong>What happened:</strong> On 20-21 November, the US AI Safety Institute held a convening in San Francisco to kickstart a new technical collaboration between the global network of AI safety institutes, ahead of the upcoming <a href="https://www.elysee.fr/en/sommet-pour-l-action-sur-l-ia">AI Action Summit</a> in Paris in February 2025. The UK AISI also held a convening to share best practices for how to develop AI safety frameworks, like Google DeepMind&#8217;s <a href="https://deepmind.google/discover/blog/introducing-the-frontier-safety-framework/">Frontier Safety Framework.</a></p></li><li><p><strong>What&#8217;s interesting:</strong> As announced in their <a href="https://www.nist.gov/system/files/documents/2024/11/20/Mission%20Statement%20-%20International%20Network%20of%20AISIs.pdf">mission statement</a>, the new global network of AISIs will focus on advancing research to understand the capabilities and risks of advanced AI systems, as well as building best practices for testing them. They also completed their first<a href="https://www.nist.gov/system/files/documents/2024/11/21/Improving%20International%20Testing%20of%20Foundation%20Models-%20%20%20A%20Pilot%20Testing%20Exercise%20from%20the%20International%20Network%20of%20AI%20Safety%20Institutes.pdf"> joint testing exercise</a>, on Llama-3.1, and shared insights about how to improve multilingual AI testing.</p></li><li><p>This is the first time the AISIs have met and announced shared priorities. They showed interest in coordinating more and exchanging best practices. Synthetic content was one of the three areas discussed during the convening, with the global network of AISIs announcing $11 million to fund new research on how to mitigate risks in this area. The global network of AISIs - currently numbering 10 - may expand further in 2025, and they may look to conduct more joint testing exercises. This collaboration could potentially reduce the risk of separate bilateral conversations and fragmentation in the AISIs&#8217; mandates.</p></li></ul><h1><strong>Sector spotlight</strong></h1><h2>Central Banks begin to scale up their use of AI</h2><ul><li><p><strong>What Happened: </strong>Central Banks play a critical role in most modern economies, where among other mandates, they are typically responsible for maintaining the stability of prices. As outlined in the <a href="https://www.bis.org/publ/arpdf/ar2024e3.htm">BIS Annual Economic Report 2024</a>, central banks are increasingly using AI to improve how they make decisions. Key focus areas include economic forecasting, financial supervision, and payment systems.</p></li><li><p><strong>What&#8217;s Interesting:</strong> One area central banks are focussing on is identifying signals and anomalies in the vast datasets they have access to. For example, the Bank of England and the European Central Bank are using AI to <a href="https://www.bis.org/ifc/publ/ifcb57.pdf">look for </a>unexpected transaction patterns or spikes in asset price volatility that may signal liquidity issues. Similarly, the Deutsche Bundesbank is focused on detecting outliers in major financial data sets, which could signal risks such as mispricing in the market. This anomaly detection can be difficult for humans to do reliably, and so AI could help central banks improve resilience against different kinds of financial risks and crime.</p></li><li><p>Central Banks are also using AI to synthesise insights, including sentiment and trends, from unstructured text, to improve their forecasting. For example, the Bank of France is using AI to better gauge public perceptions on inflation, which can provide insights about how future &#8216;sticky&#8217; price growth may be. Meanwhile, Malaysia&#8217;s central bank uses AI to analyse financial news articles to help forecast key indicators, such as GDP and consumer spending.</p></li><li><p>Finally, central banks are exploring the merits of <em>tokenized payment systems</em>&#8212;digital versions of money that use blockchain technology&#8212;and <em>unified ledgers</em> - systems that combine financial and other records in one place. These technologies could potentially make transactions faster, more transparent, or more secure. A <a href="https://www.atlanticcouncil.org/cbdctracker/">growing number</a> of central banks are planning to launch their own tokenized payment systems, such as Central Bank Digital Currencies. AI's role in these developments could take several forms, including detecting and preventing fraud in real-time, or monitoring transactions to adhere to anti-money laundering or counter-terrorism regulations.</p></li><li><p>In the next 1-2 years, we will see central banks explore if AI could also support monitoring and mitigating emerging AI-driven risks to financial stability, such as market disruptions or collusion by autonomous agents - an area that has received relatively little attention in discussions about AI safety risks.</p><p></p></li></ul><p><em>As always, please let us know your thoughts on these updates and what you have found most interesting in the world of AI policy in the last month.  </em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives ! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#15)]]></title><description><![CDATA[Data governance, legal services, and agent communication]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-october-2024</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-october-2024</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Tue, 12 Nov 2024 14:16:48 GMT</pubDate><enclosure url="https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw"><img src="https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" width="3000" height="1688" data-attrs="{&quot;src&quot;:&quot;https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1688,&quot;width&quot;:3000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;a bunch of balloons that are in the air&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="a bunch of balloons that are in the air" title="a bunch of balloons that are in the air" srcset="https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1717501218037-2a88e2cbd2f6?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><p>In the October edition of the AI Policy Primer, we look at research exploring a shift towards data governance and away from compute governance, how AI is being used in the legal sector, and new work outlining a communication protocol for AI agents. As usual, leave a comment or let us know if you have any feedback at aipolicyperspectives@google.com.</p><h2><strong>What we&#8217;re reading&nbsp;</strong></h2><h3>From compute governance to data governance&nbsp;&nbsp;&nbsp;</h3><ul><li><p><strong>What happened: </strong>A team at Berkeley announced a new initiative, and an accompanying <a href="https://arxiv.org/pdf/2409.17216">paper</a>, which calls for AI governance efforts to shift away from relying on &#8216;compute&#8217; to identify a &#8216;frontier&#8217; or risky AI model, and towards approaches that centre &#8216;data&#8217; as well.&nbsp;</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>Several AI governance initiatives, such as the <a href="https://www.europarl.europa.eu/topics/en/article/20230601STO93804/eu-ai-act-first-regulation-on-artificial-intelligence">EU AI Act</a> and the <a href="https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/">US Executive Order on AI</a>, use some measure of compute, such as total parameter count and/or FLOPs, to identify the most powerful &#8216;frontier&#8217; AI models. These models are in turn subject to various governance measures and safety assessments. Other forms of <a href="https://arxiv.org/pdf/2402.08797">&#8216;AI compute governance&#8217;</a> include export controls on certain chips.&nbsp;</p></li><li><p>The Berkeley team argue&#8217;s that relying on compute to gauge the risk posed by an AI model&nbsp;is imperfect because advances in efficiency and <a href="https://arxiv.org/abs/2402.16828">distributed models of training</a> may &#8216;decouple&#8217; model performance from computational cost. Some state-of-the-art models are also relatively small. In image segmentation, for example, the authors note that a <a href="https://arxiv.org/abs/2312.01623">smaller model</a> by researchers in China outperforms the much larger <a href="https://ai.google.dev/gemma/docs/paligemma">PaliGemma model</a> from Google on the <a href="https://paperswithcode.com/dataset/refcoco">RefCOCO</a> dataset. In AI biology, smaller models like AlphaFold outperform much larger models like <a href="https://www.evolutionaryscale.ai/blog/esm3-release">ESM-3 </a>on tasks like protein structure prediction.&nbsp;</p></li><li><p>The authors also argue that it is increasingly the quality and use of <em>data</em>, both training data and deployment data (such as prompts or RAG documents) that determine model performance. They propose that it will be the combination of models exposed to specific datasets that will lead to risk, rather than models alone. In response, the group suggests making better use of existing data regulations to respond to risks posed by AI, and calls on AI labs to invest more in techniques to evaluate, red-team, and filter datasets &#8211; including to assess the incremental uplift they provide to models.&nbsp;</p></li></ul></li><li><p><strong>Looking ahead: </strong>Compute will remain the primary metric that policymakers use to identify the most powerful new AI models. But as datasets become more important, we may see new efforts to tweak existing regulations as well as more market solutions &#8211; such as new types of AI licensing regimes.&nbsp;</p></li></ul><h2><strong>Sector spotlight</strong></h2><h3>AI and legal services</h3><ul><li><p><strong>What happened: </strong>AI continues to impact different services. One of the most interesting of these is the legal profession, where AI increasingly is becoming a key tool for lawyers and paralegals. On July 29th 2024 the American Bar Association published <a href="https://www.americanbar.org/content/dam/aba/administrative/professional_responsibility/ethics-opinions/aba-formal-opinion-512.pdf">Formal Opinion 512,</a> which lays out the ground rules for lawyers who want to use generative AI. This is representative of a key trend in which various professions carefully consider the technology, issue guidance and come together around sector norms.&nbsp;</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>The opinion addresses six key areas: competence (lawyers must understand AI's capabilities and limitations and independently verify its output); confidentiality (lawyers must evaluate risks of disclosure and obtain informed consent before inputting client information); communication (lawyers must inform clients about AI use in certain circumstances); duties to courts (lawyers must verify AI outputs for accuracy in court submissions); supervision (law firms must establish clear AI policies and ensure proper training); and fees (lawyers must charge reasonable fees for AI use and cannot bill clients for time saved through AI efficiency or for learning to use AI tools).&nbsp;</p></li><li><p>The opinion acknowledges AI's potential to improve legal service efficiency while emphasising that AI cannot replace professional judgment and that lawyers remain fully responsible for their work product. It also recognises that AI technology is rapidly evolving and that further guidance may be needed as these tools develop.&nbsp;</p></li></ul></li><li><p><strong>Looking ahead: </strong>Formal Opinion 512 is an example of how professional social norms regulate the use of artificial intelligence. This kind of regulation usually gets much less attention than legislation, but remains one of the most important tools for a society adapting to the complexity that new technology introduces in different fields. The legal industry will become a core arena for shaping AI regulation alongside many other high-risk areas.&nbsp;&nbsp;</p></li></ul><h2><strong>What we&#8217;re reading&nbsp;</strong></h2><h3>Agora: Enabling agent collaboration</h3><ul><li><p><strong>What happened</strong>:&nbsp;</p><ul><li><p>Building <em>collaborative</em> AI systems where multiple LLMs specialise in different tasks is challenging. Imagine trying to coordinate a team where everyone speaks a different language and with different cultural expectations &#8211; a similar issue arises when diverse LLMs attempt to interact. To study this problem, researchers at Oxford University have <a href="https://arxiv.org/html/2410.11905v1">introduced</a> Agora, a new communication protocol designed to enable more efficient and scalable collaboration between large language models.&nbsp;</p></li><li><p>The communication bottleneck described by the authors stems from what they call the &#8216;agent communication trilemma&#8217;, which captures three distinct challenges:&nbsp;agents vary significantly in their architecture and training data (<strong>heterogeneity</strong>); language models are general-purpose tools, making it impractical to define and pre-program every possible interaction scenario (<strong>generality</strong>); and agents are computationally expensive (<strong>cost</strong>).&nbsp;</p></li></ul></li><li><p><strong>What&#8217;s interesting</strong>:&nbsp;</p><ul><li><p>Agora aims to break this trilemma, which makes it hard to design cost-effective ways for agents to communicate across different types of scenarios. It employs pre-defined routines, similar to APIs, for common tasks. This greatly reduces the computational overhead compared to relying solely on language models for every interaction. For less frequent communications, Agora leverages structured data like JSON, which seeks to offer a balance between flexibility and efficiency.</p></li><li><p>Only in rare cases, such as unexpected errors or the need for complex negotiation, does Agora fall back on natural language. A key innovation of Agora is the use of &#8220;protocol documents&#8221; (PDs) &#8211; machine-readable descriptions of communication protocols. Agents can share and learn these PDs, allowing them to automatically adapt their communication strategies without human intervention.&nbsp;</p></li></ul></li><li><p><strong>Looking ahead</strong>: We expect the research and production of agents to continue to grow in the coming months and years. As such we anticipate the need for streamlined communication between agents will only grow over time, and this will in turn incentivise the development of specialised protocols like Agora. Which <em>specific</em> protocol will be ultimately adopted by industry remains to be seen.&nbsp;</p></li></ul><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives! Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#14)]]></title><description><![CDATA[Compute, energy, and competitive measures]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-september-2024</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-september-2024</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Wed, 16 Oct 2024 13:09:12 GMT</pubDate><enclosure url="https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw"><img src="https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" width="3000" height="1688" data-attrs="{&quot;src&quot;:&quot;https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1688,&quot;width&quot;:3000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;a close up of a blue and green structure&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="a close up of a blue and green structure" title="a close up of a blue and green structure" srcset="https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1717501218198-816a64915f81?fm=jpg&amp;q=60&amp;w=3000&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><p>In this edition of the AI Policy Primer, we have pieces on investment in compute, work assessing AI&#8217;s energy footprint, and moves by the EU to boost the competitiveness of its AI capabilities. As usual, leave a comment or let us know if you have any feedback at aipolicyperspectives@google.com.</p><h2><strong>What we&#8217;re reading&nbsp;</strong></h2><h3>Compute investments add up </h3><ul><li><p><strong>What happened: </strong>Trials in the TSMC Arizona plant <a href="https://www.bloomberg.com/news/articles/2024-09-06/tsmc-s-arizona-trials-put-plant-productivity-on-par-with-taiwan">reportedly</a> put the fab&#8217;s productivity on par with some of the firm&#8217;s locations in Taiwan. The plant, which will <a href="https://www.nytimes.com/2024/02/19/technology/semiconductor-chip-factories-delays.html">begin</a> commercial operations in the first half of 2025, was the subject of intense <a href="https://www.nytimes.com/2024/08/08/business/tsmc-phoenix-arizona-semiconductor.html">scrutiny</a> just one month ago. According to the report,&nbsp;in trial production, the yield rate - how many usable chips a company can produce during a single manufacturing process - is similar to comparable facilities in the southern Taiwanese city of Tainan.&nbsp;</p></li><li><p><strong>What&#8217;s interesting: </strong>The news comes at a busy time for chip companies and data centre developers. Intel, the American semiconductor giant facing stiff competition, <a href="https://www.intc.com/news-events/press-releases/detail/1710/a-message-from-intel-ceo-pat-gelsinger-to-employees">announced</a> greater autonomy for its foundry business. In a press release, the firm said that Intel Foundry would be established as an independent subsidiary to provide &#8220;future flexibility to evaluate independent sources of funding.&#8221; Earlier this month, Intel was also <a href="https://www.businesswire.com/news/home/20240916866974/en/Intel-Awarded-up-to-3B-by-the-Biden-Harris-Administration-for-Secure-Enclave">awarded</a> up to $3 billion from the CHIPS and Science Act, which seeks to bring chipmaking to the U.S. via the &#8220;Secure Enclave&#8221; program in partnership with the Department of Defense.</p><ul><li><p>Technology firms in the US also continue to <a href="https://www.bbc.co.uk/news/articles/cx25v2d7zexo">invest</a> in compute capacity to train large models, while in the UK, new developer DC01UK <a href="https://www.uktech.news/big-data/plans-submitted-herfordshire-data-centre-20240912">submitted plans</a> for a &#163;3.8bn data centre in Hertfordshire - although questions were <a href="https://www.uktech.news/cloud/who-is-dc01uk-the-firm-behind-plans-for-a-huge-billion-pound-data-centre-20240912">raised</a> about the structure and experience of the group.</p></li><li><p>For their part, policymakers are considering how to best assess and manage this growing demand for data centres and chips and what it means for energy supplies. In the US, the Department of Energy recently convened <a href="https://www.energy.gov/sites/default/files/2024-08/Powering%20AI%20and%20Data%20Center%20Infrastructure%20Recommendations%20July%202024.pdf">a Working Group on Powering AI and Data Center Infrastructure</a>, which published a set of recommendations ranging from advancing different types of efficiencies for LLM training and inference to exploring new types of&nbsp;generation, storage and grid technologies to power data centres. It also noted that any efforts to predict energy demand were &#8220;fraught with uncertainties&#8221; (the subject of our next update below).&nbsp;</p></li></ul></li><li><p><strong>Looking ahead: </strong>Investment in compute capacity and energy infrastructure will continue to increase dramatically. In the next two years, we may see several developers implement plans to train models requiring over 1 gigawatt of power - <a href="https://www.energy.gov/eere/articles/how-much-power-1-gigawatt">equivalent </a>to approximately 300 utility-scale wind turbines. </p></li></ul><h2><strong>What we&#8217;re reading&nbsp;</strong></h2><h3>Power consumption under the spotlight&nbsp;&nbsp;&nbsp;</h3><ul><li><p><strong>What happened: </strong>Tim Fist at the Institute for Progress <a href="https://x.com/fiiiiiist/status/1836471413198459331">assessed</a> a recent estimate from a Washington Post <a href="https://www.washingtonpost.com/technology/2024/09/18/energy-ai-use-electricity-water-data-centers/">article</a> which claimed that GPT-4 consumes 0.14kWh of energy to produce a 100-word email. Fist suggests that the estimate, which was produced in collaboration with researchers at the University of California, Riverside, was off by a factor of at least 350x.&nbsp;</p></li><li><p><strong>What&#8217;s interesting: </strong>If correct, Fist&#8217;s analysis would put the Washington Post article&nbsp;in line with other past claims about dramatic energy usage, or greenhouse gas emissions, from digital technology that were subsequently questioned or debunked. For example, George Kamiya, then at the IEA, <a href="https://www.iea.org/commentaries/the-carbon-footprint-of-streaming-video-fact-checking-the-headlines">explained</a> why claims by an NGO, <a href="https://theshiftproject.org/en/article/shift-project-really-overestimate-carbon-footprint-video-analysis/">The Shift Project</a>, during COVID-19 that watching 30 minutes of Netflix generated 1.6kg of CO2 were off by 90x. Jonathan Koomey and Eric Masanet have also<a href="https://www.sciencedirect.com/science/article/pii/S2542435121002117"> cautioned</a> about regular missteps in this area, such as conflating increased data, or internet use, with increased energy use, which can ignore important questions such as whether demand affects peak capacity.&nbsp;</p><ul><li><p>Experts have created <a href="https://github.com/lfwa/carbontracker">tools</a> to estimate the energy use and emissions from AI models more reliably, though mainly focussed on training rather than inference. These estimations can be complex and require information that is not always available, such the efficiency of the data centre, the source of energy, the type of chips used, and the training protocols. </p></li><li><p>Beyond these methodological challenges, any estimates for the energy used to train or deploy an AI model &#8211; and the emissions generated &#8211; will only ever offer a partial answer to the broader question about how an AI model will impact the environment. A lifecycle approach also requires considering other types of emissions, such as the embodied emissions from building a data centre or device, and any indirect effects on emissions from applications that AI enables. These estimates, in turn, would need to be considered against counterfactual scenarios that do not use AI, as few actions in the modern economy produce no emissions.&nbsp;</p></li><li><p>One of the main reasons for disagreements in this space is the upfront energy costs from training AI models are growing and tend to appear quickly, with relative certainty, while the downstream benefits - from a more efficient Internet to potentially helping to enable new kinds of materials for solar panels or batteries - are potentially much more consequential, but also less immediate and certain.&nbsp;</p></li></ul></li><li><p><strong>Looking ahead: </strong>As models increase in size &#8211; and despite the emergence of more efficient training and inference procedures &#8211; require more energy, we expect interested third parties to continue to develop new methodologies to estimate the energy required to train and run AI models. Wide-ranging debates about what this may mean for emissions is likely to follow. &nbsp;</p></li></ul><h2><strong>Policymakers taking action</strong></h2><h3>EU eyes competitive measures&nbsp;</h3><ul><li><p><strong>What happened: </strong>The European Union is heading into a new mandate and things are changing fast. The new regulatory agenda will be key to follow. At the outset of the recent <a href="https://commission.europa.eu/topics/strengthening-european-competitiveness/eu-competitiveness-looking-ahead_en">Draghi report</a>, authored by European eminence grise Mario Draghi, the paper suggests that Europe needs to invest massively to compete with the rest of the world - and not just regulate - but really grow a tech sector of its own. Draghi did not pull punches: &#8220;Technological change is accelerating rapidly. Europe largely missed out on the digital revolution led by the internet and the productivity gains it brought: in fact, the productivity gap between the EU and the US is largely explained by the tech sector. The EU is weak in the emerging technologies that will drive future growth. Only four of the world&#8217;s top 50 tech companies are European.&#8221;</p></li><li><p><strong>What&#8217;s interesting: </strong>Europe is at a crossroads. The old hypothesis was that regulation would provide Europe with a seat at the global table. Will the new Commission still agree that this is the case? Efforts such as the <a href="https://digital-strategy.ec.europa.eu/en/policies/ai-factories">AI factories</a> programme, which will allow AI developers to build on the EuroHPC network of supercomputers, attempt to put the EU on the path to compete with the US. But will it be enough? The AI factories are meant to help startups and public sector efforts - but without access to capital markets and growth mechanisms, will startups stay in Europe? And will the code of practice for AI now being drafted help the European efforts to focus on productivity growth?&nbsp;</p></li><li><p><strong>Looking ahead: </strong>This is the question the European Union needs to answer: whether it will double down on regulation as a competitive advantage, or if it will pivot to policy interventions seeking to bolster innovation. For that reason, we may see the European Commission double down on opening the European market for AI, and implementing the AI-act in a way that allows for a transatlantic market to emerge over time.</p></li></ul>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#13)]]></title><description><![CDATA[Security, evaluations, and critical technologies]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-august-2024</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-august-2024</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Fri, 06 Sep 2024 08:58:08 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Pqvq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Pqvq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Pqvq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png 424w, https://substackcdn.com/image/fetch/$s_!Pqvq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png 848w, https://substackcdn.com/image/fetch/$s_!Pqvq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png 1272w, https://substackcdn.com/image/fetch/$s_!Pqvq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Pqvq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png" width="1432" height="776" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:776,&quot;width&quot;:1432,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2114634,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Pqvq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png 424w, https://substackcdn.com/image/fetch/$s_!Pqvq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png 848w, https://substackcdn.com/image/fetch/$s_!Pqvq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png 1272w, https://substackcdn.com/image/fetch/$s_!Pqvq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ec25223-396e-4deb-895d-8a9c7843a13e_1432x776.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind </figcaption></figure></div><p>In the latest edition of the AI Policy Primer, we have pieces on security in AI, &#8216;questionable practices&#8217; in evaluations, and a new league table comparing national performance across various critical technologies. As always, leave a comment or let us know if you have any feedback at aipolicyperspectives@google.com.</p><h2><strong>What we&#8217;re reading&nbsp;</strong></h2><h3>Security under the spotlight&nbsp;&nbsp;&nbsp;</h3><ul><li><p><strong>What happened:&nbsp;</strong></p><ul><li><p>In July, at the Aspen Security Forum, Google <a href="https://blog.google/technology/safety-security/google-coalition-for-secure-ai/?utm_source=tw&amp;utm_medium=social&amp;utm_campaign=og&amp;utm_content=&amp;utm_term=">introduced the Coalition for Secure AI </a>(CoSAI) alongside Anthropic, Microsoft, and OpenAI. CoSAI aims to support a collective investment in AI security, initially by focusing on three areas: software supply chain security for AI systems, preparing defenders for a changing cybersecurity landscape, and AI security governance.&nbsp;&nbsp;</p></li><li><p>The same month, the Frontier Model Forum <a href="https://www.frontiermodelforum.org/updates/issue-brief-foundational-security-practices/">outlined a set of foundational security best practices</a>, which noted that as frontier AI systems become more capable, developing and implementing a security strategy that &#8220;effectively layers and integrates both traditional and tailored [security] approaches&#8221; will be vital. Recommendations from the work include applying fundamental security principles to AI, establishing proactive security management measures, securing the deployment and distribution of AI models, implementing insider threat detection programs, and developing and testing robust incident response and recovery procedures.</p></li></ul></li><li><p><strong>What&#8217;s interesting: </strong>A robust approach to AI security requires both adapting and standardising concepts from software security, and introducing novel thinking and experimentation about the unique technical aspects of frontier systems. In addition to thinking about security in the traditional sense, AI security may also be conceived of more broadly: red-teaming, post-deployment monitoring, and dynamic responses are all measures that can boost security.&nbsp;</p></li><li><p><strong>Looking ahead: </strong>Discussions about what constitutes &#8220;good enough&#8221; frontier AI security continue to intensify. Policymakers are also starting to introduce more prescriptive proposals for how to secure AI systems, such as California&#8217;s Safe and Secure Innovation for Frontier Artificial Intelligence Models Act.</p></li></ul><h2><strong>Sector spotlight&nbsp;</strong></h2><h3>Questionable practices in machine learning&nbsp;&nbsp;</h3><ul><li><p><strong>What happened: </strong>Researchers from the University of Bath, University of Bristol, and Arb Research took aim at the challenges in the evaluation of large language models. In a <a href="https://arxiv.org/pdf/2407.12220">paper</a> published in July, the group lists 43 ways machine learning evaluations can be &#8220;misleading or actively deceptive.&#8221; Taking inspiration from psychological science, they call these instances &#8220;questionable research practices&#8221;, which they group into three categories.</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p><strong>Contamination: </strong>This group includes the various ways that information can leak from one part of the model training process to another. The most well-known example of this phenomenon is training contamination, which sees data from the training set (the set of examples a model learns in pre-training) migrate to the test set (a new set of examples that it shouldn&#8217;t have seen before used to assess performance).&nbsp;</p></li><li><p><strong>Cherrypicking: </strong>The second group involves choosing amongst runs to make a system look more impressive than it is. In practice, this sees researchers &#8216;hack&#8217; experiments by selecting those under which their model works better than others after testing multiple times. This group also includes techniques such as prompt hacking (choosing the best prompt strategy like implementing chain of thought approaches that work better for some models than others) and benchmark hacking (picking the easiest benchmarks for a particular model).&nbsp;</p></li><li><p><strong>Misreporting:</strong> Finally, the paper considers the ways in which researchers mayindulge in misleading calculations or presentations. This bucket includes methods such as under-reporting the size of a particular model, failing to report negative benchmark studies, and pretraining a model on benchmark or instruction data.&nbsp;</p></li></ul></li><li><p><strong>Looking ahead:</strong> Both developers and third party observers stress the importance of evaluations for determining the capabilities and risk profiles of AI systems. As a result, more critical work is likely to appear in the future as evaluations remain a topic of lively discussion.&nbsp;</p></li></ul><h2><strong>Sector spotlight&nbsp;</strong></h2><h3>A shift in research leadership towards the Indo-Pacific&nbsp;&nbsp;</h3><ul><li><p><strong>What happened: </strong>The Australian Strategic Policy Institute (ASPI) released a major update to its <a href="https://techtracker.aspi.org.au/">Critical Technology Tracker</a>, which compares the adoption of strategically-relevant technologies around the world.&nbsp;The dataset now covers the top 10% of the most highly cited research publications from the past 21 years (2003&#8211;2023) across 64 critical technologies as &#8220;an indicator of a country&#8217;s research performance, strategic intent and potential future science and technology capability&#8221;.</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>The tracker <a href="https://www.economist.com/science-and-technology/2024/06/12/china-has-become-a-scientific-superpower">reinforces</a> a dramatic shift in leadership over the past two decades. While the US held a commanding lead in the early 2000s, leading in 60 out of 64 technologies, its dominance has eroded while China has made major gains, surging from a lead in just three technologies in the 2000s, to a current lead in 57 out of 64 (including machine learning). The US, however, retains a lead in natural language processing. Though only the US or China lead in any technology, India now ranks in the top 5 countries for 45 of 64 technologies (an increase from just four in the 2000s), while the UK was in the top 5 for 36 technologies.&nbsp;</p></li><li><p>ASPI argues that maintaining scientific and research leadership is not a simple &#8216;on-off switch&#8217;, and requires sustained investment in scientific knowledge, talent, and high-performing institutions over the long term. They argue that countries that have scaled back investment in research &#8211; often in domains where they previously held a competitive advantage &#8211; now face a significant challenge in maintaining their position in the future. The report also acknowledges that research excellence, while a critical starting point, is just one part of the equation. Translating research breakthroughs into tangible technological gains and commercial success requires a range of complementary factors, including a healthy manufacturing base and supportive policy frameworks.</p></li></ul></li></ul><p><strong>Looking ahead: </strong>The US and China will likely continue to dominate the critical technologies tracker for the foreseeable future. While American industrial policy may provide the impetus to improve the USA's position in some areas, its effects are unlikely to be felt in the near term. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives! Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#12)]]></title><description><![CDATA[Weather forecasting, AI agents, and synthetic data]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-july-2024</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-july-2024</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Thu, 01 Aug 2024 15:05:03 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!DO6Y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DO6Y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DO6Y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png 424w, https://substackcdn.com/image/fetch/$s_!DO6Y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png 848w, https://substackcdn.com/image/fetch/$s_!DO6Y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png 1272w, https://substackcdn.com/image/fetch/$s_!DO6Y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DO6Y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png" width="1456" height="811" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:811,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4542957,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DO6Y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png 424w, https://substackcdn.com/image/fetch/$s_!DO6Y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png 848w, https://substackcdn.com/image/fetch/$s_!DO6Y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png 1272w, https://substackcdn.com/image/fetch/$s_!DO6Y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96d4cd6b-4ef5-459d-8330-4f9bdd504eb1_2704x1506.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><p>In July&#8217;s edition of the AI Policy Primer, we take a look at weather forecasting models, the governance of AI agents, and recent debates surrounding synthetic data. As always, leave a comment or let us know if you have any feedback at aipolicyperspectives@google.com.</p><h2><strong>What we&#8217;re reading</strong></h2><h3>Taking the temperature of weather models&nbsp;</h3><ul><li><p><strong>What happened: </strong>Google DeepMind&#8217;s <a href="https://deepmind.google/discover/blog/graphcast-ai-model-for-faster-and-more-accurate-global-weather-forecasting/">GraphCast</a>, a state-of-the-art AI weather prediction model, won the <a href="https://macrobertaward.raeng.org.uk/">MacRobert Award</a> hosted by the Royal Academy of Engineering. GraphCast can predict hundreds of weather variables up to ten days in advance, and is faster and more accurate than traditional weather models. The system, which Google DeepMind <a href="https://github.com/google-deepmind/graphcast">open-sourced</a>, was joined in 2024 by <a href="https://windbornesystems.com/blog/windborne-breaks-world-record-for-most-accurate-global-weather-forecasts">WeatherMesh</a> &#8211; a model developed by weather forecasting start-up WindBorne. Google Research also recently released <a href="https://research.google/blog/fast-accurate-climate-modeling-with-neuralgcm/">NeuralGCM</a>, a model that can simulate Earth&#8217;s atmosphere.</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>GraphCast goes beyond standard weather prediction by offering earlier warnings of extreme weather events. It can predict the tracks of cyclones with great accuracy further into the future, characterise atmospheric rivers associated with flood risk, and predict the onset of extreme temperatures. These abilities have the potential to save lives through greater preparedness and faster emergency response, and address environmental challenges.</p></li><li><p>Weather is a domain where the state takes on prediction tasks, for example, via the <a href="https://www.weather.gov/">National Weather Service</a> (NWS) under the <a href="https://www.noaa.gov/">National Oceanic and Atmospheric Administration</a> (NOAA) in the United States and the <a href="https://www.metoffice.gov.uk/">Met Office</a> of the Department for Science, Innovation and Technology (DSIT) in the UK. Before GraphCast, the High-Resolution Forecast (HREF) developed by the independent intergovernmental organisation European Centre for Medium-Range Weather Forecasts&#8217; (ECMWF) was the state-of-the art model.&nbsp;</p></li><li><p>GraphCast, trained on ECMWF&#8217;s ERA 5 dataset, is now being <a href="https://charts.ecmwf.int/products/graphcast_medium-mslp-wind850?base_time=202407240000&amp;projection=opencharts_europe&amp;valid_time=202407240000">used</a> by ECMWF, marking a move towards new modes of public-private partnership in weather prediction. As AI companies increasingly contribute to public goods, we should prepare for the emergence of new types of collaboration between model makers in the private sector and model deployers in the public sector.&nbsp; To that end, the Royal Academy of Engineering <a href="https://raeng.org.uk/news/ai-weather-forecasting-tech-wins-uk-s-top-engineering-award">notes</a> the potential for GraphCast to support critical decision-making across industries and optimise resource allocation.</p></li></ul></li><li><p><strong>Looking ahead: </strong>GraphCast is part of wider research to understand the broader patterns of our climate. Alongside other GDM models &#8211; such as AlphaFold 3, GNoME, and others &#8211; it demonstrates AI's potential to accelerate scientific discovery and address some of our greatest challenges. To learn more, see the <a href="https://raeng.org.uk/news/ai-weather-forecasting-tech-wins-uk-s-top-engineering-award">page</a> for the MacRobert Award, Google DeepMind's <a href="https://deepmind.google/discover/blog/graphcast-ai-model-for-faster-and-more-accurate-global-weather-forecasting/">blog</a> about GraphCast, an accompanying <a href="https://www.science.org/stoken/author-tokens/ST-1550/full">paper</a>, and the code shared on <a href="https://github.com/google-deepmind/graphcast">GitHub</a><strong>.&nbsp;</strong></p></li></ul><h2><strong>Sector spotlight</strong></h2><h3>Governing AI agents&nbsp;&nbsp;</h3><ul><li><p><strong>What happened: </strong>The development and deployment of AI agents continues to&nbsp; spark <a href="https://www.theatlantic.com/technology/archive/2024/07/ai-agents-safety-risks/678864/">commentary</a>. Such systems aim to autonomously plan and execute complex tasks with limited human involvement (unlike AI tools like Gemini or Claude that provide task-specific assistance and respond to user queries without independent initiative or decision-making capabilities).&nbsp;&nbsp;</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>While developers have yet to deploy powerful agents, they have - along with researchers from academia and civil society - released work focused on identifying and assessing the governance mechanisms needed to allow for the safe deployment of such systems.&nbsp;Google DeepMind, for example, published a <a href="https://storage.googleapis.com/deepmind-media/DeepMind.com/Blog/ethics-of-advanced-ai-assistants/the-ethics-of-advanced-ai-assistants-2024-i.pdf">collection of papers</a> considering issues such as value alignment, safety and misuse, economic and environmental impact, epistemic security, and access in the context of agentic AI systems.&nbsp;&nbsp;</p></li><li><p>The University of Toronto&#8217;s Noam Kolt <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4772956">looked</a> at governance challenges connected to discretionary authority (making sure the agent doesn&#8217;t vicariously use authority to act unreasonably), loyalty (determining how best to keep an agent acting in the user&#8217;s best interests), delegation (how to manage the creation of subagents), and information asymmetry (managing situations in which the agent knows more than the person, or &#8216;principal&#8217;, employing it). Kolt also examines visibility, the subject of a <a href="https://arxiv.org/abs/2401.13138">paper</a> by authors at Mila and GovAI, proposing measures including agent identifiers, real-time monitoring, and activity logging.</p></li></ul></li><li><p><strong>Looking ahead: </strong>Developing successful governance structures also requires understanding how agents might actually be used in practice. One method in this vein is Seth Lazar&#8217;s <a href="https://arxiv.org/pdf/2404.06750">work</a> considering the cultural and epistemic impact of AI agents, which outlines the different forms that agents may take: as &#8216;companions&#8217; offering comfort and support, as &#8216;attention guardians&#8217; that could help people decide where to focus, and as &#8216;universal intermediaries&#8217; that mediate our interactions with the digital world.&nbsp;</p></li></ul><h2><strong>Sector spotlight</strong></h2><h3>Climbing the data wall&nbsp;&nbsp;</h3><ul><li><p><strong>What happened:</strong> Debates about the availability of training data, the amount of data likely to be used as models scale, and potential bottlenecks and solutions continue to run. Last month, Epoch AI <a href="https://epochai.org/blog/will-we-run-out-of-data-limits-of-llm-scaling-based-on-human-generated-data">estimated</a> with an 80% confidence interval that the existing high quality training data stock will be fully depleted at some point between 2026 and 2032, bringing new energy to discussions about the &#8220;data wall&#8221; and potential remedies for the problem.&nbsp;</p></li><li><p><strong>What&#8217;s interesting:</strong></p><ul><li><p>Data availability is crucial for AI development. As a <a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">rule of thumb</a>, researchers generally accept that more &#8211; and higher quality data &#8211; tends to lead to better model performance (with the <a href="https://arxiv.org/abs/2407.05694v1">caveat</a> that models at the same capacity have been getting better over time). Some <a href="https://www.dataprovenance.org/Consent_in_Crisis.pdf">research</a> suggests we may soon exhaust available training data, which may in turn stymie the development of frontier models.&nbsp;&nbsp;</p></li><li><p>Google researchers <a href="https://arxiv.org/abs/2304.08466">showed</a> that, when fine-tuning the Imagen text-to-image model, increasing the size of the synthetic dataset monotonically improved the model's accuracy; synthetic data was also used to train Anthropic&#8217;s Claude 3, as outlined in its <a href="https://www-cdn.anthropic.com/de8ba9b01c9ab7cbabf5c33b80b7bbc618857627/Model_Card_Claude_3.pdf">technical report</a>; and comments from <a href="https://www.youtube.com/watch?v=9TU0XjJqpOg">Mark Zuckerberg</a> and <a href="https://twitter.com/SquawkCNBC/status/1782799614501957840">Dario Amodei</a> highlight the importance of synthetic data for scaling.&nbsp;&nbsp;&nbsp;</p></li><li><p>An opposing view, however, proposes that synthetic data may not be enough to overcome the data wall. Studies from <a href="https://arxiv.org/pdf/2305.17493.pdf">Oxford</a> and <a href="https://arxiv.org/pdf/2307.01850.pdf">Rice University</a> both suggest that the use of synthetic data degrades model quality over time, while other <a href="https://arxiv.org/abs/2305.17493v2">research</a> shows that compounding errors from training on synthetic text online may result in a phenomenon known as &#8216;model collapse&#8217;. Recent <a href="https://arxiv.org/pdf/2404.01413">research</a>, however, shows that &#8216;model collapse&#8217; tends only to occur when synthetic data is substituted for real data, rather than added to it.&nbsp;</p></li></ul></li><li><p><strong>Looking ahead:</strong> Developers will likely face other challenges too, like <a href="https://ieeexplore.ieee.org/document/9710332">ensuring the factuality and fidelity of synthetic data</a>, and the potential for synthetic data to <a href="https://arxiv.org/abs/2105.04144">amplify</a> or <a href="https://www.researchgate.net/publication/360377045_A_Methodology_for_Controlling_Bias_and_Fairness_in_Synthetic_Data_Generation">introduce</a> biases. AI itself may be part of this solution, as it can help annotate and curate data, making it more accessible and useful to AI labs and other researchers.&nbsp;</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives! Subscribe for free to receive new posts and support this work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#11)]]></title><description><![CDATA[State-level legislation, science, and energy]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-june-2024</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-june-2024</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Mon, 08 Jul 2024 10:51:56 GMT</pubDate><enclosure url="https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw"><img src="https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" width="3000" height="1688" data-attrs="{&quot;src&quot;:&quot;https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1688,&quot;width&quot;:3000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;a group of hands reaching up into a pile of food&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="a group of hands reaching up into a pile of food" title="a group of hands reaching up into a pile of food" srcset="https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1717501219905-2711c58ab655?fm=jpg&amp;w=3000&amp;auto=format&amp;fit=crop&amp;q=60&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><p>In this month&#8217;s AI Policy Primer, we look at state-level legislation in the US, new work exploring the opportunities and challenges associated with using AI in science, and recent debates about AI and energy. As always, leave a comment or let us know if you have any feedback at aipolicyperspectives@google.com.</p><h2><strong>Policymakers taking action</strong></h2><h3>US state-level legislation moves forward&nbsp;</h3><ul><li><p><strong>What happened: </strong>As Congress struggles to find consensus on federal AI legislation, states are moving forward to fill the vacuum. Over 600 AI-related bills have been introduced in <a href="https://www.multistate.ai/artificial-intelligence-ai-legislation">45 states</a> during this year&#8217;s legislative sessions alone.&nbsp;</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>California faces a 31 August legislative deadline as it attempts to pass <a href="https://leginfo.legislature.ca.gov/faces/billNavClient.xhtml?bill_id=202320240SB1047">far-reaching legislation</a> to impose various requirements related to the development of advanced models, including safety assessments, increased liability for AI developers, and the creation of a new state regulator for AI models. </p></li><li><p>Colorado recently passed <a href="https://leg.colorado.gov/bills/SB24-205">first-of-its-kind legislation</a> requiring developers of high-risk AI systems to prevent algorithmic discrimination by establishing a rebuttable presumption of reasonable care linked to compliance with disclosure, reporting, risk management, and other requirements. Governor Jared Polis signed the measures into law along with an <a href="https://drive.google.com/file/d/1i2cA3IG93VViNbzXu9LPgbTrZGqhyRgM/view">unusual signing statement</a> expressing &#8220;reservations&#8221; about the impacts and encouraging lawmakers to make amendments before the law takes effect in 2026. Similar bills were introduced in over a half-dozen states this year.</p></li></ul></li><li><p><strong>Looking ahead: </strong>Some states are taking a more incremental approach to developing AI policies. For example, <a href="https://custom.statenet.com/public/resources.cgi?mode=show_text&amp;id=ID:bill:IN2024000S150&amp;verid=IN2024000S150_20240313_0_EF&amp;">Indiana</a>, <a href="https://custom.statenet.com/public/resources.cgi?mode=show_text&amp;id=ID:bill:OR2024000H4153&amp;verid=OR2024000H4153_20240327_0_EF&amp;">Oregon</a>, <a href="https://custom.statenet.com/public/resources.cgi?mode=show_text&amp;id=ID:bill:WA2023000S5838&amp;verid=WA2023000S5838_20240318_0_ESE&amp;">Washington</a>, and <a href="https://custom.statenet.com/public/resources.cgi?mode=show_text&amp;id=ID:bill:WV2024000H5690&amp;verid=WV2024000H5690_20240327_0_E&amp;">West Virginia</a> each enacted bills this year establishing multi-stakeholder public-private AI Task Forces to develop legislative and policy recommendations. The flurry of state-level activity could further fracture the AI policy landscape, potentially creating a patchwork of regulations and compliance requirements.&nbsp;</p></li></ul><h2><strong>What we&#8217;re reading</strong></h2><h3>Reports tackle science in the age of AI&nbsp;</h3><ul><li><p><strong>What happened: </strong>The Royal Society's recent publication, "<a href="https://royalsociety.org/news-resources/projects/science-in-the-age-of-ai/">Science in the Age of AI</a>", looks at the opportunities and challenges associated with using AI in science. The report follows publications from the <a href="https://scientificadvice.eu/advice/artificial-intelligence-in-science/">European Commission</a> and <a href="https://www.oecd.org/publications/artificial-intelligence-in-science-a8d820bd-en.htm">OECD</a>, signalling a growing global recognition of AI's transformative potential in scientific research and the need for supportive policy frameworks.&nbsp;</p></li><li><p><strong>What&#8217;s interesting: </strong>These reports note that AI is reshaping science across many fields. With applications spanning medicine, materials science, robotics, climate modelling and more, the capacity for more sophisticated data analysis, pattern recognition, and simulation is changing how scientists approach complex problems. Additionally, the reports highlight that:</p><ul><li><p><strong>Infrastructure is key. </strong>Plentiful, well-maintained, and robust data and compute resources are essential for AI's success in science. The reports emphasise the need for investment in public research infrastructure, data sharing and open science principles.</p></li><li><p><strong>Scientists must adapt. </strong>The evolving role of AI necessitates new skills and greater AI and data literacy, including a nuanced understanding of AI's limitations and potential risks, and a desire and ability to work across disciplines.</p></li><li><p><strong>Strategic policy interventions are crucial. </strong>These include investments in infrastructure, more public-private partnership and knowledge exchange, standardised tools and methods and governance frameworks to catalyse the integration of AI into scientific workflows.&nbsp;</p></li></ul></li><li><p><strong>Looking ahead: </strong>Despite showing promise, however, many challenges remain. Responsible AI use requires a balanced approach that embraces its potential while addressing challenges such as reproducibility, transparency, and bias. To that end, the reports also caution against an overreliance on industry-led research, noting that risks could include proprietary tool lock-in, a decline in basic science research, and brain drain from public institutions.</p></li></ul><h2><strong>Sector spotlight</strong></h2><h3>AI and energy sparks debate</h3><ul><li><p><strong>What happened: </strong>The relationship between AI and energy is under the spotlight. Leopold Aschenbrenner <a href="https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf">estimated</a> that the total energy required to train and deploy AI systems may require up to 20% of US electricity production by 2030, while a new<a href="https://epochai.org/blog/how-much-does-it-cost-to-train-frontier-ai-models"> post</a> from Epoch AI proposed that "a naive extrapolation suggests that AI supercomputers will require gigawatt-scale power supply by 2029" for a single model. Meanwhile, Hugging Face <a href="https://arxiv.org/pdf/2311.16863">looked</a> at the factors driving AI energy use, and the International Energy Agency <a href="https://iea.blob.core.windows.net/assets/6b2fd954-2017-408e-bf08-952fdd62118a/Electricity2024-Analysisandforecastto2026.pdf">said</a> that in 2023 NVIDIA shipped 100,000 units that consume 7.3 TWh of electricity annually.&nbsp;</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>Predicting AI&#8217;s future energy consumption is challenging. One way of understanding the variables is to create a &#8216;<a href="https://en.wikipedia.org/wiki/Drake_equation">Drake equation</a> for energy consumption&#8217;. A rough version of this framework may contain the following factors: <strong>E</strong> (growth in energy consumption of AI) = <strong>C</strong> (annual growth rate of compute) x <strong>U</strong> (proportion of time AI systems are in use) x <strong>P</strong> (power consumption per unit of compute) x <strong>E<sub>c</sub></strong> (efficiency improvements in compute usage) x <strong>E<sub>e</sub></strong><sub> </sub>(efficiency improvements in energy use).&nbsp;&nbsp;</p></li><li><p>Taking both the IEA&#8217;s 2023 <a href="https://iea.blob.core.windows.net/assets/6b2fd954-2017-408e-bf08-952fdd62118a/Electricity2024-Analysisandforecastto2026.pdf">figure</a> &#8211; with the caveats that it is 1) global and 2) only accounts for NVIDIA chips &#8211; of 7.3 TWh for AI consumption and a total US electricity <a href="https://yearbook.enerdata.net/electricity/world-electricity-production-statistics.html">production</a> of 4,510 TWh, we can roughly estimate AI&#8217;s 2023 US energy usage to be 0.16%. Based on this figure, the growth rate needed for AI to account for 20% of US electricity production by 2030, starting from 2023, is just shy of 100% per year. Meanwhile, the annual growth rate needed for AI to account for 2% of US electricity production by 2030 is approximately 43%. (Both figures, however, assume negligible growth in energy capacity - see below.)&nbsp;</p></li><li><p>Returning to our model, we can change individual variables for a range of predictions for the annual growth rate of AI energy consumption. In the 2% scenario, the required rate of compute growth (<strong>C</strong>) would be around 35% per year if we assume a 80% contribution, 22% with a 50% contribution, and less than 9% with a 20% contribution. For 20%, however, we would require compute growth of 79% for an 80% contribution, 50% for a 50% contribution, and 20% for a 20% contribution. This back of the envelope calculation may, however, look very different based on compute efficiency savings in both compute and energy.&nbsp;&nbsp;</p></li></ul></li><li><p><strong>Looking ahead: </strong>Meeting this demand may require building new power infrastructure. Aschenbrenner&#8217;s work <a href="https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf">suggests</a> that utilities firms are already pricing in a 4.7% growth rate over the next five years, rather than the previous 2.6% figure (though he acknowledges this is far short of what he sees as required capacity increases). Finally, there is the carbon question. While it is currently unclear whether a surge in demand for energy can be met using green energy sources, it may be possible to provide the necessary power using renewables (depending on how much each factor we identify contributes to effective compute capacity).&nbsp;</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives! Subscribe for free to receive new posts and support this work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#10)]]></title><description><![CDATA[Seoul Summit, global values, and systemic safety]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-may-2024</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-may-2024</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Mon, 03 Jun 2024 13:55:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!DXH3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DXH3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DXH3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png 424w, https://substackcdn.com/image/fetch/$s_!DXH3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png 848w, https://substackcdn.com/image/fetch/$s_!DXH3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png 1272w, https://substackcdn.com/image/fetch/$s_!DXH3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DXH3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png" width="1456" height="770" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:770,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1536758,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DXH3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png 424w, https://substackcdn.com/image/fetch/$s_!DXH3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png 848w, https://substackcdn.com/image/fetch/$s_!DXH3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png 1272w, https://substackcdn.com/image/fetch/$s_!DXH3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4fcabd81-275a-499b-b80d-7a964772be0a_1498x792.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><p>In this month&#8217;s AI Policy Primer, we look at the Seoul Summit, recent research centering global values in large language models, and the UK AI Safety Institute&#8217;s new work on systemic safety. We also published an <a href="https://www.aipolicyperspectives.com/p/the-ai-policy-atlas">overview</a> of the AI policy landscape earlier this week, which introduces a 4-box model to organise the topics that we think AI policy practitioners may need to understand. As always, let us know if you have any feedback at aipolicyperspectives@google.com.</p><h2><strong>Policymakers taking action</strong></h2><h3>Seoul Summit strengthens AI safety coordination&nbsp;</h3><ul><li><p><strong>What happened: </strong>The Republic of Korea and the UK co-hosted the AI Seoul Summit. The follow-up to last year&#8217;s Summit at Bletchley Park, the event convened representatives from 28 governments (including the US &amp; China) industry, academia and civil society to discuss &#8216;three critical priorities on AI: safety, innovation and inclusivity&#8217;.&nbsp;</p></li><li><p><strong>What&#8217;s interesting: </strong>The Summit produced several outputs that pushed forward international cooperation on frontier AI safety.&nbsp;</p><ul><li><p><strong><a href="https://www.gov.uk/government/news/historic-first-as-companies-spanning-north-america-asia-europe-and-middle-east-agree-safety-commitments-on-development-of-ai">Frontier AI Safety Commitments</a>: </strong>commitments from 16 leading AI companies to publish safety frameworks (if they have not done so already) by the next Summit in France in February 2025 about how they will measure the risks of frontier models. Our recent <a href="https://deepmind.google/discover/blog/introducing-the-frontier-safety-framework/">Frontier Safety Framework</a> outlined Google DeepMind&#8217;s approach, which comes as the emerging dynamic of <a href="https://www.gov.uk/government/publications/emerging-processes-for-frontier-ai-safety/emerging-processes-for-frontier-ai-safety">responsible capability scaling</a> continues to gain traction.&nbsp;</p></li><li><p><strong><a href="https://www.gov.uk/government/news/global-leaders-agree-to-launch-first-international-network-of-ai-safety-institutes-to-boost-understanding-of-ai">International Network of AI Safety Institutes</a>: </strong>a new agreement&#8212;backed by 10 countries including the US &amp; UK in addition to the EU&#8212;to build &#8220;complementarity and interoperability&#8221; between technical work and approaches to safety to promote the safe, secure and trustworthy development of AI.&nbsp;</p></li><li><p><strong><a href="https://assets.publishing.service.gov.uk/media/66474eab4f29e1d07fadca3d/international_scientific_report_on_the_safety_of_advanced_ai_interim_report.pdf">Interim International Scientific Report on the Safety of Advanced AI</a>: </strong>a new report, loosely inspired by the Intergovernmental Panel on Climate Change, that aims to provide an independent and inclusive &#8216;state of the science&#8217; report on the capabilities and risks of frontier AI.&nbsp;</p></li></ul></li><li><p><strong>Looking ahead: </strong>The new safety frameworks published by AI labs ahead of the France Summit is likely to establish <a href="https://www.gov.uk/government/publications/emerging-processes-for-frontier-ai-safety/emerging-processes-for-frontier-ai-safety">responsible capability scaling </a>as an industry norm. This will kickstart a process of industry best practices being agreed and adopted by a critical mass of labs within the next two years,&nbsp;and may spur a wave of empirical research into scaling, safety and capabilities evaluations. We also published a <a href="https://deepmind.google/discover/blog/looking-ahead-to-the-ai-seoul-summit/">blogpost</a> with ideas about how the Summits in Seoul, France and beyond can galvanise international cooperation on frontier AI safety.&nbsp;</p></li></ul><blockquote></blockquote><h2><strong>Study watch&nbsp;&nbsp;</strong></h2><h3>Researchers eye global values&nbsp;&nbsp;</h3><ul><li><p><strong>What happened:</strong> Researchers from the University of Oxford, New York University, Meta, Cohere, and elsewhere released a <a href="https://arxiv.org/abs/2404.16019">study</a> looking at how preferences for language models differ across the world. The group compiled a database, <a href="https://huggingface.co/datasets/HannahRoseKirk/prism-alignment">PRISM</a>, which represents the end result of a large-scale experiment in which 1,500 participants from 75 countries provided details of their background, familiarity with LLMs and stated preferences for fine-grained behaviours (i.e. specific information about how they want an LLM to behave).&nbsp;</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>The work explored how different constituencies are likely to use language models. It found, for example, that older people (55+) are more likely to talk about elections and seek travel recommendations compared to younger people (18-24 years) who are more likely to discuss managing relationships or job searches.&nbsp;</p></li><li><p>The work is the latest in the long line of work exploring with which values language models ought to be aligned, while earlier this month OpenAI <a href="https://openai.com/index/introducing-the-model-spec/">released</a> its &#8216;model spec&#8217; to explain how it makes decisions about how it shapes model behaviour (e.g. how ChatGPT <a href="https://cdn.openai.com/spec/model-spec-2024-05-08.html#dont-respond-with-nsfw-content">responds</a> to NSFW requests). While developers often seek to empower users to change certain aspects of model behaviour through functions like <a href="https://help.openai.com/en/articles/8096356-custom-instructions-for-chatgpt">user instructions</a> and <a href="https://cloud.google.com/vertex-ai/generative-ai/docs/multimodal/configure-safety-attributes">custom safety filters</a>, developers are increasingly considering a &#8220;<a href="https://arxiv.org/pdf/2303.05453">personalisation within bounds</a>&#8221; model that sets overall guardrails for model behaviour while allowing for some flexibility within this boundary.&nbsp;&nbsp;&nbsp;&nbsp;</p></li></ul></li><li><p><strong>Looking ahead: </strong>In the future, we anticipate that labs will introduce personalisation tools into consumer platforms to allow users to shape behaviour on sensitive queries. If adopted, this approach will represent a significant shift away from platform policies in which the platform makes all content decisions.</p></li></ul><h2><strong>What we&#8217;re hearing&nbsp;&nbsp;</strong></h2><h3>UK launches systemic AI safety programme</h3><ul><li><p><strong>What happened: </strong>The UK government announced an &#163;8.5 million <a href="https://www.aisi.gov.uk/grants">grants programme</a> to fund research into systemic AI safety. The programme will be led by the UK AI Safety Institute (AISI) in partnership with UK Research and Innovation (UKRI) and the Alan Turing Institute.</p></li><li><p><strong>What's interesting:&nbsp;</strong></p><ul><li><p>The grants aim to broaden the AISI's remit to include 'systemic AI safety', which seeks to manage societal-level impacts of AI and help existing institutions, systems and infrastructure adapt to the diffusion of AI. As the group explains, &#8220;addressing AI's risks to people and society requires looking beyond AI models' capabilities.&#8221;&nbsp;</p></li><li><p>Potential research areas include curbing the spread of AI-generated misinformation, understanding how to adapt infrastructure and systems for a world with widespread AI usage, and generating ideas for safely deploying AI in society. The programme aims to attract proposals from researchers in academia, industry and the public sector. Those that are particularly promising may receive further funding to support their development into fuller, longer-term projects.</p></li><li><p>The move comes as governments increasingly recognise the need to proactively<em> </em>use AI to mitigate risks. The grants build on to the "defensive AI acceleration" (d/acc) concept advocated by <a href="https://vitalik.eth.limo/general/2023/11/27/techno_optimism.html">Vitalik Buterin</a> and further amplified by <a href="https://www.joinef.com/posts/introducing-def-acc-at-ef/">Matt Clifford</a>, which argues that we need to build defensive technologies to protect against AI threats (see our recent <a href="https://www.aipolicyperspectives.com/p/models-on-the-frontline-ais-defensive">post</a> on the same topic).&nbsp;</p></li></ul></li><li><p><strong>Looking ahead: </strong>As we continue to see capabilities&#8212;and the number of AI applications&#8212;grow, we expect new government initiatives seeking to use AI in prosocial ways to bolster societal infrastructure and enhance societal defences. We anticipate that the first major initiative in this vein will have links to existing government priorities such as climate change or cybersecurity.</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives. Subscribe for free to receive new posts and support this work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#9)]]></title><description><![CDATA[AISI MoU, cyber harms, and AI 'shadow use']]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-april-2024</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-april-2024</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Tue, 30 Apr 2024 13:45:18 GMT</pubDate><enclosure url="https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw"><img src="https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" width="1000" height="563" data-attrs="{&quot;src&quot;:&quot;https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:563,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;a close up of an inflatable object with a sky background&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="a close up of an inflatable object with a sky background" title="a close up of an inflatable object with a sky background" srcset="https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1692605914283-c3dd7254d513?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><p>Welcome to another monthly installment of the AI Policy Primer. As a reminder, we&#8217;re also sharing the AI Policy Primer alongside other content&#8212;such as <a href="https://www.aipolicyperspectives.com/p/the-line-between-risk-and-progress">essays</a> and <a href="https://www.aipolicyperspectives.com/p/book-review-inspectors-for-peace">book reviews</a>&#8212;on AI Policy Perspectives. If you have any feedback, let us know at aipolicyperspectives@google.com.</p><p>For this month&#8217;s Primer, we take a look at the Memorandum of Understanding signed by the US and UK AI Safety Institutes, a survey of cybersecurity practitioners to understand the real-world harms that they are witnessing from deployed AI systems, and reports on AI usage figures in science and business. &nbsp;</p><h2><strong>Policymakers taking action</strong></h2><h3>AI Safety Institutes sign MOU&nbsp;</h3><ul><li><p><strong>What happened: </strong>In April, the US and UK AI Safety Institutes (AISIs) <a href="https://www.gov.uk/government/publications/collaboration-on-the-safety-of-ai-uk-us-memorandum-of-understanding/collaboration-on-the-safety-of-ai-uk-us-memorandum-of-understanding">signed</a> a Memorandum of Understanding (MoU) to collaborate on AI safety. <strong>&nbsp;</strong>In the document, the US and UK AISIs agreed to develop a shared approach to AI model evaluations (methodologies, infrastructures and processes), collaborate on technical AI safety technical research, and&nbsp;perform at least one joint testing exercise on a publicly accessible model. They will also explore personnel exchanges and similar collaborations with other countries to manage frontier AI risks.&nbsp;&nbsp;</p></li><li><p><strong>What&#8217;s interesting: </strong></p><ul><li><p>The move comes as the EU and US also begin to collaborate more closely on AI safety. In a <a href="https://ec.europa.eu/commission/presscorner/detail/en/statement_24_1828">joint statement</a>, the EU-US Trade and Technology Council&nbsp; announced that the US AISI and EU AI Office had "briefed one another on respective approaches and mandates" and agreed to a "scientific exchange on benchmarks/risks/future trends." The organisations are also developing a joint roadmap on evaluation tools for trustworthy AI and risk management.</p></li><li><p>Canada has also launched its own AI Safety Institute. The <a href="https://www.pm.gc.ca/en/news/news-releases/2024/04/07/securing-canadas-ai">announcement</a> notes that the Canadian government is planning to dedicate $50 million &#8220;to further the safe development and deployment of AI&#8221; alongside a further $2 billion for compute and infrastructure.</p></li></ul></li><li><p><strong>Looking ahead: </strong>Since the Bletchley Summit, we have seen a multiplication of AI Safety Institutes, and it&#8217;s possible that an increasing number of governments decide to create their own dedicated institute to better understand advanced AI models. In the future, we also expect to see increased international collaboration between national safety institutes.&nbsp;</p></li></ul><blockquote></blockquote><h2><strong>Study watch&nbsp;&nbsp;</strong></h2><h3>What AI cyber harms are actually occurring?&nbsp;&nbsp;</h3><ul><li><p><strong>What happened:</strong> A new <a href="https://ojs.aaai.org/index.php/AAAI/article/view/30347">study</a>, led by Kathrin Grosse at the Swiss Federal Institute of Technology Lausanne (EPFL), surveyed cybersecurity practitioners to understand the real-world harms that they are witnessing from deployed AI systems. To date, most policy discussions about how AI might affect cybersecurity have been theoretical. This new study is a rare example of a post-deployment evaluation, with the authors surveying &gt;200 practitioners to understand the cybersecurity harms they have witnessed from AI.&nbsp;</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>AI is a double-edged sword for cybersecurity. Threat actors could potentially use AI to identify vulnerabilities, generate more persuasive phishing emails or compile malicious code. Powerful AI systems could also be the target of, and <a href="https://arxiv.org/pdf/2305.15324.pdf">perhaps one day even carry out</a>, cyberattacks. <a href="https://services.google.com/fh/files/misc/how-ai-can-reverse-defenders-dilemma.pdf">AI could also boost cybersecurity</a>, for example, if practitioners use it to write more secure code, identify anomalies, and better triage alerts. Over the longer-term, AI could potentially enable more automated protection for software by identifying vulnerabilities and generating rapid fixes. &nbsp;</p></li><li><p>The authors find that less than 5% of practitioners have witnessed real-world harms from AI, although it&#8217;s difficult to specify what constitutes an &#8216;AI&#8217; harm. This small sample size makes it challenging to extrapolate, but the data suggests 1) that attacks on data and infrastructure may be more common than attacks on models; 2) that the healthcare, automotive and security industry may be key targets; 3) that unintentional <em>accidents</em> may be a bigger near-term challenge than<em> intentional </em>attacks; and 4) that employees who feel threatened by AI systems may look to attack them (for example, by sabotaging data labelling efforts - a real-life example).&nbsp;</p></li></ul></li><li><p><strong>Looking ahead: </strong>As the study notes, there are few robust programmes to reliably track the harms that are occurring due to AI. As deployment increases, addressing this will likely require policy responses that go beyond ad-hoc surveys. This could include: more formal <a href="https://datainnovation.org/2024/04/tracking-ai-incidents-and-vulnerabilities/">post-market surveillance programmes</a>, building on early examples, such as the the <a href="https://avidml.org/">AI Vulnerability Database</a>; funding <a href="https://scholar.google.com/citations?view_op=view_citation&amp;hl=en&amp;user=q4qDvAoAAAAJ&amp;sortby=pubdate&amp;citation_for_view=q4qDvAoAAAAJ:vRqMK49ujn8C">adversarial research</a>; opening up AI models for third-party access and testing; designing programmes for AI model developers to <a href="https://arxiv.org/abs/2404.02675">responsibly report known risks</a>; and designing bug bounties for third-parties to report harms.&nbsp;</p></li></ul><h2><strong>What we&#8217;re hearing&nbsp;&nbsp;</strong></h2><h3>AI &#8216;shadow use&#8217; on the rise&nbsp;</h3><ul><li><p><strong>What happened: </strong>Researchers at Stanford University <a href="https://arxiv.org/abs/2404.01268">analysed</a> almost 1m papers published between January 2020 and February 2024 on arXiv, bioRxiv, and the Nature portfolio of journals. The group found that the use of large language models for writing research papers is on the rise across the board, with the largest and fastest growth observed in computer science papers (up to 17.5%). In contrast, the authors said that mathematics papers and the Nature portfolio showed the least LLM usage (up to 6.3%).&nbsp;</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>This forms part of a wider trend of 'shadow AI' use - i.e. individuals using AI tools in their workplace in a way that isn't formally directed/endorsed by their employer. In a 2023 <a href="https://www.nature.com/articles/d41586-023-02980-0">survey</a> of over 1,600 scientists, Nature reported that approximately 30% of researchers said that they had used generative AI tools to help write manuscripts, while a further 15% said they had used the tools to help with grant applications. On the benefits of AI, over half (55%) of researchers cited translation, a finding replicated in a <a href="https://erc.europa.eu/sites/default/files/2023-12/AI_in_science.pdf">poll</a> by the European Research Council (ERC) in 2023. With respect to risks, around 70% of researchers said that it could lead to &#8220;more reliance on pattern recognition without understanding&#8221; while a further 59% said the technology may entrench bias.&nbsp;</p></li><li><p>Science isn&#8217;t the only sector in which AI usage is on the up. In a March 2024 poll, Pew Research <a href="https://www.pewresearch.org/short-reads/2024/03/26/americans-use-of-chatgpt-is-ticking-up-but-few-trust-its-election-information/">found</a> that 43% of American adults aged 18-29 have used ChatGPT, a figure that has increased 10 percentage points since last summer. Within this group, Pew found that approximately one third (31%) have used ChatGPT for work. The figures contrast with significantly lower <a href="https://www2.census.gov/library/working-papers/2024/adrm/ces/CES-WP-24-16.pdf">figures</a> collated by businesses about how workers are using AI. A report from the U.S. Census Bureau found that, between September 2023 and February 2024, estimates of AI use rate rose from 3.7% to 5.4%. These figures are directly provided by the leadership of 1.2 million businesses to show the proportion of firms using AI within a two week period. The stats add some colour to reports in the <a href="https://aiindex.stanford.edu/wp-content/uploads/2024/04/HAI_AI-Index-Report-2024.pdf?utm_source=substack&amp;utm_medium=email&amp;tpcc=NL_Marketing">Stanford HAI Index</a>, which said that 55% of organisations in 2023 had tried to use AI in some capacity, marking a slight increase from 50% in 2022 and a significant jump from 20% in 2017.&nbsp;</p></li></ul></li></ul><p><strong>Looking ahead: </strong>Worker shadow usage may continue to increase ahead of officially reported statistics by businesses. While growth is likely to remain steady across many demographic groups, it is possible that young adults in particular will continue to play an outsized role in driving the adoption of AI for work.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives. Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#8)]]></title><description><![CDATA[AI safety institutes, open models, and biotechnology]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-march-2024</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-march-2024</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Wed, 03 Apr 2024 12:55:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!kiUz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kiUz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kiUz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png 424w, https://substackcdn.com/image/fetch/$s_!kiUz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png 848w, https://substackcdn.com/image/fetch/$s_!kiUz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png 1272w, https://substackcdn.com/image/fetch/$s_!kiUz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kiUz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png" width="1454" height="728" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:728,&quot;width&quot;:1454,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1218164,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kiUz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png 424w, https://substackcdn.com/image/fetch/$s_!kiUz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png 848w, https://substackcdn.com/image/fetch/$s_!kiUz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png 1272w, https://substackcdn.com/image/fetch/$s_!kiUz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb98e049a-ffbe-4908-aa82-9b1c7b2b6a15_1454x728.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><p>Welcome to another monthly installment of the AI Policy Primer. As a reminder, we&#8217;re also sharing the AI Policy Primer alongside other content&#8212;such as <a href="https://www.aipolicyperspectives.com/p/the-line-between-risk-and-progress">essays</a> and <a href="https://www.aipolicyperspectives.com/p/book-review-inspectors-for-peace">book reviews</a>&#8212;on AI Policy Perspectives. If you have any feedback, please do get in touch with us at aipolicyperspectives@google.com.</p><p>For this month&#8217;s edition, we have a stock-take of the various national AI safety institutes, our response to the NTIA&#8217;s request for input on open-weight models, and commentary on a new report from the Tony Blair Institute addressing biotechnology in the UK. </p><h2><strong>Policymakers taking action</strong></h2><h3>EU AI Office gets up and running&nbsp;&nbsp;</h3><ul><li><p><strong>What happened: </strong>The EU Parliament <a href="https://www.europarl.europa.eu/news/en/press-room/20240308IPR19015/artificial-intelligence-act-meps-adopt-landmark-law">approved</a> the Artificial Intelligence Act to boost &#8220;safety and compliance with fundamental rights, while boosting innovation.&#8221; The regulation, which was <a href="https://www.europarl.europa.eu/news/en/press-room/20231206IPR15699/artificial-intelligence-act-deal-on-comprehensive-rules-for-trustworthy-ai">agreed</a> in negotiations with member states in December 2023, was endorsed by MEPs with 523 votes in favour, 46 against and 49 abstentions. As part of the process of operationalising the AI Act, the EU AI Office&#8212;established in the Directorate-General for Communications Networks, Content and Technology&#8212;has begun to ramp up its operations.&nbsp;</p></li><li><p><strong>What&#8217;s interesting: </strong>The AI Office is expected to employ approximately 100 staff members in total by the end of 2025 <a href="https://eu-careers.europa.eu/en/job-opportunities/technology-specialist-artificial-intelligence/ec-2024-cnect-443651">in order to</a> &#8220;play a key role in the implementation of the new EU AI Regulation (AI Act), strengthen development and use of trustworthy AI, and foster international cooperation.&#8221; As part of this process, the new group will develop tools, methodologies, and benchmarks for evaluating capabilities for general purpose AI systems, which <a href="https://www.europarl.europa.eu/doceo/document/TA-9-2024-0138_EN.pdf">includes</a> those whose &#8220;cumulative amount of computation used for training measured in FLOPs is greater than 10^25.&#8221; The move comes as the US AI Safety Institute begins to <a href="https://venturebeat.com/ai/nist-staffers-revolt-against-potential-appointment-of-effective-altruist-ai-researcher-to-us-ai-safety-institute/">create</a> a team to conduct evaluations, and follows <a href="https://www.gov.uk/government/publications/uk-ai-safety-institute-third-progress-report/ai-safety-institute-third-progress-report">efforts</a> by the UK AI Safety Institute to build capacity.&nbsp;&nbsp;&nbsp;</p></li><li><p><strong>Looking ahead:</strong> The EU&#8217;s AI Office is likely to emerge as one of the three most important new institutions in conjunction with the UK AISI and US AISI (which recently <a href="https://www.ft.com/barrier/corporate/5a2ab953-4e02-440c-8708-b126128866e5">signed</a> a partnership agreement). Like the institutes in the US and the UK, the group will continue to accelerate <a href="https://www.reuters.com/technology/ai-talent-war-heats-up-europe-2024-03-11/">efforts</a> to hire technical researchers and policy specialists.  </p></li></ul><blockquote></blockquote><h2><strong>Policymakers taking action&nbsp;</strong></h2><h2>NTIA solicits comments on open-weight models</h2><ul><li><p><strong>What happened: </strong>This month, the US (NTIA) ran a consultation on the &#8220;risks of openly available model weights,&#8221; as directed by the Executive Order on AI. Google DeepMind partnered with Google to submit a <a href="https://www.linkedin.com/feed/update/urn:li:activity:7180888388177154048/">response</a> making the case that, though we have long been supporters of open science, we recognise that open models can pose risks (and releasing them is irreversible). We also proposed that openness is not a binary, and that a more useful frame is &#8220;access&#8221; to the right capabilities for the right purposes.</p></li><li><p><strong>What&#8217;s interesting: </strong>Many parties are grappling with the question of how to assess the risks posed by open models. A <a href="https://crfm.stanford.edu/open-fms/paper.pdf">recent paper</a> from Stanford researchers, for example, made the case for focusing on <a href="https://hai.stanford.edu/sites/default/files/2024-03/Response-NTIA-RFC-Open-Foundation-Models.pdf">marginal</a><em><a href="https://hai.stanford.edu/sites/default/files/2024-03/Response-NTIA-RFC-Open-Foundation-Models.pdf"> </a></em><a href="https://hai.stanford.edu/sites/default/files/2024-03/Response-NTIA-RFC-Open-Foundation-Models.pdf">risk</a> (i.e. the extent to which open models represent a greater risk relative to their closed counterparts or existing digital technologies). Our ability to set more granular thresholds for when open models may be too risky to release will require making much more progress on safety evaluations. For this reason, we proposed that governments can help develop recommendations and best practices to help set thresholds for risks, drive progress on evaluations, and identify potential procedural requirements for open models release. Google DeepMind also recently released its own set of open models, <a href="https://opensource.googleblog.com/2024/02/building-open-models-responsibly-gemini-era.html">Gemma</a>, which was based on a set of safety and responsibility best practices. </p></li></ul><ul><li><p><strong>Looking ahead: </strong>The debate around open models will remain highly political given it exists at the intersection of concerns over competition and AI safety discussions. National security will continue to feature prominently. In parallel to discussions about &#8220;frontier&#8221; models, we may see requirements for developers who are considering releasing the weights of sophisticated models.&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</p></li></ul><h2><strong>What we&#8217;re hearing&nbsp;</strong></h2><h3>Harnessing the benefits of biotechnology&nbsp;</h3><ul><li><p><strong>What happened: </strong>We hosted the Tony Blair Institute (TBI) for the launch of its report, &#8220;<a href="https://www.institute.global/insights/politics-and-governance/a-new-national-purpose-leading-the-biotech-revolution">A New National Purpose: Leading the Biotech Revolution</a>&#8221;, which proposes policies to help the UK harness the benefits of advances in biotechnology. Benedict Macon-Cooney, chief policy strategist at the TBI,&nbsp;was joined by Sir Sajid Javid (Former Secretary of State for Health and Social Care), Sarah Korman (Isomorphic Labs&#8217; general counsel) and Hans Bishop (president of Altos Labs)<strong> </strong>to discuss how policymakers should react to a moment of rapid technological progress.&nbsp;</p></li><li><p><strong>What&#8217;s interesting: </strong></p><ul><li><p>The core recommendation in the report is the creation of a UK Laboratory of Biodesign to bring together scientists and bioengineers under one roof to focus on interdisciplinary research. This institution would, according to the report, &#8220;focus on the invention of new biotechnology that is at too early a stage for commercial investors.&#8221; The paper&#8217;s central argument is that biotechnology represents a major economic opportunity that can be realised through building and scaling globally competitive biotechnology firms. </p></li><li><p>These firms, in conjunction with the UK Laboratory of Biodesign, would benefit from network effects that can be harnessed to power the UK&#8217;s <a href="https://www.lse.ac.uk/european-institute/events/europe-at-lse/2023-24/AT/The-Knowledge-Economy-A-New-Conceptualisation-and-Index-for-Comparative-Research">knowledge economy</a>. It also identifies hurdles&#8212;and proposes solutions&#8212;to realising biotechnology&#8217;s potential, with the introduction of a new NHS-led data trust proposed to solve bottlenecks in the availability of high quality data. Finally, to address novel risks posed by the development of biotechnology, the report suggests that the Laboratory of Biodesign should deliver biosecurity advice to the government alongside a new UK Biosecurity Taskforce.</p></li></ul></li><li><p><strong>Looking ahead: </strong>There is increasing interest in bioengineering and life sciences as areas of strategic advantage for the UK. As a result, it is possible that governments may begin to explore new data sharing frameworks to securely release data for experimentation with AI in the next few years.</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives. Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#7)]]></title><description><![CDATA[Cybersecurity threat assessments, agriculture & data policies]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-primer-february-2024</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-primer-february-2024</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Wed, 14 Feb 2024 13:48:25 GMT</pubDate><enclosure url="https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw"><img src="https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" width="1000" height="563" data-attrs="{&quot;src&quot;:&quot;https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:563,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;a sculpture of a horse covered in green grass&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="a sculpture of a horse covered in green grass" title="a sculpture of a horse covered in green grass" srcset="https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1692607038909-2f33929144cc?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><p>We&#8217;re back with our monthly roundup of AI policy news, now rebranded as the AI Policy Primer. We&#8217;ll be sharing the Primer alongside other content&#8212;such as <a href="https://www.aipolicyperspectives.com/p/the-line-between-risk-and-progress">essays</a> and <a href="https://www.aipolicyperspectives.com/p/book-review-inspectors-for-peace">book reviews</a>&#8212;regularly in the coming weeks and months on AI Policy Perspectives. </p><p>For this month&#8217;s Primer, we have an assessment of the cyber threat posed by AI from the UK&#8217;s National Cyber Security Centre, a look at AI&#8217;s use in agriculture, and a rundown of recent discussions focusing on access to training data. </p><h2><strong>Policymakers taking action</strong></h2><h3>Near term AI cyber threat &#8216;evolution not revolution&#8217;</h3><ul><li><p><strong>What happened: </strong>The UK&#8217;s National Cyber Security Centre (NCSC) <a href="https://www.ncsc.gov.uk/report/impact-of-ai-on-cyber-threat">published its assessment</a> of the cyber threat from AI over the next two years. NCSC&#8217;s assessment uses the UK intelligence community&#8217;s formal probabilistic language (see <a href="https://assets.publishing.service.gov.uk/media/6421b6a43d885d000fdadb70/2019-01_PHIA_PDF_First_Edition_Electronic_Distribution_v1.1__1_.pdf">yardstick on p.29</a>) to conclude that AI will &#8220;almost certainly&#8221; increase the number and impact of cyber attacks. It notes, however, that the threat comes primarily from &#8220;evolution and enhancement&#8221; of existing techniques and approaches - and, as NCSC CEO Lindy Cameron surmised, &#8220;does not transform the risk landscape in the near term.&#8221; The NCSC expects the impacts to 2025 will include:</p><ul><li><p>More convincing &#8216;social engineering&#8217; attacks and information gathering capabilities - think fewer typos and more compelling prose in phishing emails - which will boost less sophisticated cyber criminals. The NCSC judges this &#8220;will likely&#8221; also contribute to the global ransomware threat.</p></li><li><p>More sophisticated uses of AI in cyber attacks, such as malware development and vulnerability research, &#8220;will continue to rely on human expertise&#8221; and are therefore &#8220;highly likely to be restricted to threat actors with access to quality training data, significant expertise...and resources&#8221;. This refers to highly capable state actors and some established (and capable) criminal groups.</p></li></ul></li><li><p><strong>What&#8217;s interesting: </strong>The cyber risks from AI steadily attracted policymaker attention in 2023, including most prominently at the UK&#8217;s AI Safety Summit where risks to cybersecurity featured heavily alongside biosecurity concerns. But as with many other areas of potential AI risk, there are a range of views on what exactly the threat landscape looks like, how we should allocate attention across current and future risks, and how imminent <a href="https://arxiv.org/abs/2401.05566">truly novel</a> risks are. This assessment from the NCSC gives its best effort response to some of these questions. How it compares to the one made in the forthcoming&nbsp; <a href="https://www.gov.uk/government/publications/ai-safety-summit-2023-chairs-statement-state-of-the-science-2-november/state-of-the-science-report-to-understand-capabilities-and-risks-of-frontier-ai-statement-by-the-chair-2-november-2023">&#8216;State of the Science&#8217; report</a> in May, commissioned at the UK AI Safety Summit, will be one to watch. The NCSC report is framed as further evidence of momentum following the UK Summit - and follows the UK&#8217;s publication of the first global <a href="https://www.ncsc.gov.uk/collection/guidelines-secure-ai-system-development">guidelines</a> on secure AI development, endorsed by 18 countries including the US, in late 2023.</p></li><li><p><strong>Looking ahead: </strong>Made easier by well-established security alliances, cybersecurity and AI may prove to be a bright spot for international collaboration on AI governance in 2024, including at the South Korea and France Safety Summits. Watch for new international R&amp;D collaborations on using AI for cyber defence and more formal information sharing agreements between allies on emerging cyber risks. It is also possible that the cyber conversation focuses more on <a href="https://arxiv.org/pdf/2401.03315.pdf">the use of open source models</a> by malicious actors.</p></li></ul><blockquote></blockquote><h2><strong>Sector spotlight</strong></h2><h3>Agricultural AI ploughs ahead</h3><ul><li><p><strong>What happened: </strong>AI is being <a href="https://arxiv.org/abs/2401.06171">used</a> by farmers around the world to enable precision farming, crop monitoring, and climate-resilient agricultural practices. The technology is also being deployed to measure soil health, which the Ecological Society of America <a href="https://www.esa.org/esa/wp-content/uploads/2012/12/carbonsequestrationinsoils.pdf">says</a> contains around 75% of all carbon stored on land, by underpinning the creation of &#8216;digital twins&#8217; of farmland to quantify sequestration (long term storage of carbon in oceans, soils, vegetation, and geologic formations). A recent <a href="https://www.polarismarketresearch.com/industry-analysis/artificial-intelligence-in-agriculture-market">estimate</a> put the global artificial intelligence and agriculture market size at $1.44 billion, predicting that the sector will generate an estimated revenue of around $12 billion by 2032.&nbsp;</p></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>According to the Food and Agriculture Organization of the United Nations (FAO), almost half of the Earth&#8217;s population <a href="https://www.fao.org/3/cc4337en/cc4337en.pdf">lives</a> in households that are &#8220;linked&#8221; to livelihoods dependent on agrifood systems. While only about 3% of all employment in high-income countries is typically in the agricultural sector, the figure can <a href="https://data.worldbank.org/indicator/SL.AGR.EMPL.ZS?most_recent_value_desc=true">reach</a> as high as 85% in some countries. However, while AI can be used to boost yields and minimise loss, it also <a href="https://www.omfif.org/2023/11/ai-in-agriculture-calls-for-imaginative-policy-making/">risks</a> consolidating power in the hands of a small number of farming groups and creating labour displacement effects that fall disproportionately on low and middle-income countries. Additionally, its success is likely to be <a href="https://arxiv.org/abs/2401.06171">contingent</a> on the provision of technological infrastructure, measures to boost data accessibility, and efforts to close skill gaps.</p></li><li><p>Our protein folding system, AlphaFold, has been used in research related to crops, plants, and agriculture. For example, it has been used to study <a href="https://pubmed.ncbi.nlm.nih.gov/35915586/">potato blight</a>, the plant pathogen <a href="https://nph.onlinelibrary.wiley.com/doi/epdf/10.1111/nph.18378">white blister rust</a>, and the growth of <a href="https://pubmed.ncbi.nlm.nih.gov/34947066/">rice blast fungus</a>. Google DeepMind has a number of additional former and current projects in this space, from historical <a href="https://deepmind.google/discover/blog/using-machine-learning-to-accelerate-ecological-research/">efforts</a> to study the impact of poaching, climate abnormalities, and agriculture on animal behaviour to the GraphCast model that <a href="https://www.science.org/stoken/author-tokens/ST-1550/full">provides</a> faster and more accurate global weather forecasting.</p></li></ul></li><li><p><strong>Looking ahead: </strong>AI&#8217;s use in agriculture may primarily be <a href="https://www.polarismarketresearch.com/industry-analysis/artificial-intelligence-in-agriculture-market">driven</a> by the United States and Europe in the near term, which could mitigate its immediate impact on employment in the agricultural sectors of low and middle-income countries. Over the long term, however, a core policy global challenge will be to ensure that productivity gains in these geographies are realised in a way that protects livelihoods connected to the agrifood sector.&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</p></li></ul><h2><strong>Issue spotlight</strong></h2><h3>Policy discussions focus on data</h3><ul><li><p><strong>What happened:&nbsp;</strong></p><ul><li><p>Data has become one of the focal points in AI policy discussions. Developers consider the availability of high quality data a prerequisite for increases in capability, while policymakers are increasingly looking to regulate specific types of data that are used to train large models (e.g. copyrighted data or personal data). The <a href="https://www.consilium.europa.eu/en/press/press-releases/2023/12/09/artificial-intelligence-act-council-and-parliament-strike-a-deal-on-the-first-worldwide-rules-for-ai/">recently-agreed</a> EU AI Act requires the developers of &#8220;general-purpose AI systems&#8221; to provide high level disclosures of copyrighted content used for their training.&nbsp;</p></li><li><p>Meanwhile, a number of high profile lawsuits have emerged in which creators of certain types of content (such as news publishers) argue for compensation for the use of their data to train large models. They <a href="https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html">suggest</a> that their proprietary data is particularly important to the usability and performance of certain AI systems, or is particularly sought-after by its users.&nbsp;</p></li></ul></li><li><p><strong>What&#8217;s interesting:&nbsp;</strong></p><ul><li><p>However, the extent to which certain data sources elicit particular capabilities is unclear. Recent <a href="https://arxiv.org/abs/2401.06751">research</a> proposes that LLMs trained on &#8220;easy&#8221; data (for example, a dataset of grade-school subject questions) perform well on &#8220;hard&#8221; data tasks (for example, graduate level STEM questions). They demonstrate that, surprisingly, models can learn to solve complex problems by training on easily-obtained, simpler data.&nbsp;</p></li><li><p>The paper suggests that models may not actually need large datasets of specialised &#8211;&nbsp;and often copyrighted &#8211; content to reach high performance. Given that &#8216;hard&#8217; datasets tend to be restricted and expensive, the dynamic has implications for the ability of different actors to train capable models. It may also diminish the importance of providers of highly specialised information, which has recently been drawn into focus by the <a href="https://arxiv.org/abs/2305.05862">use</a> of prompting regimes to enable general models to surpass those trained on proprietary data sources.&nbsp;<strong>&nbsp;</strong></p></li></ul></li><li><p><strong>Looking ahead: </strong>The debate will continue through legislative action and in the courts, with parties taking hard stances about whether interventions are best focused at the level of inputs (e.g. hard restrictions on models training on certain types of data) or outputs (e.g. obligations to apply certain types of filters).</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#6)]]></title><description><![CDATA[The next AI Safety Summit, US Executive Order, and policy discovery]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-perspectives-november-2023</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-perspectives-november-2023</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Thu, 30 Nov 2023 12:18:29 GMT</pubDate><enclosure url="https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw"><img src="https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D" width="1000" height="563" data-attrs="{&quot;src&quot;:&quot;https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:563,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;a group of colorful sculptures sitting on top of a table&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="a group of colorful sculptures sitting on top of a table" title="a group of colorful sculptures sitting on top of a table" srcset="https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 424w, https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 848w, https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1272w, https://images.unsplash.com/photo-1692607334484-40e695e364e6?q=80&amp;w=1000&amp;auto=format&amp;fit=crop&amp;ixlib=rb-4.0.3&amp;ixid=M3wxMjA3fDB8MHxwaG90by1wYWdlfHx8fGVufDB8fHx8fA%3D%3D 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualising AI by Google DeepMind</figcaption></figure></div><p>We&#8217;re back with another edition of AI Policy Perspectives: a monthly rundown of the topics that Google DeepMind&#8217;s policy team has been reading, thinking about and working on over the past few weeks.</p><p>This month, we have preparations for the next AI Safety Summit in South Korea, reflections on the US Executive Order, and outputs from Google DeepMind&#8217;s recent policy discovery programme delivered in partnership with civil society. </p><h2>Policymakers taking action </h2><h3>South Korea prepares to host next AI Safety Summit</h3><ul><li><p><strong>What happened: </strong>Preparations are <a href="https://www.gov.uk/government/news/landmark-sci-tech-deal-with-the-republic-of-korea-to-boost-cooperation-in-critical-technologies-such-as-ai-and-semiconductors?utm_medium=email&amp;utm_campaign=govuk-notifications-topic&amp;utm_source=a9e22eaa-180d-4c6f-a306-f9e04e2d2f08&amp;utm_content=immediately">underway</a> ahead of the next AI Safety Summit, which will be co-hosted by the Republic of Korea and the UK next year. The event, which will take place virtually, is <a href="https://bfpg.co.uk/2023/11/ai-safety-summit-a-new-era-for-global-tech-governance/">expected</a> to focus on the development of frameworks, guidelines, and policies connected to elements of the U.S. Executive Order, the EU&#8217;s AI Act, and the G7 principles. The Carnegie Endowment for International Peace <a href="https://carnegieendowment.org/2023/11/09/uk-ai-safety-summit-opened-new-chapter-in-ai-diplomacy-pub-90968">speculated</a> that the summit will &#8220;include how to gauge increases in AI model capabilities, as well as institutional design problems affecting the world&#8217;s capacity to spread access to frontier-level AI technology without increasing risks of misuse.&#8221;&nbsp;</p></li><li><p><strong>What&#8217;s interesting: </strong>The preparations come after the inaugural UK Safety Summit culminated in the &#8216;Bletchley declaration&#8217;, an <a href="https://www.gov.uk/government/publications/ai-safety-summit-2023-the-bletchley-declaration/the-bletchley-declaration-by-countries-attending-the-ai-safety-summit-1-2-november-2023">agreement</a> from 28 states to work together on safety standards to maximise the upside and minimise the risks posed by frontier AI systems. Amidst two days of workshops, keynotes, and demos, the US Secretary of Commerce Gina Raimondo used the Summit as an opportunity to <a href="https://www.commerce.gov/news/speeches/2023/11/remarks-commerce-secretary-gina-raimondo-ai-safety-summit-2023-bletchley">highlight</a> new policy interventions from the Biden administration, while Chinese Vice Minister Wu Zhaohui urged attendees to &#8220;ensure AI always remains under human control&#8221; and that governments should work to &#8220;build trustworthy AI technologies that can be monitored and traced.&#8221;</p></li><li><p><strong>Looking ahead:</strong> The South Korean summit&#8217;s most significant contribution may prove to be the State of the Science <a href="https://www.gov.uk/government/publications/ai-safety-summit-2023-chairs-statement-state-of-the-science-2-november/state-of-the-science-report-to-understand-capabilities-and-risks-of-frontier-ai-statement-by-the-chair-2-november-2023">report</a>, an effort led by Yoshua Bengio to identify emerging risks associated with frontier AI. </p></li></ul><h2>Policymakers taking action </h2><h3><strong>Executive Order reshapes US AI policy landscape</strong></h3><ul><li><p><strong>What happened: </strong>On 30 October the Biden Administration released its long-anticipated <a href="https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/">Executive Order</a> (EO) on artificial intelligence. The EO builds on other actions taken by the Biden administration on AI, including the White House <a href="https://www.whitehouse.gov/ostp/ai-bill-of-rights/">Blueprint for an AI Bill of Rights</a>, the National Institute for Standards and Technology (NIST) <a href="https://nvlpubs.nist.gov/nistpubs/ai/NIST.AI.100-1.pdf">Risk Management Framework</a>, and the <a href="https://www.whitehouse.gov/briefing-room/statements-releases/2023/07/21/fact-sheet-biden-harris-administration-secures-voluntary-commitments-from-leading-artificial-intelligence-companies-to-manage-the-risks-posed-by-ai/">voluntary White House commitments</a> made by leading AI companies.</p></li><li><p><strong>What&#8217;s interesting: </strong>The comprehensive EO touches a broad range of federal agencies and AI issues, from workforce development to support for research, to sectors spanning healthcare, education, energy and others. Additionally, the EO establishes a new interagency White House AI Council which will be responsible for coordinating AI-related policy, including implementation of the EO. It also gives the Department of Commerce and NIST a leading role in implementation and tasks the White House Office of Management and Budget (OMB) with formulating guidance for federal agencies&#8217; use and procurement of AI.&nbsp;</p></li><li><p>Other noteworthy provisions include reporting requirements for developers of &#8220;dual-use foundation models&#8221; using a certain compute threshold (greater than 10^26 flops for general models). Commerce will be providing more details about the definition of "dual-use foundation models" as well as what such requirements will look like. Additionally, the EO introduces requirements for US cloud service providers to report when a foreign person or reseller transacts to train a large AI model that could be used for malicious purposes.&nbsp;&nbsp;</p></li><li><p><strong>Looking ahead:</strong>The EO does not need to be passed into law to go into effect, and with a divided Congress and prospects for major AI legislation uncertain, it is likely to represent the primary instrument for US AI regulation in the near-term.</p></li></ul><h2>What we&#8217;re hearing</h2><h3>Civil society groups drive policy discovery</h3><ul><li><p><strong>What happened:</strong> Throughout 2023, we heard from a broad range of groups calling for policies like equitable data practices, upskilling efforts, and measures to build trust and enable participation in AI development. To surface these policies, we co-authored a new <a href="https://static1.squarespace.com/static/652479cdfe87ba58b3f5392a/t/654131d140c3056dd933c87d/1698771431230/The+Changing+Landscape+of+AI%3A+Lessons+from+a+Year+of+Policy+Discovery">report</a> with civil society organisations that summarised dialogues with a global set of participants from academia, governments, start-ups and the private sector, including those with experience in communities and sectors that will be most affected by the deployment of AI systems. The programme built on our work with the Aspen Institute, &#8216;<a href="https://www.aspeninstitute.org/publications/blueprint-for-equitable-ai/">A Blueprint for Equitable AI</a>,&#8217; which highlighted the need to encourage democratic dialogue about how AI might be built, used, and governed.</p></li><li><p><strong>What&#8217;s interesting: </strong>AI labs are experimenting with methodologies like <a href="https://cip.org/alignmentassemblies">citizens assemblies</a> and <a href="https://www.wired.com/story/meta-ran-a-giant-experiment-in-governance-now-its-turning-to-ai/">community fora</a> to incorporate public input into the AI development process. Private sector participatory efforts, however, come with a host of <a href="https://arxiv.org/abs/2306.09871">challenges</a>: power imbalances, information asymmetries, a lack of shared definitions and competing or contradictory goals. For these reasons, we partnered with civil society organisations to lead the creation of discussion agendas, recruitment of participants, and development of pre-reading materials. Many of the reports include lessons for improving how governments, civil society and the private sector might work together toward ensuring equitable AI outcomes.&nbsp;</p></li><li><p><strong>Looking ahead: </strong>AI developers should strive to make sure that their models are reflective of and responsive to the rest of the AI ecosystem and the world beyond it. To understand some of our work in this space, read a summary of insights from the programme in the report &#8216;<a href="https://static1.squarespace.com/static/652479cdfe87ba58b3f5392a/t/654131d140c3056dd933c87d/1698771431230/The+Changing+Landscape+of+AI%3A+Lessons+from+a+Year+of+Policy+Discovery">The Changing Landscape of AI: Lessons From a Year of Policy Discovery</a>&#8217;. Additionally, each of the organisations we worked with produced their own reports of the individual roundtable discussions, which can be found <a href="https://www.aipolicydiscovery.co.uk/roundtables">here</a>.</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Policy Perspectives ! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Policy Primer (#6)]]></title><description><![CDATA[The Hiroshima process, scientists' views on AI, and the UK AI Safety Summit]]></description><link>https://www.aipolicyperspectives.com/p/ai-policy-perspectives-october-2023</link><guid isPermaLink="false">https://www.aipolicyperspectives.com/p/ai-policy-perspectives-october-2023</guid><dc:creator><![CDATA[AI Policy Perspectives]]></dc:creator><pubDate>Thu, 02 Nov 2023 10:30:04 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/cbebb122-f827-4f1d-ab58-d4e1a1e9142c_3840x2160.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Welcome to the latest edition of AI Policy Perspectives: a rundown of the topics that Google DeepMind&#8217;s policy team has been reading, thinking about and working on over the past few weeks.<br><br>This month's version looks at the G7&#8217;s International Code of Conduct, a recent study examining how scientists are using AI, and new approaches to evaluations as the AI Safety Summit gets underway.&nbsp;<br><br>As ever, feedback and questions are very welcome. We&#8217;re planning a few updates to AI Policy Perspectives in the future, so make sure to watch this space.</p><h1><strong>Policymakers taking action</strong></h1><h2>G7 announces International Code of Conduct</h2><ul><li><p><strong>What happened: </strong>The G7 this week <a href="https://ec.europa.eu/commission/presscorner/detail/en/ip_23_5379">announced</a> an International Code of Conduct for Organisations Developing Advanced AI Systems as a result of its &#8220;Hiroshima Process.&#8221; The principles are largely modelled on previously agreed commitments including those companies made at the White House in July &#8211;&nbsp;including measures to limit misuse; invest in cybersecurity; and identify vulnerabilities, including through red-teaming. Additional principles focus on sharing risk management policies and implementing controls on models&#8217; data inputs and outputs.</p></li><li><p><strong>What&#8217;s interesting: </strong>The nascency of AI policy is enabling fast progress: from a <a href="https://www.whitehouse.gov/briefing-room/statements-releases/2023/05/04/readout-of-white-house-meeting-with-ceos-on-advancing-responsible-artificial-intelligence-innovation/">meeting</a> with four CEOs at the White House in May, to <a href="https://www.whitehouse.gov/briefing-room/statements-releases/2023/07/21/fact-sheet-biden-harris-administration-secures-voluntary-commitments-from-leading-artificial-intelligence-companies-to-manage-the-risks-posed-by-ai/">company commitments</a> in July to G7 leaders agreeing an international set of principles by October. That's leaving aside the extensive Executive Order that the White House also <a href="https://www.politico.com/news/2023/10/27/white-house-ai-executive-order-00124067">published this week</a>. Amidst the supposed demise of multilateralism, the G7 process provides a welcome reminder that governments can collaborate constructively on shared global issues.</p></li><li><p>Developing interoperable, international frameworks for AI safety is one of our policy priorities, and we&#8217;ve been supportive of more international backing for the practices we committed to the White House. As other fora also consider these topics &#8211;&nbsp;such as the upcoming UK AI Safety Summit and discussions at the UN &#8211;&nbsp;we will continue sharing our perspective on things like <a href="https://deepmind.google/discover/blog/exploring-institutions-for-global-ai-governance/">potential new institutions</a>.</p></li><li><p><strong>Looking ahead: </strong>We could see the OECD and upcoming Italian G7 Presidency collaborate closely in 2024 to operationalise the principles in alignment with the proposed EU AI Act, for which trilogue negotiations are still ongoing.&nbsp;</p></li></ul><h1><strong>Study watch</strong></h1><h2>Scientists share hopes and concerns about AI</h2><ul><li><p><strong>What happened: </strong>Two recent Nature surveys shed light on how <a href="https://www.nature.com/articles/d41586-023-02980-0">scientists</a> and <a href="https://www.nature.com/articles/d41586-023-03235-8">postdocs</a> are using AI, including various generative AI tools.</p></li><li><p><strong>What&#8217;s interesting: </strong>Scientists&#8217; use of AI is growing relatively quickly, although it is not yet transforming most practitioners&#8217; research. While <a href="https://www2.deloitte.com/content/dam/Deloitte/uk/Documents/technology-media-telecommunications/deloitte-uk-digital-consumer-trends-2023-deck.pdf">8% of the UK public</a> use generative AI tools at work, a third (31%) of postdocs do so, mainly to refine their writing, debug code, adapt content for different formats (e.g. LaTex), and summarise literature. Looking ahead, scientists hope that using machine learning in their research will enable them to speed up data processing, use new kinds of data, and tackle otherwise prohibitive research problems. Smaller shares of scientists expect AI to directly generate new research hypotheses or make new discoveries.&nbsp;</p></li><li><p>Scientists also worry that GenAI tools will lead to more misinformation, fraud, and inaccuracies, and compound issues relating to biases in datasets - although AI may also help address some of these challenges, such as detecting fraudulent images in papers.&nbsp; When it comes to using machine learning in their research, scientists worry that over-reliance on pattern recognition may come at the expense of deeper understanding. Although new <a href="https://arxiv.org/abs/2310.16410">interpretability research</a> focussed on the chess playing AI system AlphaZero, also suggests that we may one day be able to expand human knowledge by studying how AI systems learn.&nbsp;</p></li><li><p>When asked about obstacles to using AI in their work, scientists placed the lack of skills, training, and funding above a lack of compute and data, but also worried that only a small number of companies and universities could operate at the cutting edge.&nbsp;</p></li></ul><ul><li><p><strong>Looking ahead:</strong> In the next 1-2 years, scientists could be among a group of professions, including programmers, whose use of AI will outpace that of the broader labour force. This may provide early insights into the broader benefits and risks of AI.&nbsp;</p></li></ul><h1><strong>What we&#8217;re thinking about&nbsp;</strong></h1><h2>The UK AI Safety Summit </h2><ul><li><p><strong>What happened:</strong> At the UK AI Safety Summit, which began yesterday, evaluations and &#8220;effective model risk assessments&#8221; will be a priority <a href="https://www.gov.uk/government/publications/ai-safety-summit-programme/ai-safety-summit-day-1-and-2-programme">discussion item</a> on day one. In the run up to the Summit, the Frontier Model Task Force announced that it was <a href="https://www.gov.uk/government/publications/frontier-ai-taskforce-second-progress-report/frontier-ai-taskforce-second-progress-report">partnering</a> with Humane Technology&#8217;s Rumman Chowdhury to expand its capacity to evaluate societal impacts from AI.&nbsp;</p></li><li><p><strong>What&#8217;s interesting: </strong>Our recent <a href="https://arxiv.org/pdf/2310.11986.pdf">paper</a> identifies three main types of sociotechnical evaluations of AI safety risks: (a) those that assess a model's capabilities; (b) those that assess risks stemming from how people interact with an AI model; and (c) those that evaluate longer-term societal effects, such as employment or environmental effects, as AI becomes more widely used across society.&nbsp;</p></li><li><p>The paper surveys the current state of AI evaluations and identifies gaps, particularly for non-text modalities, human-AI interaction and societal impact evaluations. This tallies with the efforts of the Frontier Model Task Force, which, although focussed primarily on assessing AI models&#8217; dangerous capabilities, recently <a href="https://www.gov.uk/government/publications/frontier-ai-taskforce-first-progress-report/frontier-ai-taskforce-first-progress-report">partnered</a> with the Collective Intelligence Project to conduct social evaluations of powerful models.&nbsp;</p></li><li><p>Evaluations are <a href="https://www.anthropic.com/index/evaluating-ai-systems">challenging</a> to conduct. Arvind Narayanan and Sayash Kapoor recently <a href="https://www.cs.princeton.edu/~arvindn/talks/evaluating_llms_minefield/">described</a> evaluating LLMs as a &#8216;minefield&#8217; due to what they characterised as prompt sensitivity - results depending on prompts rather than model properties;&nbsp; construct validity - failing to adequately model the real world; and data contamination - when training data isn&#8217;t properly separated from test data.&nbsp; This wide range of challenges highlights how policymakers, AI labs and civil society groups will need to significantly step up AI evaluation efforts and co-design new evaluations approaches.&nbsp;</p></li><li><p><strong>Looking ahead:</strong> In the next year, the number of society-focused AI evaluations may increase, but will remain vastly under-studied relative to capability-focused approaches.</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.aipolicyperspectives.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>