The venue will be picturesque: a 19th-century pile north of London that during the second world war was home to Alan Turing, his code-breaking crew and the first programmable digital computer. The attendees will be an elite bunch of 100 world leaders and tech executives. And the question they will strive to answer is epochal: how to ensure that artificial intelligence neither becomes a tool of unchecked malfeasance nor turns against humanity.
The “AI Safety Summit”, which the British government is hosting on November 1st and 2nd at Bletchley Park, appears destined for the history books. And it may indeed one day be seen as the first time global power-brokers sat down to discuss seriously what to do about a technology that may change the world. As Jonathan Black, one of the organisers, observed, in contrast with other big policy debates, such as climate change, “there is a lot of political will to do something, but it is not clear what.”
Efforts to rein in AI abound. Negotiations in Brussels entered a pivotal stage on October 25th as officials grappled to finalise the European Union’s ambitious AI act by the end of the year. In the days leading up to Britain’s summit or shortly thereafter, the White House is expected to issue an executive order on AI. The G7 club of rich democracies will this autumn start drafting a code of conduct for AI firms. China, for its part, on October 18th unveiled a “Global AI Governance Initiative”.
The momentum stems from an unusual political economy. Incentives to act, and act together, are strong. For starters, AI is truly a global technology. Large language models (LLMs), which power eerily humanlike services such as ChatGPT, travel easily. Some can be run on a laptop. It is of little use to tighten the screws on AI in some countries if they remain loose in others. Voters may be in favour. More than half of Americans “are more concerned than excited” about the use of AI, according to polling by the Pew Research Centre.
The Beijing effect
Regulatory rivalry is adding more urgency. Europe’s AI act is intended in part to cement the bloc’s role as the setter of global digital standards. The White House would love to forestall such a “Brussels effect”. Neither the EU nor America wants to be outdone by China, which has already adopted several AI laws. They were cross with the British government for inviting China to the summit—never mind that without it, any regulatory regime would not be truly global. (China may actually show up, even if its interest is less to protect humanity than the Communist Party.)
Another driver of AI-rulemaking diplomacy is even more surprising: the model-makers themselves. In the past the technology industry mostly opposed regulation. Now giants such as Alphabet and Microsoft, and AI darlings like Anthropic and OpenAI, which created ChatGPT, lobby for it. Companies fret that unbridled competition will push them to act recklessly by releasing models that could easily be abused or start developing minds of their own. That would really land them in hot water.
The will to act is there, in other words. What is not there is “anything approaching consensus as to what the problems are that we need to govern, let alone how it is that we ought to govern them”, says Henry Farrell of Johns Hopkins University. Three debates stand out. What should the world worry about? What should any rules target? And how should they be enforced?
Start with the goals of regulation. These are hard to set because AI is evolving rapidly. Hardly a day passes without a startup coming up with something new. Even the developers of LLMs cannot say for sure what capabilities these will exhibit. This makes it crucial to have tests that can gauge how risky they might be—something that is still more art than science. Without such “evals” (short for evaluations), it will be hard to check whether a model is complying with any rules.
Tech companies may back regulation, but want it to be narrow and target only extreme risks. At a Senate hearing in Washington in July, Dario Amodei, Anthropic’s chief executive, warned that AI models will in a few years be able to provide all the information needed to build bioweapons, enabling “many more actors to carry out large scale biological attacks”. Similar dire forecasts are being made about cyber-weapons. Earlier this month Gary Gensler, chairman of America’s Securities and Exchange Commission, predicted that an AI-engineered financial crisis was “nearly unavoidable” without swift intervention.
Others argue that these speculative risks distract from other threats, such as undermining the democratic process. At an earlier Senate hearing Gary Marcus, a noted AI sceptic, opened his testimony with a snippet of breaking news written by GPT-4, OpenAI’s top model. It convincingly alleged that parts of Congress were “secretly manipulated by extraterrestrial entities”. “We should all be deeply worried,” Mr Marcus argued, “about systems that can fluently confabulate.”
The debate over what exactly to regulate will be no easier to resolve. Tech firms mostly suggest limiting scrutiny to the most powerful “frontier” models. Microsoft, among others, has called for a licensing regime requiring firms to register models that exceed certain performance thresholds. Other proposals include controlling the sale of powerful chips used to train LLMs and mandating that cloud-computing firms inform authorities when customers train frontier models.
Most firms also agree it is models’ applications, rather than the models themselves, that ought to be regulated. Office software? Light touch. Health-care AI? Stringent rules. Facial recognition in public spaces? Probably a no-go. The advantage of such use-based regulation is that existing laws would mostly suffice. The AI developers warn that broader and more intrusive rules would slow down innovation.
Until last year America, Britain and the EU seemed to agree on this risk-based approach. The breathtaking rise of LLMs since the launch of ChatGPT a year ago is giving them second thoughts. The EU is now wondering whether the models themselves need to be overseen, after all. The European Parliament wants model-makers to test LLMs for potential impact on everything from human health to human rights. It insists on getting information about the data on which the models are trained. Canada has a harder-edged “Artificial Intelligence and Data Act” in its parliamentary works. Brazil is discussing something similar. In America, President Joe Biden’s forthcoming executive order is also expected to include some tougher rules. Even Britain may revisit its hands-off approach.
These harder regulations would be a change from non-binding codes of conduct, which have hitherto been the preferred approach. Last summer the White House negotiated a set of “voluntary commitments”, which 15 model makers have now signed. The firms agreed to have their models internally and externally tested before release and to share information about how they manage AI risks.
Then there is the question of who should do the regulating. America and Britain think existing government agencies can do most of the job. The EU wants to create a new regulatory body. Internationally, a few tech executives now call for the creation of something akin to the Intergovernmental Panel on Climate Change (IPCC), which the UN tasks with keeping abreast of the latest research into climate change and with developing ways to gauge its impact.
Given all these open questions, it comes as no surprise that the organisers of the London summit do not sound that ambitious. It should mainly be thought of as “a conversation”, says Mr Black, the organiser. Still, the not-so-secret hope is that it will yield a few tangible results, in particular on day two when only 20 or so of the most important corporate and world leaders remain in the room. They could yet endorse the White House’s voluntary commitments and recommend the creation of an IPCC for AI or even globalising Britain’s existing “Frontier AI Taskforce”.
Such an outcome would count as a success for Britain’s government. It would also speed up the more official efforts at global AI governance, such as the G7’s code of conduct. As such, it would be a useful first step. It will not be the last. ■
Source: Business - economist.com