Source: Getty
commentary

The Importance of AI Safety Institutes

This essay traces the evolution of AI safety institutes around the world, explores different national approaches, and examines the need for an AI safety institute in India.

Published on June 28, 2024

Introduction

At the AI Seoul Summit held in May 2024, over twenty countries affirmed the role of artificial intelligence (AI) safety institutes to “enhance international cooperation on AI risk management and increase global understanding in the realm of AI safety and security.”

Countries at the forefront of AI development, including the United States, the United Kingdom, South Korea, Japan, Singapore, and the European Union, signed the Seoul declaration. This has laid the groundwork for a more collaborative and multilateral approach to AI safety through established institutions.

India has participated in the Seoul ministerial statement, suggesting there is political appetite for a national AI safety institute that can contribute to future engagements on AI safety.

In this essay, we trace the evolution of AI safety institutes around the world, explore different national approaches, and examine the need for an AI safety institute in India.

The Bletchley Effect

The launch of ChatGPT in November 2022 spurred policymakers to address the emerging risks of AI. These deliberations have resulted in the adoption of global principles for the safe development and use of AI at the G7, G20, GPAI, and the UN.

Meanwhile, the United Kingdom is taking the lead in identifying risks posed by “frontier models,” defined as “highly capable general purpose AI models that can perform a wide variety of tasks and can match or exceed the capabilities of current advanced models.” The inaugural AI Safety Summit, hosted by the United Kingdom at Bletchley Park in November 2023, helped develop a shared understanding of the risks of frontier models and the need for global norms on testing of AI models.

Till this point, the norm was for companies to brief government officials on the potential risks of frontier models. At Bletchley, participating companies committed to independent testing of their latest AI models – a significant move towards transparency.

The concept of an “AI Safety Institute” also emerged from this process. The UK launched the AI Safety Institute to coordinate research and build its internal capabilities for testing of frontier models. This was soon followed by AI safety institutes in the United States, Japan, Singapore, Canada, and the EU, in a process that is now being called the “Bletchley effect.

The concept of AI safety institutes gained momentum at the Seoul Summit, jointly organized by the UK and South Korea in May. The Seoul Declaration not only endorsed the establishment of AI safety institutes in different countries but also proposed the creation of an international network of such institutes for greater multilateral collaboration on AI safety.

A Global Network for AI Safety

The “international network of AI Safety Institutes” is based on a statement of intent signed by ten countries at the Seoul Summit—Australia, Canada, France, Germany, Italy, Japan, South Korea, Singapore, the United States, and the UK—along with the EU. The purpose of this global network is to enable governments to develop best practices, share resources for testing, exchange findings, and collaborate on monitoring specific AI harms and incidents.

AI safety institutes are usually publicly funded and state-backed. They offer technical expertise and resources to governments to help understand and mitigate the risks of AI. They also help policymakers formulate risk-based strategies, and evaluate the safety of AI models (the UK’s AI Safety Institute has already tested five AI models).

National approaches may differ in some respects. For example, the AI Safety Institute set up by the U.S. is established under the National Institute of Standards and Technology (NIST), a standard-setting body. Singapore, meanwhile, has designated the existing Digital Trust Centre in the Nanyang Technological University as its AI safety institute. Differences in enforcement powers also exist. For example, the  EU AI office, which acts as the EU’s AI safety institute, can exercise legal powers under the Artificial Intelligence Act, whereas the UK’s AI Safety Institute does not have any regulatory powers.

Ensuring that the frameworks developed by these institutes are ‘interoperable’, despite their structural and functional differences, will be a key metric of success of the global network.

An AI Safety Institute for India?

Ensuring AI safety is a policy imperative for India. It features prominently as the ‘seventh pillar’ of the AI mission, and anchors the current debate on AI regulation. With the development of indigenous tools and frameworks to promote AI safety already in the works, a national AI safety institute could provide the required expertise and institutional backing.

The private sector will likely support the establishment of a national AI safety institute. Building national capacity in AI safety and testing could help counteract burdensome regulation and promote interoperability across international markets.

Moreover, India has actively engaged in the global discourse on AI safety. It is a signatory to the Bletchley declaration and a part of the Seoul ministerial statement, both if which endorse the idea of AI safety institutes. Establishing a national AI safety institute of its own would ensure India’s continued involvement in these conversations, including at the next AI safety summit in France.

Now is the right time for the Indian government to bring together key stakeholders and convene a strategic dialogue on the need for a national AI safety institute.

Carnegie does not take institutional positions on public policy issues; the views represented herein are those of the author(s) and do not necessarily reflect the views of Carnegie, its staff, or its trustees.