S/PV.9821 Security Council
Provisional
The meeting was called to order at 9.35 a.m.
Adoption of the agenda
The agenda was adopted.
Maintenance of international peace and security Artificial intelligence
I warmly welcome the Secretary- General and the Ministers and other high-level representatives present in the Security Council Chamber. Their presence today underscores the importance of the subject matter under discussion.
In accordance with rule 39 of the Council’s provisional rules of procedure, I invite the following briefers to participate in this meeting: Mr. Yann LeCun, Chief AI Scientist, Meta, and Jacob T. Schwartz Professor of Computer Science, Data Science, Neural Science and Electrical and Computer Engineering at New York University; and Ms. Fei-Fei Li, Sequoia Professor in the Computer Science Department at Stanford University, co-Director of Stanford’s Human- Centered AI Institute and member of the Secretary- General’s Scientific Advisory Board.
The Security Council will now begin its consideration of the item on its agenda.
I now give the floor to His Excellency Secretary- General António Guterres.
I thank the United States for convening this meeting on artificial intelligence (AI) and the maintenance of international peace and security.
I briefed the Security Council about AI in July 2023 (see S/PV.9381). As I said then, those who feel like technology is moving very fast must understand a simple fact: technology will never move in the future as slowly as it is moving today.
In the short time since, artificial intelligence has moved at breakneck speed. Fuelled by record investments, today’s AI models keep getting more powerful, more versatile and more accessible — not only combining language, image, sound and video but also automating decisions. Artificial intelligence is not just reshaping our world; it is revolutionizing it. Tasks that required years of human expertise are now completed in a heartbeat.
But the risks are equally huge. This rapid growth is outpacing our ability to govern it — raising fundamental questions about accountability, equality, safety and security and about humankind’s role in the decision-making process. Artificial intelligence without human oversight would leave the world blind, and perhaps nowhere more perilously and recklessly than in global peace and security.
AI tools are already making a positive difference in countries suffering from conflict and insecurity — identifying food insecurity, predicting displacements caused by extreme events and climate change, and detecting and clearing landmines. And soon, AI could spot patterns of unrest before violence erupts.
But AI has also entered the battlefield in more troubling ways. Recent conflicts have become testing grounds for AI military applications. AI’s expansion into security systems raises fundamental concerns about human rights, dignity and the rule of law — from autonomous border surveillance to predictive policing and beyond.
I have long warned about the unforeseen consequences of AI-enabled systems — each advance creates new and unimaginable vulnerabilities. The AI arms race creates fertile ground for misunderstanding, miscalculation and mistakes. AI-enabled cyberattacks could cripple a country’s critical infrastructure and paralyse essential services. Most critically, AI is eroding the fundamental principle of human control over the use of force. From intelligence-based assessments to target selection, algorithms have reportedly already been used in making life-and- death decisions. The convergence of AI with other technologies amplifies those risks exponentially. The integration of AI with nuclear weapons is particularly alarming, with potentially disastrous consequences. We must avoid it at all costs. And looking ahead, quantum AI systems could breach the strongest defences and rewrite the rules of digital security overnight. Let us be clear: the fate of humankind must never be left to the black box of an algorithm. Humans must always retain control over decision-making functions, guided by international law, including international humanitarian law, international human rights law and ethical principles. Humankind’s hand created AI — humankind’s hand must guide it forward.
Beyond weapons systems, we must also address the other risks to peace and security posed by artificial
intelligence. AI creates highly realistic content that can spread instantly across online platforms, manipulating public opinion, threatening information integrity and making truth indistinguishable from outright lies. Deepfakes could trigger diplomatic crises, incite unrest and undermine the very foundations of societies. The environmental footprint of AI also poses distinct security risks. The massive energy and water consumption of AI data centres, combined with the rush for critical minerals, is creating dangerous competition for resources and geopolitical tensions.
Unprecedented global challenges call for unprecedented global cooperation. In July 2023, I welcomed calls from some Member States for the creation of a new United Nations entity to support collective efforts to govern AI and establish and administer internationally agreed frameworks and mechanisms of monitoring and governance. Since then, a series of initiatives has prompted high-level discussions around international peace and security implications, including on responsible applications of AI in the military domain. Declarations on AI have been issued by Member States, regional groups and international organizations. The United Nations has pursued efforts to reduce the fragmentation of AI governance and help bring those separate initiatives towards a common framework. The General Assembly adopted resolutions 78/265 and 78/311 on AI this year, thereby promoting enhanced global cooperation and capacity-building. In addition, a draft resolution focusing on AI in the military domain was recently recommended by the First Committee and will be considered by the General Assembly in the coming days.
Drawing on extensive global consultations, my High-level Advisory Body on AI developed in record time a blueprint for addressing both the profound risks and the opportunities that AI presents to humankind. Their work laid the foundation for a framework that connects existing initiatives and ensures that every nation can help shape our digital future. Adopted by leaders at the Summit of the Future, the Global Digital Compact (resolution 79/1, annex I) transforms that shared vision into action and represents the first universally endorsed framework on AI governance. It commits to establishing an Independent International Scientific Panel on AI and initiating a Global Dialogue on AI Governance within the United Nations, giving every country a seat at the table. The Compact requests options for innovative financing to build AI capabilities
where they are needed most, ensuring that developing countries receive our full support. A world of AI haves and have-nots would be a world of perpetual instability. We must never allow AI to stand for advancing inequality. Only by preventing the emergence of fragmented AI spheres can we build a world in which technology serves all humankind.
(spoke in French)
Our next steps will be crucial and the choices we make now will define our future. Every moment of delay in establishing international guardrails increases the risk for us all. I urge Member States to move swiftly in establishing the Independent International Scientific Panel on AI and launching the Global Dialogue on AI Governance within the United Nations.
I also reiterate my call for banning lethal autonomous weapons. We must establish new prohibitions and restrictions on autonomous weapons systems by 2026. No country should design, develop, deploy or use military applications of AI in armed conflict that violate international law, humanitarian law and human rights. That includes relying on AI to select or engage targets autonomously.
The members of the Security Council must lead by example and ensure that competition over emerging technologies does not destabilize international peace and security. I urge all members to join forces to build a safe, secure and inclusive AI future.
I thank the Secretary-General for his briefing.
I now give the floor to Mr. LeCun.
Mr. LeCun: Artificial intelligence (AI) will profoundly transform the world in the coming years and decades. Simply put, AI will amplify human intelligence. AI is accelerating progress in science and medicine. AI is facilitating access to information and knowledge. AI is making more people more informed, more creative and more productive. In the coming decade, AI will become pervasive. Everyone will have access to a virtual staff of AI assistants at all times. They will help us in our daily lives like a staff of human assistants. We will interact with them through wearable devices like smart glasses and smartphones. They will provide easy access to knowledge in every language in the world. AI will supercharge productivity and human creativity. It will help countries with ageing and decreasing populations. It will turbocharge scientific
and technological progress. It will help us solve some of the biggest problems facing humankind, such as climate change and the treatment of cancer as well as endemic or chronic diseases.
AI systems will eventually constitute a repository of all human knowledge. AI systems are produced in two phases. The first phase is training a foundation model and the second phase is fine-tuning it for a particular application. Foundation models must be trained on all of the world’s cultural materials in all languages if we want them to be accessible and useful to everyone around the world. Since all of our digital diets will eventually be mediated by AI systems, fine-tuned systems need to be numerous and diverse to represent all cultures and value systems around the world.
Two conditions are necessary for that to happen. First, foundation models must be free and open source, and secondly, training must be performed in a collaborative and distributed fashion in multiple data centres around the world. Governments and the private sector must work together to ensure that this global network of infrastructure exists to support the development of AI, enabling people all over the world to participate in the creation of a common resource.
The future of AI is inevitably one in which free and open-source foundation models will dominate. History shows that infrastructure software platforms always end up being open-source. For example, the software infrastructure of the Internet and the mobile communication networks are entirely open-source. It is clearly unacceptable for many Governments around the world that the digital diet of their citizens be controlled by a handful of companies. Meta has taken a leading role in producing and distributing free and open-source foundation models such as the Large Language Model Meta AI (LLAMA) family. LLAMA has enabled the emergence of an entire AI industry around the world and has been downloaded 650 million times since mid- 2023. A total of 85,000 open models derived from the LLAMA have been published around the world — and all are open-source. A fast-growing number of companies large and small, as well as Government institutions, non-profits and so forth, are building custom products and services by fine-tuning LLAMA foundation models. The applications include sectors such as agriculture and food production, education, healthcare, manufacturing and environmental preservation. We have observed an acceleration of progress thanks to the open-source nature of LLAMA and other similar models, through
contributions from academic labs, independent developers, start-up companies and large corporations.
There is also a big question about safety. AI applications must of course be deployed safely. Foundation models must go through rigorous testing and red-teaming. Historically, though, open-source platforms have been more secure than proprietary ones, and predictions of a flood of AI-generated disinformation due to the availability of AI systems have clearly been overblown. There is no evidence that current forms of AI present any existential risk or even a significantly greater threat than traditional technologies such as search engines and textbooks. Current AI technology is very focused on text and language rather than on the real world. That limits its applicability at the moment. Current AI systems do not understand the real world, do not have persistent memory and cannot really reason or plan. They cannot learn new skills with the same speed and efficiency as humans or even animals.
But AI will make dramatic progress over the next decade. There is no question that at some point in the future, AI systems will match and surpass human intellectual capabilities. They will be very different from current AI systems. They will be capable of understanding the physical world, as well as remembering, reasoning and planning. They may have some level of common sense. That will not happen tomorrow. Probably over the next decade or two, such superintelligence systems will do our bidding and remain under our control. They will accomplish tasks that we give them, subject to safety guardrails. Guardrails will shape their behaviour, similarly to the way that inviolable laws would shape human behaviour.
It is often said that AI is enabling the next industrial revolution. I think that the effect of AI on society may be more akin to the invention of the printing press and the wide dissemination of knowledge through printed material. By amplifying human intelligence, AI may bring not just a new industrial revolution, but a new renaissance — a new period of enlightenment for humankind. In that way, AI can contribute towards the maintenance of international peace and security — the vital mission of the Security Council — by supercharging the diffusion of knowledge and powering global economic growth. As with the Internet, international cooperation should focus on two initiatives. The first should be aimed at collecting cultural material, providing AI-focused
supercomputers in multiple regions around the world and establishing a modus operandi for the distributed training of a free and open universal foundation model. The second initiative should unify the regulatory landscape so that the development and deployment of open-source foundation models is not hindered.
I thank Mr. LeCun for his very thoughtful briefing.
I now give the floor to Ms. Li.
Ms. Li: I would like to thank you, Secretary Blinken, for inviting me here today. It is truly an honour. I also thank Secretary-General Guterres and Mr. LeCun for their remarks.
I have spent my entire career working in the field of artificial intelligence (AI), with more than 25 years dedicated to studying, developing and understanding that transformative technology. One thing is clear. Never before have we stood at such an extraordinary intersection of scientific possibilities and urgent global responsibility. My area of expertise in AI is computer vision, deep learning, robotics learning and AI for healthcare. Recently, I have focused on a new technology called spatial intelligence in the age of generative AI — that is, how AI systems perceive and interact with the three-dimensional virtual and physical worlds. That work has illuminated further promises of the technology, bringing us to some of the most exciting frontiers of innovation — for example, robots that can navigate disaster zones to save lives, precision agriculture systems that can address food insecurity and advanced medical imaging tools that can improve healthcare outcomes. Those advances highlight the incredible potential of this technology to help people drive scientific discovery and improve our world.
Yet we must also remain vigilant. The same capabilities that hold such promise can be misused. That duality — the ability to profoundly help or harm — lies at the heart of AI’s impact on peace and security. That is why public-sector leadership and a human-centred AI approach are critical. I want to reflect today on how we — Governments, researchers and global citizens — can ensure that AI serves humankind rather than undermining it.
First, a vibrant and healthy AI ecosystem is essential to ensuring that the transformative benefits of AI reach everyone. Right now, owing to the vast amount of computing and data required to train these
systems, much of the innovation in AI is concentrated in the hands of a few very large corporations and select nations. While that is one needed perspective, we must broaden the access and benefits of AI. We need a well- resourced AI public sector to ensure that AI’s benefits are widely distributed and aligned with public interests.
Governments should take bold steps to address that gap. We need what we call a moonshot mentality for AI — a commitment to visionary public-sector investment that goes beyond incremental reforms. The United States has already taken an important step with its National AI Research Resource Pilot programme, which aims to democratize access to computing resources and Government data sets. But that effort must be scaled globally, especially to ensure that no countries are left behind. I urge Governments to view AI not just as a technology to regulate, but as a strategic asset to invest in. That means funding basic research, supporting education and workforce development and creating inclusive platforms for global collaboration. Only with sustained public investment can we ensure that AI reflects the diverse needs and values of humankind.
Secondly, global collaboration has been a cornerstone of my career, from interdisciplinary research teams to cross-country partnerships. It is also the cornerstone of effective AI governance. I have had the privilege of serving on the Secretary- General’s Scientific Advisory Board, which studies how scientific and technological progress can support efforts to achieve the Sustainable Development Goals. I also appreciate the growing recognition of the need for global safety standards that ensure that AI systems are designed and deployed responsibly. However, we need to go further. I envision a multilateral AI research institute, a network of research hubs bringing together experts from across disciplines and pooling resources across nations. Such an institute would do more than advance technical innovation; it would set global norms for responsible AI development and deployment, rooted in democratic values and a commitment to international peace and security.
Global collaboration must also address the persistent digital divides that threaten to marginalize some regions of the world. The benefits of AI should not be reserved for wealthy nations. It is our collective responsibility to ensure that everyone has equitable access to AI tools, training and infrastructure. That is not just a matter of fairness — it is a matter of global stability.
Finally, AI’s transformative potential is matched by its complexity, which demands careful and evidence-based governance. Policies grounded in rigorous research will pave the way for innovation and global progress. Although we have identified risks such as algorithmic bias, disinformation and the misuse of autonomous systems, much remains to be understood about their scale and impact. Those are hurdles we can overcome through collaboration and ingenuity. A global research agenda is essential to fill those knowledge gaps and inform targeted interventions. That dedication to evidence-based governance bridges the gap between developers and policymakers, fostering an environment in which AI is not only safe and equitable but also a source of inspiration and progress.
As a researcher, a teacher and a mother, I often reflect on the world we are creating for future generations. Realizing the potential of AI requires vigilance, collaboration and a shared commitment to human dignity and global stability. I urge the Security Council to act with urgency and unity. By fostering public- sector leadership, championing global collaboration and advancing evidence-based policymaking, we can unlock AI’s transformative potential while safeguarding its responsible development.
I thank Secretary Blinken and Secretary-General Guterres for allowing me the privilege of speaking to the Council today, which is truly an honour.
I thank Ms. Li for sharing her insight and ideas with the Security Council, which are deeply appreciated.
I shall now make a statement in my capacity as the Secretary of State of the United States.
Let me begin by thanking both of our briefers, Mr. LeCun and Ms. Li, for sharing their thoughts with us today.
As we just heard and as I think so many of us know, artificial intelligence (AI) has the potential to do enormous good. Scientists are using AI to discover medications that could fight antibiotic-resistant bacteria. AI models are predicting natural disasters more accurately so that communities can better prepare. Such tools are identifying new crystal structures that could help us build the next generation of electric vehicle batteries. In those and in so many other ways, AI could accelerate our progress on nearly 80 per cent of the Sustainable Development Goals.
At the same time, as we also heard, if misused AI can pose tremendous threats to the international peace and security that the Council is charged with upholding. With AI, hackers can make cyberattacks more destructive and harder to trace. Repressive regimes are using AI- enabled surveillance to target journalists and political dissidents, destabilizing societies. If algorithms are built into weapons systems and they malfunction, they could accidentally spark a conflict. By setting rules of the road for AI, we can minimize those risks. We can harness the exceptional promise of that technology, and we can realize the vision that the United Nations enshrined in the Global Digital Compact: a future in which technology is inclusive, open, sustainable, fair, safe and secure for people everywhere.
Over the last few years, the United States has been leading international efforts towards those common goals. As home to the world’s leading tech companies, we have a responsibility to influence the evolution of artificial intelligence. We are also committed to mobilizing a collective response. We have teamed up with partners in Governments, the private sector and civil society in countries all across the globe to address both the perils and the opportunities of AI. First, our Government secured commitments from leading American companies to make AI systems safer. For example, they have agreed to establish and create tools such as watermarks that help users recognize AI-generated content. They will also strengthen their cybersecurity to protect AI models from hackers.
With Japan’s leadership, the Group of Seven expanded those pledges into a code of conduct for AI developers all across the world. It recommends that they run tests to identify safety risks, prioritize research into potential harms and publicly report on the limitations of AI in order to increase accountability. Earlier this year, the United States introduced the first stand-alone General Assembly resolution on AI (resolution 78/265), which was adopted by consensus. We have committed to promoting safe, secure, trustworthy AI systems that respect human rights and further economic and social progress.
We also agree to make the benefits of AI more accessible, in part by closing the digital divide that still exists around the world. That is something that we underscored in a second General Assembly resolution on AI (resolution 78/311), which was introduced by China and adopted by consensus in June. The United States and our partners have developed a global consensus
on AI and now we are building on it. Last month, the United States launched the International Network of AI Safety Institutes, in which researchers and experts are creating shared benchmarks for testing and evaluating AI systems. Their recommendations will offer practical guidance for developers and for tech companies.
We are also getting and setting ground rules for Governments. This year, the European Union, the United States and nine other countries signed the first international treaty on AI. We pledged to protect human rights, democracy and the rule of law when we use AI. That means safeguarding data privacy, adopting transparency and accountability measures and implementing other strategies that would limit any harms. The United States has rallied nearly 60 Governments to commit to guidelines for militaries, too. For example, we want to make sure that senior officials oversee the development and deployment of AI, including in weapons systems, and that those tools are used in ways that follow international humanitarian law. Separately, at a meeting in November, President Biden and President Xi Jinping affirmed that only humans should control the decision to use nuclear weapons. While we work to uphold our shared principles for AI, the United States is improving access to that technology so that communities everywhere can benefit. We are teaming up with leading tech companies to host trainings to build local data sets to provide AI tools for developers and researchers. This summer, the United States and Morocco also established a group at the United Nations — open to all Member States — in which experts from every region are sharing best practices for adopting artificial intelligence.
Now, that is real progress. But for all the progress, I think we all know that far more work remains to be done. Nations with leading tech sectors must do more to uphold security standards and prevent AI from being abused. The international community needs to stand together against irresponsible misuses of AI systems. Today, State and non-State actors are increasingly using those tools to influence, to distort public opinion, to manipulate geopolitical narratives and to make offensive cyberoperations more effective. That is only going to get worse as AI advances. The United States opposes the malicious use of AI by any actor, and we call on the other members of the Council to reject and to condemn those practices. We must adhere to our shared norms and build AI systems that are genuinely safe and secure.
In the months and years ahead, the Council will have an important responsibility. Since its inception, it has adapted to address the greatest threats to international peace and security — conflicts, terrorism, the spread of nuclear weapons. If the Security Council is going to continue holding that responsibility — and the United States believes that it must — it is incumbent on its members to grapple with the evolving risks of artificial intelligence. That requires leveraging our collective power to help set, update and eventually enforce international norms on AI, because, simply put, that will be vital to lasting security.
Now, even with the brilliant minds that we have heard from today, who have done so much in working on and thinking about the evolution of AI, I do not think any of us can fully predict what the future holds for AI. To fully understand how the technology changes over time and to stay ahead of the risks that it can pose, we have to continue collaborating. We need to keep working with developers, business leaders and members of civil society. If we do that, I am convinced that we can shape AI for the better so that it remains a force for progress and for the advancement of people all around the world.
I now resume my functions as President of the Council.
I call on the Minister for Foreign Affairs and Human Mobility of Ecuador.
I thank the United States for organizing this important meeting. I also thank Secretary-General Guterres and the other briefers for their briefings.
We are at a defining moment for humankind, characterized by the rapid advance of emerging technologies such as artificial intelligence, whose development presents both unprecedented opportunities and significant risks. Artificial intelligence has transformative potential. It can revolutionize key sectors, drive the Sustainable Development Goals, strengthen peacekeeping operations through early- warning systems and support conflict mediation. However, the uncontrolled development of those technologies without adequate regulation and respect for human rights poses significant risks. They include the potential to undermine global stability, concentrate power in a troubling manner, exacerbate geopolitical tensions and weaken democratic processes. In addition, access to those technologies by non-State actors, in
particular terrorist and criminal groups, poses a threat to international security. Those tools can be diverted for purposes of recruitment, coordination or incitement to hatred and violence.
Ecuador recognizes that the speed with which advances in artificial intelligence are being made requires global and collective responses. The governance of these technologies, which transcend national borders, must be the result of a coordinated international effort. In that regard, we welcome the initiatives that have led to the adoption of resolutions on artificial intelligence in the General Assembly, which Ecuador was honoured to co-sponsor. The initiation of a global dialogue within the framework of the United Nations, as agreed in the Global Digital Compact, is a fundamental step in balancing the opportunities and risks associated with such technologies and ensuring that artificial intelligence becomes a driver for the collective welfare, leaving no country behind. Ecuador will continue to advocate for artificial intelligence systems that are designed and used ethically, with strict respect for human rights, and that promote comprehensive sustainable development, encompassing its economic, social and environmental dimensions.
While celebrating the potential of artificial intelligence to decode global challenges and even to contribute to the eradication of poverty, the Global South must not allow critical barriers to stand in the way of its adoption. Universal, secure and accessible digital connectivity must be ensured to drive innovation, facilitate the achievement of the Sustainable Development Goals and promote international cooperation, enabling all States to take advantage of the opportunities of the digital age. At the national level, Ecuador is already promoting the adoption and ethical development of artificial intelligence. We are working in collaboration with UNESCO and the United Nations Development Programme to ensure that those technologies are used for the benefit of our society and in a responsible manner.
Where the military domain is concerned, it is a matter of urgency to address the challenges arising from the use of artificial intelligence in weapon systems. It will be crucial to move towards a legally binding regulatory framework governing the design and use of lethal autonomous systems, ensuring that they are developed ethically and in accordance with international law.
In the current race for leadership in the development of artificial intelligence, it is essential that governance and regulatory practices move equally swiftly in order to ensure that international peace and security are safeguarded. Geopolitical conflict should not prevent the eventual adoption of a new international regulatory body for the physical products that make up artificial intelligence. Nor should it sharpen divisions over the intangible assets that the technology requires. Once again, the emerging legal regime we are calling for must not serve to entrench a divided world order in which broad-based collective solutions fail. Neither should it instigate a new technological arms race. In this era of faltering global resolve, our common goal must rather be to prevent the misuse of artificial intelligence without restricting innovation in the process. As has been proposed, the idea of creating an international panel similar to the Intergovernmental Panel on Climate Change is an interesting one with regard to informing Governments on the current state of artificial intelligence capabilities and presenting evidence-based forecasts of what is to come.
As the Secretary-General said, we must engage in a race for the common good — a race to develop artificial intelligence that fosters peace and enables us to build a more equitable, secure and sustainable future for all nations, with a special emphasis on supporting developing countries.
I now call on the Minister for Foreign Affairs and Cooperation of Mozambique.
I would like to commend the United States of America for including among the signature events of its presidency an essential topic of great relevance to our lives as States, Governments and citizens of the contemporary world, demonstrating its commitment to addressing crucial issues that affect us all and to seeking collaborative solutions to global challenges. We also want to thank His Excellency Secretary- General António Guterres for his impactful briefing on artificial intelligence. The vision and guidance that he provided on the benefits and precautions to be adopted in relation to artificial intelligence are essential and will ensure that we continue to reflect on these issues in our societies.
As Mozambique’s term in the Security Council is coming to end, I would like to express, on behalf of His
Excellency President Filipe Jacinto Nyusi of the Republic of Mozambique, and on behalf of the Government and the team that represented our country on the Council, our deep gratitude to the United States and all the other permanent and non-permanent members of the Security Council, as well as the Secretariat, for the support and collaboration they have provided since we were elected to undertake that noble mission. We also wish the new members every success and reiterate our readiness to collaborate on issues of peace and security for the good of humankind.
It is an honour to address the Council to discuss the critical implications of artificial intelligence in maintaining international peace and security. The topic is extremely relevant in the current context of rapid technological changes and growing geopolitical challenges. Advances in artificial intelligence offer notable opportunities, such as improving decision-making, stimulating agricultural production, eradicating diseases, predicting pandemics and preventing conflicts. Those innovations and benefits, however, also present risks, such as the amplification of disinformation, the facilitation of cybercrimes and their use by terrorist networks for harmful purposes.
Those challenges require coordinated and collaborative approaches. For Mozambique, it is essential that the international community establish norms and standards that promote trust and cooperation among States and safeguard human rights. Global governance of artificial intelligence must ensure that its use is ethical, responsible and duly supervised by human beings.
The rapid acceleration of technological innovation seen in recent decades has drastically reduced the time that international decision makers and diplomats have to formulate appropriate policies. By way of example, while the electrical power grid took 50 years to reach 100 million users, recent artificial intelligence applications like ChatGPT reached the same milestone in just two months in 2022. That pace of development, coupled with the dual-use nature of artificial intelligence-based technologies, points to the urgency of anticipatory governance to prevent unintended consequences and mitigate potential risks. The integration of artificial intelligence in the context of peacekeeping and security indeed raises crucial questions about the control and enhanced design of systems operated by artificial intelligence in conflict situations, compliance with international humanitarian
law and the ethical implications of autonomous and automatic decisions in war scenarios.
We are encouraged by the fact that the international community is taking these challenges seriously, having recently promoted several initiatives, including multiple debates within the Security Council. We highlight the adoption of General Assembly resolution 78/265, which marked a significant step in the creation of a common framework for the use of artificial intelligence in a safe, secure and reliable manner.
During this debate, we would like to highlight the importance of aligning artificial intelligence systems with the Charter of the United Nations and the Universal Declaration of Human Rights, highlighting the importance of strengthening capabilities and reducing the digital divide between developed and developing nations. Furthermore, the recently adopted Pact for the Future (General Assembly resolution 79/1) addresses issues relating to lethal autonomous weapons, the prevention of arms races in outer space and the protection of critical infrastructure. That is a clear recognition from world leaders of the attention that must be paid to those scientific and technological innovations. The Global Digital Compact further reinforces our shared commitment to using technology for the common good. Those initiatives represent a foundation on which more robust and inclusive international governance mechanisms can be built. They also present the foundations on which Governments, private institutions and other interested parties can build effective strategies to deal with the challenges posed by artificial intelligence. Those initiatives play a fundamental role in stimulating dialogue on the ethical use of artificial intelligence in war contexts and in the prevention of an arms race in the field of technology.
My country is committed to developing a legislative and regulatory framework that promotes innovation and the judicious use of artificial intelligence. In November 2022, my country’s National Institute of Information and Communication Technologies presented a draft cybersecurity law to reinforce protection in the digital world. Additionally, efforts are under way to align national legislation with regional standards and to train qualified human resources in data science and programming. We recognize the need for further investment in infrastructure, education and policies that maximize the potential of artificial intelligence for sustainable development and peacebuilding.
Allow me to put forward some recommendations on the responsible use of artificial intelligence in the context of maintaining international peace and security.
First, it is essential to strengthen international cooperation, through mechanisms for sharing knowledge in this area between States, the private sector, civil society and the promotion of a multilateral dialogue on the risks and opportunities associated with artificial intelligence.
Secondly, it is important to enhance international governance structures, developing legally binding standards that regulate the responsible use of artificial intelligence in armed conflicts and prevent its misuse to destabilize regions or undermine the sovereignty of States.
Thirdly, it is critical to invest in training young people and women on technology, particularly in developing countries, in order to bridge the digital divide through technology transfer and technical support.
Fourthly and lastly, there is a need to ensure that artificial intelligence systems are used to promote transparency, accountability and justice, safeguarding respect for human rights and international humanitar ian law.
Artificial intelligence offers a unique opportunity to transform the approach to international peace and security. This Council has a crucial role in leading efforts to ensure that artificial intelligence is a force for global peace, progress and stability. We reiterate our commitment to the responsible use of artificial intelligence, and we are committed to participating in this global debate. It is our conviction that only with a joint effort will we be able to face the challenges and fully reap the benefits of technology, especially artificial intelligence.
Artificial intelligence (AI) has had a major impact on United Nations diplomacy over the past two years and has also marked Switzerland’s term as an elected member of the Security Council. The impact of that technology on the maintenance of international peace and security is a pertinent and timely topic. We would like to thank the United States for organizing this event and the speakers for their insightful presentations.
The spread of artificial intelligence is character ized by its extremely rapid pace and its profoundly dis ruptive nature. However, it is taking place within an immutable framework: that of public international law,
including international humanitarian and human rights law. Switzerland strongly reaffirms that the existing legal framework is fully applicable to AI, as we have emphasized on numerous occasions. Furthermore, ethi cal considerations must also be taken into account. The Security Council must carefully consider the implica tions of AI for the maintenance of peace and security.
In that regard, Switzerland wishes to highlight three points.
First, the Council must foster an inclusive discussion on AI governance. Effective rules are essential if we are to ensure that artificial intelligence systems are safe, secure and responsibly managed. The countries at the forefront of developing such systems are among the members of the Council and they have a special responsibility. However, the United Nations also provides a platform for ongoing dialogue that can ensure the interoperability of governance approaches around the world. To be effective, the meaningful inclusion of all State and non-State stakeholders is essential. Through a number of resolutions, including in the military domain, as well as the Global Digital Compact, the General Assembly has already begun to establish certain principles based on almost 20 years of United Nations experience in the digital field since the World Summit on the Information Society. That work is complemented by other initiatives, such as the Responsible Artificial Intelligence in the Military Domain Summit. The Council must also articulate its perspective on inclusive governance in order to contribute to existing processes and deliberations.
Secondly, artificial intelligence is an opportunity for a more effective implementation of the Council’s mandates. As a tool, artificial intelligence can facilitate the implementation of the mandates decided by the Security Council. Switzerland organized an Arria formula meeting in May that illustrated that potential in a tangible manner, in particular in the context of peace operations. In collaboration with the Geneva- based DiploFoundation, Switzerland has created an AI- based tool that facilitates access to, and the analysis of, data from 10 Council meetings, with a particular focus on the New Agenda for Peace.
Thirdly, the Council must promote measures throughout the life cycle of artificial intelligence systems in order to ensure that they are safe, secure and responsibly managed. As a priority, the Council must better anticipate the impact of technologies such
as AI on the maintenance of peace and security and act preventively. Switzerland organized a Council briefing on the topic in October (see S/PV.9753) and the issuance of presidential statement S/PRST/2024/6, in which the Council expressed its determination to take into account more systematically the scientific advances that could have an impact on international peace and security.
Human beings and their dignity must always be top priorities when it comes to the development and use of artificial intelligence systems. Human rights due diligence and human oversight are two essential elements that must be taken into account. AI also has important implications in the context of armed conflicts and humanitarian crises. Having organized two exhibitions in collaboration with the International Committee of the Red Cross, entitled “Digital Dilemmas” and “Deepfake and You”, Switzerland made use of its Council presidencies to raise awareness of that issue and its related challenges. In partnership with the Office of the United Nations High Commissioner for Refugees, Switzerland is committed to protecting forcibly displaced persons and humanitarian actors from digital risks, such as misinformation and online hate speech. It is therefore exploring how artificial intelligence can help to mitigate those risks, in particular in humanitarian contexts.
The emergence and spread of technologies such as artificial intelligence have a profound implication for global peace and security. Switzerland is convinced of the importance for the Security Council to systematically address those developments in order to identify the necessary political responses and concrete actions. Switzerland is proud to have contributed ideas and impetus during its term as an elected member over the past two years.
I begin by thanking you, Mr. President, for convening and presiding over today’s high-level briefing on this highly topical issue. I also thank the briefers for their valuable insights.
The responsible development and use of artificial intelligence (AI) require clear principles and international governance to uphold human dignity through ethical oversight and accountability. The ethical frameworks shared by all stakeholders are essential to safeguard against harmful effects, including biases affecting marginalized groups. Safeguards to mitigate risks during the design, development and deployment of AI systems must be integrated, thereby ensuring
that they do not undermine information integrity or exacerbate digital divides. To ensure that AI supports human dignity, ethical oversight must be embedded at every stage. National legislation should mandate principles such as fairness, transparency, accountability and gender equality, while preventing discrimination. Human oversight must remain integral, preventing automated decisions from eroding individual rights. By prioritizing safeguards, we can create systems that serve humankind, while minimizing risks. With the advancement of AI also in the military domain, it would be relevant for the international community to reach consensus on the way forward on the discussions being held within the Group of Governmental Experts on lethal autonomous weapons systems in Geneva, so as to have adequate regulation on autonomous weapons.
Decades of research into AI have underscored a number of key lessons.
First, inclusive governance frameworks, such as UNESCO’s Recommendation on the Ethics of Artificial Intelligence, emphasize the importance of embedding ethical considerations into AI development. Those frameworks ensure that AI serves humankind without undermining it. Leadership from women in technology and governance is essential to ensuring that gender perspectives are included in AI development.
Secondly, prioritizing human judgment in AI systems reinforces respect for rights and addresses biases affecting underrepresented communities.
Thirdly, fostering transparency and accountability builds trust, thereby ensuring alignment with societal values.
Fourthly and lastly, capacity-building and equitable, mutually agreeable technology transfer empower communities to benefit from AI advancements, without compromising dignity.
Governing AI demands diplomatic efforts and consensus-building, as seen in the Global Digital Compact. Collaborative efforts establish best practices for testing and help to bridge digital divides and enhance security. International cooperation should also prioritize capacity-building for developing countries. Empowering all States to participate in AI governance ensures fair representation and equitable access. General Assembly resolution 78/311 on enhancing international cooperation on capacity-building of artificial intelligence, which was adopted in July,
underscores the need for global collaboration that is aligned with the Charter of the United Nations and international law. Such collaboration should promote safe, trustworthy AI technologies, while advancing peace, security and development. Member States can take proactive steps to foster a robust international AI ecosystem. Building consensus on ethical AI use and unified risk mitigation approaches is essential.
The proposed Global Digital Compact office could enhance United Nations coordination, ensuring that AI technologies respect human rights and support sustainable development. Multi-stakeholder collaboration should remain central, incorporating perspectives from Governments, the private sector and civil society in order to uphold human dignity. Member States should prioritize equitable access to AI benefits by investing in data access and capacity- building initiatives. Establishing frameworks for cross- border regulation is crucial to mitigating risks such as monopolization and to ensuring fair competition. Public-private partnerships can bridge the technology gap for developing nations.
Prioritizing human judgment in AI systems reinforces respect for rights and addresses biases affecting underrepresented communities. Fostering transparency and accountability builds trust, ensuring alignment with societal values. Robust accountability frameworks and national policies aligned with human rights standards are essential to preventing potential abuses. Gender equality must be considered in military AI applications to protect women and girls in conflict zones. Member States must strengthen safeguards against the misuse of AI, particularly in such areas as surveillance and encryption. Independent oversight mechanisms can ensure that military applications of AI uphold global, legal and ethical standards. By balancing AI capabilities with ethical use, we emphasize human oversight, transparency and accountability.
In conclusion, the responsible governance and development of AI hinge on collaboration, ethical oversight and adherence to international law. Let us work together to advance safe, trustworthy AI technologies that support peace, security and development for all.
Let me start by thanking the United States presidency for convening this important meeting and for the international leadership that it has shown in the field of artificial intelligence (AI) governance. That includes the first-ever General
Assembly resolution (resolution 78/265) on artificial intelligence. I am grateful to Ms. Li and Mr. LeCun for their insightful briefings. I would also like to pay tribute to the Secretary-General for his consistent campaign to bring to our attention the major opportunities and significant risks of this transformative technology and shape the international response.
It has been 18 months since the United Kingdom convened the first Security Council meeting (see S/PV.9381)) on the security implications of artificial intelligence. As the Secretary-General said, things have moved swiftly since then. Significant progress has been made in international collaboration, including the Global Digital Compact and the United Kingdom- initiated AI Summit series. In addition to the United States resolution, the General Assembly also adopted a Chinese-led resolution (resolution 78/311) on AI capacity-building, both of them by consensus. And two Nobel Prizes were awarded to AI innovators, both of them British scientists. Today I will focus on our priorities for peace and security, the opportunities that AI presents in that field and how we can share those benefits through capacity-building.
First, the United Kingdom recognizes that AI brings risks that could escalate conflicts and cause harm. But it also presents significant opportunities for supporting international peace and security, including better decision-making, improved early-warning systems and enhanced planning to support resilience and emergency preparedness. The United Kingdom welcomes progress on increasing international understanding in a military context through the Summit on Responsible Artificial Intelligence in the Military Domain and the United States-led political declaration on the responsible military use of AI. The draft resolution submitted to the General Assembly on artificial intelligence in the military domain (First Committee draft resolution A/C.1/79/L.43) presents an important opportunity to advance efforts for AI’s safe and responsible development. AI also offers significant opportunities for peacekeeping. AI can improve missions’ data collection, analysis efforts, situational awareness and decision-making. At the request of the Department of Peace Operations, the United Kingdom is therefore authoring a paper on the potential uses of AI in peacekeeping.
Secondly, AI presents huge opportunities for broader areas that affect peace and security, such as the enjoyment of human rights, by facilitating greater access
to information or improving the interpretation of medical data to aid early health diagnoses. As an inaugural signatory to the Council of Europe’s Convention on Artificial Intelligence, the United Kingdom is committed to protecting us from the risks posed by AI to human rights, democracy and the rule of law.
Lastly, the United Kingdom is also committed to sharing those opportunities through capacity-building. It is important that in contributing to international peace and security we aim to narrow digital divides. Through AI for development collaboration, the United Kingdom has donated £58 million to funding multidisciplinary responsible AI research in six African countries, with more planned for 2025. Ultimately, we must seize the opportunities and grasp the challenges of AI, including those for international peace and security, decisively and optimistically.
I would like to begin by expressing my gratitude to the United States for convening this important meeting and my sincere appreciation to the Secretary-General for his insightful briefing. I also thank Mr. Yann LeCun and Ms. Fei-Fei Li for their valuable contributions.
As an ultimate enabling technology with multiplier effects for all elements of national power, AI has an immeasurable potential impact on international peace and security. As many of us pointed out at the first Security Council briefing on AI last year (see S/PV.9381), it can promote international peace and security in various ways, as the Secretary-General and Secretary Blinken also outlined today. At the same time, AI can exacerbate risks such as the proliferation of weapons of mass destruction (WMDs) and sophisticated cyberattacks, when used by irresponsible actors such as North Korea, which is engaging in cryptoheists to finance WMD programmes. AI can also be an unstoppable spreader of disinformation and misinformation — or it can guard against them. In order to ensure that AI fosters prosperity and benefits rather than posing risks to humankind, it is critical to ensure that AI is not just responsibly used but that it is safe, secure and trustworthy, while promoting innovation.
In that regard, the Republic of Korea is making active efforts to ensure that AI will be conducive to international peace and security. In 2023, we co-hosted the first Summit on Responsible Artificial Intelligence in the Military Domain (REAIM) with the Netherlands. In May Korea hosted the AI Seoul Summit and
adopted the Seoul declaration, which underscores the importance of safety, innovation and inclusivity in AI global governance. In September we co-hosted the second REAIM Summit with the Netherlands, Singapore, Kenya and the United Kingdom. The Summit culminated in the adoption of a blueprint for action, which considers the impact of AI on international peace and security and the key principles for the responsible military application of AI. In October, the Republic of Korea and the Kingdom of the Netherlands introduced draft resolution A/C.1/79/L.43 in the First Committee, on AI in the military domain and its implications for international peace and security. Thanks to the overwhelming support for the draft resolution from 165 Member States, we are ready to continue to promote discussions on the topic. Based on those efforts, I would like to emphasize three points today.
First, the application of AI must be ethical and human-centred throughout its entire life cycle. To that end, AI applications need to be developed, deployed and used in accordance with international law, including, as applicable, the Charter of the United Nations, international humanitarian law and international human rights law. In that context, Korea has been introducing the biennial draft resolution on new and emerging digital technologies and human rights at the Human Rights Council since 2019.
Secondly, we need to identify, internalize and operationalize relevant principles to ensure responsible AI. We believe that the key principles laid out in the Blueprint for Action adopted at this year’s REAIM Summit held in Seoul can serve as a valuable stepping stone for the international community to achieve the responsible use of AI in the military domain.
Thirdly, it is absolutely essential to share information and strategies among all stakeholders, including the Governments of developed and developing countries, industries, academia and civil society, in order to promote a collective understanding of AI technology and its implications and grasp each other’s priorities and approaches, because no single country or entity can govern AI alone or establish relevant norms. To that end, the Republic of Korea plans to launch capacity-building initiatives next year as a follow-up to the REAIM Summit.
In his final book Genesis: Artificial Intelligence, Hope, and the Human Spirit, Henry Kissinger observed that AI could fundamentally challenge
the entire international system, including even the Westphalian system itself. We are now only starting to comprehend the profound implications of AI. Facing the immense promises of AI, countries can react differently. Some may focus on safety, some on innovation and some on finding the best application. However, future challenges to international peace and security involving AI can come not only from humans abusing AI but also from AI being out of human control. In today’s hyper-connected world, AI can be a potentially unpredictable threat, as its knowledge can be unlimited, its applications underguided and its sphere of influence unbounded. Our best hope against that potentially boundless threat is borderless cooperation by the peoples of the United Nations.
We thank Secretary-General Guterres and the briefers for their contributions to the discussion.
The Russian Federation attaches great importance to the development of advanced technologies designed to serve the good of humankind, including artificial intelligence (AI). At the same time, we are invited today to discuss AI in the context of threats to international peace and security. Recalling the previous discussion, I am sure we can agree on one thing — we must not allow AI to dominate human beings and human values. However, when it comes to the threats we may face on the path towards it, the Security Council is far from being unanimous.
I would like to begin by saying that it was quite entertaining to read the presidency’s concept note. The focus is precisely what we expected to see. The authors are attempting to determine how they can develop an international system of AI governance, while at the same time ensuring that some industries or States do not fall under that governance. It sounds familiar, does it not? In that context, I would like to recall similar approaches to addressing other topics, as the same things appeared today in their full colours.
If we talk about the international trade system, for example, we can recall the Atlantic Charter of August 1941, which contained a reference to the principles of non-discrimination and market efficiency, which subsequently formed the premise for the General Agreement on Tariffs and Trade and ultimately the World Trade Organization (WTO). And what do we see today? The United States, which was one of the originators of this process, having secured its leadership in world trade
and finding itself unwilling to put up with its waning dominance, is now blocking the work of the WTO and expanding the practices of using illegitimate sanctions and other unlawful means to deal with its competitors.
Another example is the fight against climate change. American scientists and politicians have indeed made a significant contribution to developing a sort of religion of climate alarmism. And what do we see today? Washington does not hesitate to disrupt the long and hard-won global climate cooperation in those areas in which it contradicts its interests. Another interesting fact is that in recent years, the United States has only increased the extraction of hydrocarbons on its territory and expanded their export, which was of course achieved based on unfair competition practices.
Let us take a broader example. For many years, we were assured that the process of globalization would bring the benefit of development to peoples if the process were governed by the principles of democracy, the laws of the free market and the inviolability of private property. So-called summits for democracy were held. However, it is no secret that Washington is trampling today on the principles and foundations of sovereign equality, as well as the inviolability of public property, which became clear after Washington, together with London and Brussels, openly plundered other countries’ gold and foreign exchange reserves. As a result, as President Vladimir Putin said, in the existing American-style world order there is only one rule left, which is that there are no rules at all.
Of course, we understand what the United States is guided by when reflecting on the fate of humankind in the era of rapidly developing AI. However, it is now important for the country that initiated today’s discussion to realize that if it continues with the approaches adopted in the past — imposing rules on others while simultaneously freeing itself from those rules — it will repeat the mistakes it made in the past on the path of developing global cooperation. If we look at the United States initiative on AI at the General Assembly, we will see that it is weighed down by the same old mistakes. We refer to the United States- sponsored General Assembly resolution 78/265 on AI, which speaks very eloquently about providing “a fair, open, inclusive and non-discriminatory business environment, economic and commercial activities”. Unfortunately, the authors of the resolution continue to unscrupulously get rid of competitors in practice. On 11 December, the United States Bureau of Industry
and Security decided to blacklist another four Russian information technology companies. One of them is NtechLab, which specializes in facial recognition. That company is known because its team was one of the five finalists in the MegaFace competition organized by the University of Washington, having beaten the team of developers from Google. How do Council members like that type of fair and inclusive environment? No, if things go in that direction, as was the case following the agrarian and subsequently the industrial revolutions, progress will again be achieved only in the flourishing “garden”, whose sole concern will be how to extinguish the flames in the “jungle”.
We have an opportunity to combine our efforts and together prevent the global South from falling behind, so that we can avoid new forms of colonialism and discrimination. It is precisely the aggravation of existing technological inequalities that constitutes one of the main threats to the development of AI, including the threats to international peace and security. The transition to AI-based innovations requires bridging the so-called digital divide in order to reduce poverty and facilitate sustainable development — and that divide is indeed vast. After all, according to forecasts global public investment in artificial intelligence could reach $632 billion by 2028. The undisputed leaders in that area are the United States, the countries of Western Europe, Japan and South Korea, not to mention the private sector, primarily the major technology companies such as Microsoft, Meta, Alphabet and Amazon, which are planning to allocate approximately $250 billion to creating AI infrastructure by 2025. Of course, the countries of the global South also have their own leaders in that area, namely, our BRICS partners, China and India. However, the majority of developing countries do not have the necessary resources for such acceleration.
That acceleration requires international cooperation aimed at transferring technologies, developing human capital, building the necessary infrastructure and improving the quality of jobs. In that regard, the Russian Federation, as part of the group of like-minded organizations established on the initiative of China, participated in drafting General Assembly resolution 78/311 on enhancing international cooperation on capacity-building of artificial intelligence, which is in line with the priorities of our national artificial intelligence strategy.
We intend to provide the countries of the global South and the global East with technical assistance and
are planning to carry out joint projects based on equal access to knowledge and technologies. It is on the basis of those approaches that Moscow held the AI Journey conference last week. Having its own latest-generation AI developments is one of the key conditions for the scientific and technological sovereignty of any country. In addition, we would like to announce the launch of the BRICS AI Alliance Network, which will include national associations and development institutions in the field of AI from BRICS countries and other interested States. At the same time, we firmly believe that the algorithms of AI systems must be trustworthy — meaning that they must be understandable, open and unbiased, and they must take into account the cultural and national specificities of countries and of each civilization, in accordance with their history, identity and traditions.
With regard to the coordinating role of the United Nations in AI development, in line with the provisions of the Kazan Declaration of the sixteenth annual BRICS summit of 23 October, we support that role as a counterweight to the various non-inclusive formats with politicized agendas and ad hoc coalitions. At the same time, the key thing for us is to reach universal agreements in which States play the lead role and hold dialogues with each other on equal terms, and in which due account is taken of all the legitimate interests of the participants in the negotiation process. As we have indicated in previous statements, we do not regard the Security Council as the appropriate platform for addressing AI as a subject, because it is a global issue, and it would be wrong to limit its discussion to the Council. The Summit for the Future has already outlined the contours of the emerging infrastructure for addressing that topic in the United Nations system and the Security Council is not part of that infrastructure.
Given that some colleagues insist on the need to discuss the military aspects of the use of AI in the Security Council, we would like to recall that there are also inclusive specialized platforms for that purpose, in particular the Group of Governmental Experts on Lethal Autonomous Weapons Systems, which operates within the framework of the Convention on Certain Conventional Weapons. Indeed, the Group of Governmental Experts has not yet reached a general understanding on even the most basic issues such as terminology. We would also welcome the consideration of this subject by the United Nations Disarmament Commission but without duplicating the efforts of other mechanisms. We believe that it is still premature
to discuss the development of the rules for responsible behaviour when it comes to weapons systems that use AI.
Accordingly, to bring into other formats the discussions on such sensitive topics — and I have not even mentioned the issue of the impact of AI technologies on other non-proliferation and disarmament matters — would be premature to say the least. Attempts to impose on the Security Council the approaches and attitudes of certain States, as well as attempts to substitute those approaches for international legal instruments, will hardly bring us closer to developing collective approaches to addressing the subject of the military use of AI.
At the outset, I would like to thank the Secretary-General for his briefing. I also thank the briefers for sharing their views with us today on this important subject.
Artificial intelligence (AI) is one of the major revolutions of this century. It is already shaping our daily lives, our economies and the way our societies and institutions function. Artificial intelligence constitutes a major challenge because of the development opportunities it creates. We need to determine the scope of that revolution and support it for the benefit of all. Our priority should be to put in place an inclusive, multi-stakeholder, responsible international governance for AI that respects fundamental rights. As more and more initiatives seeking to regulate AI emerge, it is important that we work together to prevent the fragmentation of the normative systems, which would lead to competition between rival and divergent models. We must ensure that the development of AI is accessible to all States and peoples on the planet and that it does not exacerbate the digital divide. Those concerns were at the focus of the Global Digital Compact (General Assembly resolution 79/1, annex I), which was adopted by consensus in the Pact for the Future in September.
It is with those objectives in mind that France will host the AI Action Summit in Paris on 10 and 11 February 2025, a primary goal of which is to contribute to creating a common framework for AI governance. The Summit will bring together Governments, international organizations, businesses, the technical community, researchers and civil society actors, with a view to advancing the framework for a shared vision of an AI that we can trust for the benefit of societies at large. The Summit will also explore key
issues such as the environmental impact of AI and the AI market concentration.
The development of AI has far-reaching implications for international peace and security. We are pleased that the Security Council continues to deepen its discussions on the subject, following the convening of its first meeting on the subject in July 2023 (see S/PV.9381). Some threats to international peace and security can be exacerbated by the misuse of artificial intelligence. In particular, generative AI tools are used in disinformation and information manipulation campaigns designed to destabilize societies or States. In the cyberdomain, AI further enables malicious actors to recognize and exploit vulnerabilities in information systems. The Summit to be held in Paris will provide an opportunity to identify concrete solutions to address that problem.
In that context, France actively supports international efforts to promote the responsible use of AI for peace and security, in compliance with international law. To that end, we co-sponsored draft resolution A/C.1/79/L.43 on artificial intelligence in the military domain and its implications for international peace and security, which was submitted in the First Committee of the General Assembly by South Korea and the Netherlands.
France also continues to support the work being carried out in Geneva to ensure respect for international humanitarian law in the potential development and use of autonomous lethal weapons systems. The decision to use force must always ultimately be the responsibility of human beings.
The Security Council should also continue to give further consideration to AI in the context of its work, including by integrating AI-related issues into considerations about peacekeeping operations or by exploring ways to harness the potential of that technology, such as in monitoring the implementation of sanctions regimes.
I would like to thank Secretary of State Blinken and the United States Security Council presidency for convening today’s meeting. I also thank the Secretary-General and Professors LeCun and Li for their insightful briefings.
We are currently in an era of rapid advancement of artificial intelligence (AI), a development that holds the potential to enhance our global outlook. Many members
of the Council, both past and present, have advocated for an evidence-based approach to decision-making. AI offers a unique opportunity to respond to that call, po tentially revolutionizing how we tackle issues related to climate, peace and security and the fight against food insecurity, among other issues. However, the emergence of AI has undoubtedly created new vulnerabilities that the Council must acknowledge, in addition to consider ing mitigation measures that could be implemented to safeguard international peace and security.
Today’s meeting is being held in the context of in creased global conflict, with AI playing a central role in what is now deemed the age of cyberwarfare. Several examples exist of States augmenting their military arse nals with cybercapabilities. AI therefore risks compli cating the maintenance of international peace and secu rity due to concerns about its role in offensive and re prisal attacks, including the use of drones and facial rec ognition algorithms. In the case of the Middle East, for example, we have heard reports of civilians being mon itored and sometimes attacked using drone technology.
AI weapons are also being programmed and authorized to select their targets without further human authorization, increasing protection concerns. In addition, AI-piloted fighter jets and sniper drones have shown the substantial capability of dominating human pilots in simulated interactions. Those concerns engendered by modern warfare must be addressed through the appropriate global regulations.
Those concerns were also echoed by the Secretary- General in his policy brief on the New Agenda for Peace, in which he noted that
“Developments in artificial intelligence and quantum technologies, including those related to weapons systems, are exposing the insufficiency of existing governance frameworks”.
Guyana acknowledges those concerns and underscores the need for the Council to give attention to AI governance, given its implications for the maintenance of international peace and security. As we have heard before, the General Assembly is already being proactive on the subject of AI through its adoption of resolutions 78/265 and 78/311, introduced by the United States and China, respectively. The recently adopted Global Digital Compact (resolution 79/1, annex II) also offers practical recommendations for enhancing the international governance of AI. An international framework for AI governance will ensure the full and
equal representation of all countries, thereby creating opportunities to harness AI in bridging existing gaps between developed and developing countries.
Guyana welcomes the establishment of the Secretary-General’s High-level Advisory Body on Artificial Intelligence, which launched its final report in September to address the global governance deficit of AI. Among the outcomes recommended in the report, Guyana eagerly anticipates the creation of a global fund for AI. That fund, to be specifically designed to help developing countries, will be critical, especially for small developing States, which are at risk of not benefiting from AI owing to the digital divide. We must ensure that AI benefits everyone instead of accelerating inequality. That is why the proposed creation of an AI capacity development network for United Nations-affiliated capacity development centres is welcome. That approach can serve as a catalyst to align national, regional and global AI advances, with positive implications for the implementation of the 2030 Agenda for Sustainable Development and the Pact for the Future.
In conclusion, Guyana acknowledges that AI’s rapid development and deployment have elicited both enthusiasm and anxiety. We must therefore develop the necessary innovative and inclusive frameworks for AI governance and cooperation in order to ensure that AI is used as a force for good and to mitigate its harmful use. Guyana therefore urges the Council to consider how it can advance the recommendations, including those proposed at today’s meeting, in order to further such work for the maintenance of international peace and security.
I thank Secretary Blinken for presiding over this meeting. I thank Secretary-General Guterres for the important statement, and I also thank Professors Yann LeCun and Fei-Fei Li for their important briefings.
At present, with the development of artificial intelligence (AI) accelerating and its role as an enabler ever increasing, AI has become an important force in the promotion of social development and progress. AI technology is like a double-edged sword — it can either become a force for better social governance and international security or a source of threats to fairness and justice and a danger to peace and stability. The international community should jointly strengthen global AI governance to
ensure that AI development advances in conformity with the common values of humankind and that AI is genuinely safe, reliable, fair and under control.
China has been exploring the scientific path of AI development and governance in a responsible manner. Back in 2017, the Chinese Government issued its development plan for the new generation of AI, clearly indicating the need to address the double-edged attributes of AI with a view to preventing its risks as much as possible. In 2021, China released the code of ethics for the new generation of AI, integrating ethics into its entire life cycle. In 2023, the Chinese Government issued the world’s first ever legislative document on generative AI. China has been constructively engaged in the global governance of AI. In October 2023, President Xi Jinping proposed the Global AI Governance Initiative, comprehensively illustrating China’s position in three dimensions — development, security and governance. Resolution 78/311, entitled “Enhancing international cooperation on capacity-building of artificial intelligence”, which was introduced by China, was adopted by consensus by the General Assembly during its seventy-eighth session. Subsequently, China proposed the AI Capacity-Building Action Plan for Good and for All and promoted and established the Group of Friends for International Cooperation on AI Capacity-Building. At the Heads of State and Government Summit of the Group of 20 held recently in Rio de Janeiro, President Xi Jinping stressed the need to strengthen international governance and cooperation on AI and ensure that AI works for good and for all rather than becoming a game for the rich and the powerful. Under the BRICS framework, China and other members agreed on the establishment of an AI study group for information exchange and technological cooperation, with a view to formulating an AI governance framework, norms and standards resting on a broad consensus.
Turning to the military applications of AI, an issue that has significant implications for the future and destiny of humankind, China has demonstrated its vision and sense of responsibility as a responsible major power. In 2021, China submitted a position paper on regulating the military applications of AI to the sixth Review Conference of the Convention on Certain Conventional Weapons. We maintain that all countries, in particular the major Powers, should adopt a prudent and responsible attitude towards the military development and use of AI technology and should comprehensively assess both its pros and cons and its
consequences. They should also sincerely respect the security concerns of other countries so as to avoid misunderstandings and miscalculations and prevent an arms race. It is imperative to constantly improve the safety, reliability and controllability of the relevant technologies in order to ensure that they are under human control and to oppose the misuse, abuse and proliferation of such weapons systems.
When the Heads of State of China and the United States met recently in Lima, they recalled the dialogue and cooperation between the two countries in the field of AI and affirmed the need to address the risks of AI systems in order to improve AI security and enhance international cooperation so as to ensure the use of AI for good and for all. Both leaders stressed that a prudent and responsible attitude must be adopted towards developing AI technology in the military field. Both leaders affirmed the need to uphold the decision that nuclear weapons must be under the control of humans. The law of scientific development has shown that in order to solve problems in science, knowledge and progress, one should not remain idle without even trying or allow fear to paralyse our progress.
In view of the rapid development and use of AI technology, China would like to advance the following proposals with regard to what actions the international community should take. First, clear guidelines must be established. Security and development must be given equal attention. Smart governance must be put in place, strengthening anticipatory assessment and control of risks and encouraging technological innovation and peaceful uses. It is imperative to abide by the purposes and principles of the Charter of the United Nations, observe the basic norms governing international relations and ensure that AI technology will not become a tool for waging wars and pursuing hegemony. It is imperative to take a people-centred approach, abide by international humanitarian law and ethics, ensure human dignity and human rights, and refrain from the indiscriminate use, abuse or misuse of such technologies.
Secondly, governance measures must be improved. It is necessary to actively develop reasonable and feasible technologies for AI governance, establish risk assessment and testing systems, and conduct hierarchical management by category. Relevant laws, rules and regulations must be put in place and improved. Education and training targeting practitioners must be strengthened. Prevention on both the human and
technological fronts must go hand in hand in order to prevent the systemic threats posed by hacker attacks or data fraud, among other threats, and to diminish the risk of proliferation in the military use of AI.
Thirdly, international cooperation must be strengthened. AI technology is not a cake to be divided among a small group of people and neither should its global governance be determined by a handful of countries. China firmly opposes the practice of imposing on others rules formulated by a small number of countries or the creation of small cliques that target certain countries. China firmly opposes discriminatory barriers based on ideological differences, which undermines the right of all countries, in particular developing countries, to use emerging technologies on an equal footing. Decoupling and building small enclaves with high walls runs counter to the laws of the market economy, undermines the international economic order and disrupts the stability of global production and supply chains. Only by the East and West learning from each other, implementing joint governance for the benefit of all, seeking the greatest common denominator and expanding cooperation can we embark on the right path.
Given the widening technology, digital and AI divides between the global North and the global South, it is important to uphold cooperation, enhance the representation and voices of developing countries and implement the Global Digital Compact so that AI technology can benefit developing countries. In that regard, as the most universal, representative and authoritative intergovernmental international organization, the United Nations should become the main channel for global AI governance and coordinate major issues related to the development and security of AI.
As a major AI Power, China stands ready to actively support the relevant work of the United Nations and contribute to the establishment of governance norms that enjoy universal participation and broad consensus.
I thank the United States for convening this important meeting at a critical juncture, as we face what the Secretary-General aptly described as an era of epic transformation. We thank Secretary-General Guterres for his remarks. We also listened carefully to Mr. LeCun and Ms. Li.
Artificial intelligence (AI) is changing not only our lives and work but also the peace and security paradigm and landscapes. Although the opportunities
are immense — AI can help predict conflict, enhance peacekeeping, improve humanitarian responses and accelerate sustainable development — the risks are equally significant.
Three fundamental challenges tower over us:
First, the growing AI divide threatens to leave developing countries behind. It is not just about machines and algorithms — it is about sovereignty itself, it is about security, and it is about the sacred right of nations to chart their own courses.
Secondly, the digital landscape contains new threats, as AI-powered border-proof attacks can damage societies and manipulated information can poison minds.
Thirdly, the deployment of AI systems without adequate safeguards has the potential to escalate conflicts and threaten international peace and security.
Let us be clear: international law is not optional in this new frontier. The Charter of the United Nations, humanitarian principles and human rights are not suggestions — they remain the foundation of any military applications of AI. Algeria approaches those challenges through strategic initiatives that bridge security and development. Our national research and innovation strategy on artificial intelligence for 2020–2030 provides a comprehensive framework for responsible AI development, supported by our national scientific council for AI and specialized higher education institutions. Earlier this month, Algeria hosted the third African Start-up Conference, bringing together over 500 start-ups from 50 African countries and demonstrating our continent’s unified commitment to developing AI ecosystems that serve our peoples’ needs while respecting security imperatives. Africa’s Continental Artificial Intelligence Strategy and the African Digital Compact are not only policies — they are our vision of AI intelligence as a force for peace, security and positive transformation.
To turn that vision into reality and respond to those challenges, five key measures are important:
First, the creation of inclusive international mechanisms for AI security expertise, where developing countries are not just observers but equal architects of our shared future.
Secondly, the establishment of a global framework for responsible AI that balances national sovereignty
with international cooperation — not one at the expense of the other.
Thirdly, the launch of targeted capacity-building programmes in AI education and security assess ment, because knowledge not only equals power, but also survival.
Fourthly, the development of transparent protocols for international artificial intelligence safety cooperation, because in the digital age, transparency is not optional.
Fifthly and lastly, the building of robust digital infrastructure and support for developing countries in creating artificial intelligence strategies, because no nation should be left behind in this revolution.
The time has come — not tomorrow, not next year, but now — for a binding framework that prevents the misuse of military artificial intelligence in ways that violate international law and international humanitarian law. Let us consider what we have already achieved, namely, the approval this year of the draft United Nations Convention against Cybercrime by the Ad Hoc Committee to Elaborate a Comprehensive International Convention on Countering the Use of Information and Communications Technologies for Criminal Purposes, which is chaired by my country, Algeria. That shows that meaningful multilateral action is not just possible — it is imperative. The choice before us cannot be partial: we can let artificial intelligence deepen inequalities and create new security nightmares or we can harness it as a force for peace and security and development for all nations.
I welcome Mr. Antony J. Blinken, Secretary of State of the United States of America, to the Security Council as President. Let me also thank the United States presidency for convening this important high-level meeting on artificial intelligence and the maintenance of international peace and security. I thank Mr. António Guterres, Secretary- General of the United Nations, for his insightful briefing. We take note of the thought-provoking contributions of Mr. Yann LeCun and Ms. Fei-Fei Li.
The topic of today’s high-level briefing is not only timely, but highly relevant, as we acknowledge the significant implications of artificial intelligence (AI) for the global landscape and present-day realities. That includes its potential to transform societies and economies, and, equally, the risks that it poses to peace
and security, particularly when misapplied in military contexts. The rapid development of AI technology, as highlighted in the outcomes of ongoing international dialogues, including the Secretary-General’s July 2023 policy brief, the Global Digital Compact annexed to the Pact for the Future (General Assembly resolution 79/1), and the General Assembly resolutions introduced by the United States and China (General Assembly resolutions 78/265 and 78/311), presents both tremendous opportunities and significant challenges. AI holds the potential to address a wide range of global issues within the framework of sustainable development.
However, as AI becomes increasingly pervasive, we must also recognize its ability to disrupt peace and security in both the civilian and the military domains. In that regard, Sierra Leone is of the view that AI can play a pivotal role in maintaining peace and security by enhancing decision-making, improving situational awareness, and enabling proactive conflict prevention and management. We therefore fully support the relevant ongoing international dialogue and initiatives, including the High-Level Advisory Body on Artificial Intelligence, which underscores the importance of creating a robust international ecosystem for AI governance.
As our briefers and other Council members have already noted, AI holds immense promise for advancing sustainable development, improving governance and fostering peace. However, its unregulated deployment risks exacerbating inequality, fuelling conflict and undermining the dignity of human life. For post-conflict and developing countries striving to sustain peace, the stakes could not be higher. For the African continent, the misuse of AI poses unique risks. In most parts of Africa, where digital infrastructure is still developing, there is a growing vulnerability to AI-enabled disinformation campaigns, which could destabilize fragile social fabrics and undermine democratic processes. Additionally, the potential militarization of AI, if left unchecked, could exacerbate regional insecurities, threaten peacekeeping and endanger civilian protections. Accordingly, the principles outlined in the communiqué of the Peace and Security Council of the African Union on Artificial Intelligence, adopted on 13 June, are particularly resonant for Sierra Leone. We strongly agree that AI should be human-centric, prioritizing the well-being of people over the interests of technology companies or military objectives. That includes addressing concerns about bias in AI systems, ensuring the right to privacy
and protections against the misuse of AI for surveillance and other forms of social control.
The AU’s emphasis on the responsible development and governance of AI aligns with our belief that the international community must act proactively to ensure that AI is developed and deployed in a manner that promotes peace, equity and global cooperation, while preventing its misuse. The key aspects of the communiqué of the Peace and Security Council of the African Union on AI include instructing the AU Commission to undertake a comprehensive study on the impact and implications of AI on peace, security, stability, democracy and development in Africa; mainstreaming AI in all peace processes, including facilitating its integration in peace initiatives; leveraging AI for mediation, reconciliation and post- conflict reconstruction; and developing frameworks to oversee the responsible integration of AI in military operations, ensuring compliance with ethical standards and international humanitarian law, among others. Sierra Leone therefore has been and remains supportive of initiatives at the international and regional levels that call for timely action and cooperation to ensure the deployment and use of safe, secure and trustworthy AI systems for sustainable development and peace.
It is critical to expand the participation of all countries, in particular developing countries, in digital transformation in order to harness the benefits and effective participation, including by capacity-building. Promoting knowledge-sharing activities and the transfer of technology on mutually agreed terms is an important aspect of capacity development. Further, AI capacity- building should not only be about technology transfer, but also about building local expertise, strengthening regulatory frameworks and ensuring that countries have the necessary legal and institutional structures to manage the risks associated with AI. That will require coordinated efforts from international organizations, including the United Nations, and regional bodies like the African Union. We therefore support the establishment of a global AI fund, as proposed by the High-Level Advisory Body on Artificial Intelligence, to ensure that developing countries have the resources and technical support needed to safely and effectively integrate AI into our economies and societies.
In addressing the key questions for today’s briefing on how AI can contribute to the maintenance of international peace and security, Sierra Leone wishes to underscore the following four points.
First, on conflict prevention and early warning systems, while respecting the principles of the Charter of the United Nations, AI algorithms can analyse vast datasets, including economic indicators, political events and social media trends to predict tensions, potential conflicts or instability. Also, machine-learning models can detect patterns of escalating tensions by monitoring, inter alia, troop movements or trade disruptions. Furthermore, AI can identify regions at risk of instability and suggest an optimal allocation of humanitarian aid or peacekeeping resources to prevent the escalation of potential conflict situations.
Secondly, AI can facilitate diplomacy and conflict resolution through the analysis of data to identify areas of misunderstanding or potential consensus. Additionally, AI can gauge public and leadership sentiments in conflict regions, helping mediators design more effective peacebuilding strategies.
Thirdly, AI can assist in optimizing the deployment of peacekeeping forces by analysing data on conflict intensity, geography and logistics. More importantly, AI systems can process real-time inputs from civilians and organizations to map conflict zones, refugee movements and humanitarian needs.
Fourthly and finally, as AI technologies become more integrated into military applications, with the evident risks to international security, the use of AI in military systems, especially autonomous weapons, must address the profound concerns regarding accountability, the potential for unintended escalation and compliance with the law on the use of force and international humanitarian law. The military application of AI should be governed by clear international standards that prioritize human oversight and adhere to international humanitarian law. There must be transparency, dialogue and multilateral engagement on this issue to ensure that AI in the military domain does not contribute to destabilization or the erosion of international norms.
In conclusion, Sierra Leone is fully in support of peaceful uses of AI, especially in terms of enhancing peacebuilding and sustaining peace initiatives. That is why we collaborated with Slovenia and Switzerland in calling for joint action to implement the Secretary- General’s New Agenda for Peace during our August presidency of the Council. Within that context, the aforementioned trio of countries, during the Swiss presidency in October, we also called on the Council to act early and manage uncertainty by seeking expert
advice and collaborate with academic and research institutions to provide the Council with evidence-based information and other relevant assistance to help prevent conflict. AI can ultimately assist in that regard.
I would like to thank the Secretary-General for his briefing. I also thank Mr. LeCun and Ms. Li for sharing their perspectives.
Slovenia considered it very important that the voice of all, including those from the private sector and academia, which are the engines of fast development and change, be heard in our deliberations on digital and emerging technologies.
The magnitude of the artificial intelligence (AI) revolution is evident not only in its impact on our social fabric, our economies, science and, increasingly, warfare, but also in the parallel efforts to establish effective governance. With two General Assembly resolutions on AI adopted by consensus (General Assembly resolutions 78/265 and 78/311) and the adoption of the Global Digital Compact (General Assembly resolution 79/1, annex I) earlier this year, the United Nations has recognized its unique convening power and has taken the lead in the global conversation on AI governance, alongside various other initiatives. Moreover, important work within the United Nations is under way regarding the establishment of a global dialogue on AI governance, and rightly so. AI governance is becoming increasingly essential to global stability and requires inclusive multilateral engagement at all levels, including within the Security Council.
As the Secretary-General has noted, artificial in telligence represents both the greatest opportunity and, at the same time, an existential threat. When it comes to global stability, AI can, on the one hand, be harnessed positively, for example, to support efforts for conflict prevention and peacebuilding. On the other hand, AI can enable complex, cross-border threats, such as cy beroperations, disinformation campaigns, violent ex tremist content and terrorist propaganda. It can be used to fuel the proliferation of autonomous weapons, further complicating efforts to ensure oversight and account ability. We therefore strongly believe that the Council’s primary role in maintaining international peace and security necessitates addressing AI-related risks to en sure that AI is designed, developed, deployed and used in full compliance with international law, in particu lar international humanitarian and human rights law.
The risks are not speculative or distant; they are a reality in contemporary conflicts, such as in Gaza, where reports indicate that artificial intelligence is being used as a tool to identify targets, particularly in densely populated residential areas, resulting in significant numbers of civilian casualties. In that regard, we advocate for the integration of AI-related discussions into the Council, including through briefings such as this one to ensure that members pay attention and provide responses to those risks within existing Council mandates and other geographic and thematic issues. By fostering synergies to strengthen AI governance for international peace and security, we reaffirm the Council’s complementary role in supporting the efforts pursued in the General Assembly and other forums engaged in AI governance.
Japan thanks the United States presidency for taking the initiative in organizing today’s meeting. I also thank Secretary-General António Guterres, Mr. Yann LeCun and Ms. Fei-Fei Li for their informative and insightful interventions.
We have a common understanding that artificial intelligence (AI) offers both opportunities and risks. Depending on how it is used, it can bring prosperity or pose a threat to humankind. The irresponsible use of AI in the military field could lead to the outbreak and escalation of conflicts. The misuse of AI can cause the dissemination and spread of disinformation, thereby undermining democracy and abusing human rights. In the non-military domain, we must ensure safe, secure and trustworthy AI, while maximizing its potential and mitigating its risks. The rule of law and human security should serve as foundational principles when we develop and use AI responsibly. Japan has been pursuing an AI strategy underscoring a society in which human dignity is respected. It is also essential to close all digital divides so that all countries and people can utilize AI without being left behind. Digital cooperation, including capacity-building and infrastructure development related to AI, should be promoted.
From that perspective, Japan co-sponsored General Assembly resolution 78/265, adopted in March, which was a major step forward. Let me thank again the United States for spearheading that effort. The Global Digital Compact adopted in September (General Assembly resolution 79/1, annex I) will also serve as a policy guideline for future efforts.
With respect to AI in the non-military domain, we see a growing need for building up global discussions on AI governance. It is also important to ensure interoperability among different AI governance frameworks. Under the Hiroshima AI Process, which was launched last year under Japan’s initiative, the international guiding principles for all AI actors and the international code of conduct for organizations developing advanced AI systems were developed. We are now promoting their implementation, including through the Hiroshima AI Friends Group, and will continue to cooperate to realize safe, secure and trustworthy AI.
To ensure the responsible use of AI in the military domain, appropriate measures should be implemented throughout the life cycle of military AI capabilities. Our commitment to upholding human- centred principles and applying international legal obligations, including international humanitarian law,
is essential. In that vein, Japan supports the initiatives of the Responsible Artificial Intelligence in the Military Domain Summit and the Political Declaration on Responsible Military Use of Artificial Intelligence and Autonomy. Regarding lethal autonomous weapons systems, the Group of Governmental Experts on Emerging Technologies in the Area of Lethal Autonomous Weapons Systems should continue to play a central role. Japan will actively and constructively contribute to effective cooperation for the sake of implementation by the entire international community.
In conclusion, AI has broad and diverse impacts in unexpected ways. As expressed at its high-level briefing in October (see S/PV.9753), going forward, the Security Council must systematically consider the impact of scientific and technological developments on international peace and security.
The meeting rose at 11.50 a.m.