Emerging technologies including artificial intelligence, lethal autonomous weapons systems, and hypersonic missiles pose a potentially existential threat that underscores the imperative of arms control measures to slow the pace of weaponization, according to a new report published Tuesday.
The Arms Control Association report – entitled Assessing the Dangers: Emerging Military Technologies and Nuclear (In)Stability – "unpacks the concept of 'emerging technologies' and summarizes the debate over their utilization for military purposes and their impact on strategic stability."
The publication notes that the world's military powers "have sought to exploit advanced technologies – artificial intelligence, autonomy, cyber, and hypersonics, among others – to gain battlefield advantages" but warns too little has been said about the dangers these weapons represent.
"Some officials and analysts posit that such emerging technologies will revolutionize warfare, making obsolete the weapons and strategies of the past," the report states. "Yet, before the major powers move quickly ahead with the weaponization of these technologies, there is a great need for policymakers, defense officials, diplomats, journalists, educators, and members of the public to better understand the unintended and hazardous outcomes of these technologies."
A new @ArmsControlNow report assesses the extent to which military use of emerging tech could result in an accidental use of nuclear weapons in a crisis, and provides a framework for curtailing the indiscriminate weaponization of such tech.
Available at https://t.co/gPyDbcaOcdpic.twitter.com/fw55Gczx23
— Arms Control Assoc (@ArmsControlNow) February 7, 2023
Lethal autonomous weapons systems – defined by the Campaign to Stop Killer Robots as armaments that operate independent of "meaningful human control" – are being developed by nations including China, Israel, Russia, South Korea, the United Kingdom, and the United States. The US Air Force's sci-fi-sounding Skyborg Autonomous Control System, currently under development, is, according to the report, "intended to control multiple drone aircraft simultaneously and allow them to operate in 'swarms,' coordinating their actions with one another with minimum oversight by human pilots."
"Although the rapid deployment of such systems appears highly desirable to many military officials, their development has generated considerable alarm among diplomats, human rights campaigners, arms control advocates, and others who fear that deploying fully autonomous weapons in battle would severely reduce human oversight of combat operations, possibly resulting in violations of international law, and could weaken barriers that restrain escalation from conventional to nuclear war," the report notes.
The latter half of the 20th century witnessed numerous nuclear
close calls, many based on misinterpretations, limitations, or outright failures of technology. While technologies like artificial intelligence (AI) are often touted as immune to human fallibility, the research suggests that such claims and hubris could have deadly and unforeseen consequences.
"An increased reliance on AI could lead to new types of catastrophic mistakes," a 2018 report by the Rand Corporation warned. "There may be pressure to use it before it is technologically mature; it may be susceptible to adversarial subversion; or adversaries may believe that the AI is more capable than it is, leading them to make catastrophic mistakes."
While the Pentagon in 2020 adopted five principles for what it calls the "ethical" use of AI, many ethicists argue the only safe course of action is a total ban on lethal autonomous weapons systems.
Hypersonic missiles, which can travel at speeds of Mach 5 – five times the speed of sound – or faster, are now part of at least the US, Chinese, and Russian arsenals. Last year, Russian officials acknowledged deploying Kinzhal hypersonic missiles three times during the country's invasion of Ukraine in what is believed to be the first-ever use of such weapons in combat. In recent years, China has tested multiple hypersonic missile variants using specially designed high-altitude balloons. Countries including Australia, France, India, Japan, Germany, Iran, and North Korea are also developing hypersonic weapons.
The report also warns of the escalatory potential of cyberwarfare and automated battlefield decision-making.
"As was the case during World Wars I and II, the major powers are rushing ahead with the weaponization of advanced technologies before they have fully considered – let alone attempted to mitigate – the consequences of doing so, including the risk of significant civilian casualties and the accidental or inadvertent escalation of conflict," Michael Klare, a board member at the Arms Control Association and the report's lead author, said in a statement.
"While the media and the US Congress have devoted much attention to the purported benefits of exploiting cutting-edge technologies for military use, far less has been said about the risks involved," he added.
DARPA’s HAWC program is a wrap…concluding with a successful @LockheedMartin#hypersonic missile flying more than 300 nautical miles and lots of data for the @usairforce. More: https://t.co/Yqq2Xl50jnpic.twitter.com/ilNN4xz0z4
— DARPA (@DARPA) January 30, 2023
The report asserts that bilateral and multilateral agreements between countries that "appreciate the escalatory risks posed by the weaponization of emerging technologies" are critical to minimizing those dangers.
"As an example of a useful first step, the leaders of the major nuclear powers could jointly pledge to eschew cyberattacks" against each other's command, control, communications, and information (C3I) systems, the report states. A code of conduct governing the military use of artificial intelligence based on the Pentagon's AI ethics principles is also recommended.
"If the major powers are prepared to discuss binding restrictions on the military use of destabilizing technologies, certain priorities take precedence," the paper argues. "The first would be an agreement or agreements prohibiting attacks on the nuclear C3I systems of another state by cyberspace means or via missile strikes, especially hypersonic strikes."
"Another top priority would be measures aimed at preventing swarm attacks by autonomous weapons on another state's missile submarines, mobile ICBMs, and other second-strike retaliatory systems," the report continues, referring to intercontinental ballistic missiles. "Strict limitations should be imposed on the use of automated decision-support systems with the capacity to inform or initiate major battlefield decisions, including a requirement that humans exercise ultimate control over such devices."
"Without the adoption of measures such as these, cutting-edge technologies will be converted into military systems at an ever-increasing tempo, and the dangers to world security will grow apace," the publication concluded. "A more thorough understanding of the distinctive threats to strategic stability posed by these technologies and the imposition of restraints on their military use would go a long way toward reducing the risks of Armageddon."
Brett Wilkins is is staff writer for Common Dreams. Based in San Francisco, his work covers issues of social justice, human rights and war and peace. This originally appeared at CommonDreams and is reprinted with the author’s permission.
All you have to do is watch the original Terminator movie to see the potential dangers of this kind of stuff.
yes robots are scary – until they reach the end of their extension cord.
All these reports hack at the branches. The root problem is that we’re violent hominids rapidly acquiring fairly godlike powers of destruction, and it can’t be stopped. Humanity must adapt to it. Rules and words and protocols won’t be enough. We have to change as a species.
I couldn’t agree more! The biggest problem on the planet by far is that humans evolved mentally and spiritually in the wrong direction, or didn’t evolve at all in those ways, depending on how you want to look at it. Our only legitimate role on this planet is to expand our consciousness, but instead we artificially and very harmfully manipulate the physical/natural world, to the great detriment of the Earth and all the naturally-evolved life here (not talking about domesticated animals that were bred into existence by people).
Humans should focus on expanding our empathy — with all life, not just other people — and wisdom. Instead, we’ve obsessed on intellect and ego, totally the wrong ways to go. I understand that intellect can be a very valuable tool, but that’s all it is. And while at least some ego is necessary to live as corporeal beings, the less of it, the better.
The problem is unregulated Capitalism and the wealth-hoarding Elite Class’ total control of society and their effective propaganda messaging. Neoliberalism advocates are nihilists and PMC’s who don’t care one iota about rising temperatures, deforestation, war profiteering, poverty, stagflation and the like. All this Political charade between the parties, Dem vs Rep, only serves the PMC. It’s the bourgeoisie arguing with each other about how best to save Capitalism. Read up on Elite Overproduction by Peter Turchin.
Sorry, but no. Capitalism, while a cancer that must be removed, is just a recent symptom. These problems started at least 10,000 years ago when humans began using agriculture instead of living naturally, and possibly even 60-90,000 years ago when humans began leaving Africa and caused extinctions everywhere they went. Your comment is limited to human harms, which is not my focus.
IMO modern humans don’t really fit in on this planet, almost as if we didn’t really evolve here
As a TV series character said, when nature created humans, she created something outside of herself. Or maybe we’re actually Golgafrinchans as Douglas Adams said. Either way, I agree. If humans won’t live naturally and in balance with the Earth and its ecosystems, they should just get off the planet.
Very scary stuff. I can only imagine that the future holds more fantastical and horrific weapon.
Pingback: This week in nuclear news | Nuclear Australia
Comments are closed.