Artificial intelligence and the risks of a ‘hyper-war’
Unleashing a legal and ethical debate worldwide, AI (artificial intelligence) is progressing with leaps and bounds as it portends to change human society forever. For example, if a driverless car meets with an accident involving fatalities, it is the algorithm operator who faces “product liability” rules. In the case of AI used in conventional war, machines killing humans is an ethically chilling concept. Carrying major implications, it is feared that the proto-AI technologies of today are going to evolve into true AI super-intelligence very rapidly without giving enough time for research into the pros and cons.
As apprehensions of a “hyper-war scenario” build up, the main challenge remains: how to place the human factor in AI and prevent a drastic downgrade in military security as combat involving the technology changes the dimensions of warfare. Every country today needs to re-evaluate its defense mechanisms and reinterpret its geostrategic defenses to fit in with the modern use of artificial intelligence.
Discussing the risks of “hyper-war,” August Cole, senior fellow at the Atlantic Council, predicts that, “The decision-making speed of machines is going to eclipse the political and civilian ability.” Being dual-use in nature, most AI algorithms can also be modified for security purposes and preparing for a “hyper-war” will soon be a priority.
The US and China have already announced that they intend to harness AI for military use. Recognising the military significance of AI, Russian President Vladimir Putin has termed it the future for all mankind that would introduce “colossal opportunities” and “threats that are difficult to predict.” Declaring that the country leading in artificial intelligence will rule the world, Putin felt the gravest of threats would be that involving nuclear stability.
Every country today needs to re-evaluate its defense mechanisms and reinterpret its geostrategic defenses to fit in with the modern use of artificial intelligence
Exploring the possibilities of nuclear mishaps due to AI, the RAND Corporation started a project known as Security 2040, one of the researchers, engineer Andrew Lohn, says, “This isn’t just a movie scenario, things that are relatively simple can raise tensions and lead us to some dangerous places if we are not careful.”
Basically, the fear is that computer miscalculations could lead to nuclear annihilation if the machines taught to think and learn like humans suddenly go haywire and spin out into a ‘Terminator’ kind of nightmare. Redefining the rules of nuclear deterrence, there is a probability that the “red button” may not be in human control if things go wrong.
Depending on data that can be analyzed in real time, an accessible, data-friendly ecosystem with cross-sharing enabled is imperative, thus it is nations that promote open data sources and data sharing that advance most in AI. However, there are risks that even the most sophisticated signature-based cyber protection will be constantly exposed to malware and cyber threats so whoever wins the race for AI superiority also has massive responsibilities to secure the systems sufficiently. The US ranks eighth in data-openness ratings, while China ranks at 93rd, according to a McKinsey Global Institute study.
Ahead of other countries, the US is experimenting with autonomous boats that track submarines for thousands of miles while China explores the use of “swarm intelligence” to enable teams of drones to hunt in unison. Planning an underwater doomsday drone, Russia could deliver powerful nuclear warheads to vaporize entire cities.
Employing AI technology in the autonomous weapons Project Maven, the US military is trying “to sift through the massive troves of data and video captured by surveillance and then alert human analysts of patterns or when there is abnormal or suspicious activity.” Calling on Google to abandon the US military project, more than 3,000 employees of the internet giant signed an open letter declaring that it should not be in “the business of war and outsource moral responsibility.” They also expressed the fear that Google locations history would be used for target killing. Currently, even the CIA is executing around 137 pilot projects related to artificial intelligence.
According to McKinsey Global Institute, “AI-led automation can give the Chinese economy a productivity injection that would add 0.8 to 1.4 percentage points to GDP growth annually, depending on the speed of adoption.” Thus, it is not surprising that a $150-billion domestic AI industry is planned according to China’s State Council to provide security, fight terrorism and improve speech recognition programs, and to begin with a facial recognition application by Baidu is being used to search for missing people. Even though China is behind the US and UK in AI, the sheer size of its market gives it an advantage in pilot testing and product development.
Evaluating the positive aspects of AI, it may be possible that these technologies could be suitably harnessed over time and provide stability and, ultimately, security that is reliable as there would be no room for human error. Until that time arrives, proactive measures should be taken to rule out the dangers to humanity and prevent catastrophic miscalculations between nuclear powers.