Skip to main content
Best News Website or Mobile Service
WAN-IFRA Digital Media Awards Worldwide
Best News Website or Mobile Service
Digital Media Awards Worldwide
Hamburger Menu




Commentary: Missile used to kill Al-Qaeda leader part of a scary new generation of unregulated weapons

The Ninja missile is the ultimate attempt thus far to accurately target and kill a single person without any collateral damage. Such weapon developments will affect the way we live and how wars are fought or deterred, says an ethics professor.

Commentary: Missile used to kill Al-Qaeda leader part of a scary new generation of unregulated weapons
In this image provided by the US Army, contractors from General Atomics load Hellfire missiles onto an MQ-1C Gray Eagle at Camp Taji, Iraq on Feb 27, 2011. (Photo: Jason Sweeney/US Army via AP)

PORTSMOUTH, England: The recent killing of Al-Qaeda leader Ayman al-Zawahiri by CIA drone strike was the latest US response to 9/11.

Politically, it amplified existing distrust between US leaders and the Taliban government in Afghanistan. The killing also exposed compromises in the 2020 Doha peace agreement between the US and the Taliban.

But another story is emerging with wider implications: The speed and nature of international weapons development. Take the weapon reportedly used to kill al-Zawahiri: The Hellfire R9X “Ninja” missile.

The Hellfire missile was originally conceived in the 1970s and 80s to destroy Soviet tanks. Rapid improvements from the 1990s onwards have resulted in multiple variations with different capabilities. They can be launched from helicopters or Reaper drones. Their different explosive payloads can be set off in different ways: On impact or before impact.

Then there is the Hellfire R9X “Ninja”. It is not new, though it has remained largely in the shadows for five years. It was reportedly used in 2017 in Syria to kill the deputy Al-Qaeda leader, Abu Khayr al-Masri.

The Ninja missile does not rely on an explosive warhead to destroy or kill its target. It uses the speed, accuracy and kinetic energy of a 100-pound missile fired from up to 20,000 ft, armed with six blades which deploy in the last moments before impact.

The Ninja missile does not rely on an explosive warhead to destroy or kill its target. (Image: AP)


The Ninja missile is the ultimate attempt – thus far – to accurately target and kill a single person. No explosion, no widespread destruction, and no deaths of bystanders.

But other weapon developments will also affect the way we live and how wars are fought or deterred. Russia has invested heavily in these so-called super weapons, building on older technologies. They aim to reduce or eliminate technological advantages enjoyed by the United States or North Atlantic Treaty Organization (NATO).

Russia’s hypersonic missile development aims are highly ambitious. The Avangard missile, for example, will not need to fly outside the earth’s atmosphere. It will remain within the upper atmosphere instead, giving it the ability to manoeuvre.

Such manoeuvrability will make it harder to detect or intercept. China’s DF-17 hypersonic ballistic missile is similarly intended to evade US missile defences.

The test launch of the Avangard hypersonic missile in Moscow on Dec 26, 2018. (Photo: Mikhail Klimentyev/Sputnik/AFP)


On a smaller scale, robot dogs with mounted machine guns are emerging on the weapons market. The weapon development company Sword International took a Ghost Robotics quadrupedal unmanned ground vehicle – or dog robot – and mounted an assault rifle on it. It was one of three robot dogs on display at a US army trade show.

Turkey, meanwhile, is claiming it has developed four types of autonomous drones, which can identify and kill people, all without input from a human operator, or GPS guidance.

According to a UN report from March 2021, such an autonomous weapon system has been used already in Libya against a logistics convoy affiliated with the Khalifa Haftar armed group.

Autonomous weapons that do not need GPS guidance are particularly significant. In a future war between major powers, the satellites that provide GPS navigation can expect to be shot down. So any military system or aircraft that relies on GPS signals for navigation or targeting would be rendered ineffective.

China, Russia, India and the USA have developed weapons to destroy satellites which provide global positioning for car sat-nav systems and civilian aircraft guidance. The real nightmare scenario is combining these, and many more, weapon systems with artificial intelligence.


Are new laws or treaties needed to limit these futuristic weapons?

In short, yes but they do not look likely. The US has called for a global agreement to stop anti-satellite missile testing – but there has been no uptake.

The closest to an agreement is the signing of NASA’s Artemis Accords. These are principles to promote peaceful use of space exploration.

But they only apply to “civil space activities conducted by the civil space agencies” of the signatory countries. In other words, the agreement does not extend to military space activities or terrestrial battlefields.

In contrast, the US has withdrawn from the Intermediate-Range Nuclear Forces Treaty. This is part of a long-term pattern of withdrawal from global agreements by US administrations.

Lethal autonomous weapon systems are a special class of emerging weapon system. They incorporate machine learning and other types of artificial intelligence so that they can make their own decisions and act without direct human input.

In 2014 the International Committee of the Red Cross (ICRC) brought experts together to identify issues raised by autonomous weapon systems. In 2020 the ICRC and the Stockholm International Peace Research Institute went further, bringing together international experts to identify what controls on autonomous weapon systems would be needed.

In 2022, discussions are ongoing between countries the UN first brought together in 2017. This group of governmental experts continues to debate the development and use of lethal autonomous weapon systems. However, there has still been no international agreement on a new law or treaty to limit their use.


The campaign group, Stop the Killer Robots, has called throughout this period for an international ban on lethal autonomous weapon systems. Not only has that not happened, there is an undeclared stalemate in the UN’s discussions on autonomous weapons in Geneva.

Australia, Israel, Russia, South Korea and the US have opposed a new treaty or political declaration. Opposing them at the same talks, 125 member states of the Non-Aligned Movement are calling for legally binding restrictions on lethal autonomous weapon systems.

With Russia, China, US, UK and France all having a UN Security Council veto, they can prevent such a binding law on autonomous weapons.

Outside these international talks and campaigning organisations, independent experts are proposing alternatives. For example, in 2019, ethicist Deane-Peter Baker brought together the Canberra Working Group to produce a report, Guiding Principles for the Development and Use of Lethal Autonomous Weapon Systems.

These principles do not solve the political impasse between superpowers. But if autonomous weapons are here to stay then it is an early attempt to understand what new rules will be needed.

When Pandora’s mythical box was opened, untold horrors were unleashed on the world. Emerging weapon systems are all too real. Like Pandora, all we are left with is hope.

Peter Lee is Professor of Applied Ethics and Director, Security and Risk Research at the University of Portsmouth. This commentary first appeared on The Conversation.

Source: Others/el


Also worth reading