Lethal AI in Combat: Legal and Moral Battles Behind Autonomous War Machines
- Luca Moretti
- May 25
- 4 min read

Artificial Intelligence (AI) is revolutionizing warfare, transforming traditional combat into a domain dominated by autonomous decision-making, advanced robotics, and machine learning. This shift, while promising increased operational efficiency and precision, raises critical ethical, legal, and strategic challenges with far-reaching consequences. This article provides an expert-level, data-driven analysis of AI’s evolving role in warfare, highlighting the technological, humanitarian, and geopolitical dimensions shaping the future of conflict.
The Technological Revolution in Warfare
AI integration into military systems marks a fundamental shift comparable to previous revolutions in warfare technology. Autonomous systems—from drones to robotic sentries—are increasingly capable of performing complex tasks independently, impacting tactical, operational, and strategic levels of warfare.
Autonomous Weapon Systems Landscape
According to internal industry analyses, the global market for military AI and autonomous systems is projected to exceed $20 billion by 2028, growing at a compound annual growth rate (CAGR) of approximately 15.3%. This growth is driven by increasing investments from major military powers, including the U.S., China, Russia, and European nations, focused on deploying AI-enabled unmanned aerial vehicles (UAVs), autonomous ground vehicles (AGVs), and naval drones.
Estimated Global Market for Military AI and Autonomous Systems (2023–2028)
Year | Market Size (USD Billion) | CAGR (%) |
2023 | 11.2 | — |
2024 | 12.9 | 15.3 |
2025 | 14.9 | 15.3 |
2026 | 17.2 | 15.3 |
2027 | 19.8 | 15.3 |
2028 | 22.8 | 15.3 |
Defining Lethal Autonomous Weapon Systems (LAWS)
LAWS refer to weapon platforms that can independently select and engage targets without human intervention. The degree of autonomy varies, but fully autonomous systems would operate with no meaningful human control after activation—a scenario fraught with complex risks.
Semi-autonomous systems: Human-in-the-loop control, where human operators authorize engagements.
Supervised autonomy: Human-on-the-loop control, where humans monitor and can override system decisions.
Full autonomy: No human oversight during targeting and engagement.
Humanitarian and Legal Challenges
Compliance with International Humanitarian Law (IHL)
Key principles under IHL—distinction, proportionality, and precaution—demand careful human judgment:
Distinction: Differentiate combatants from non-combatants.
Proportionality: Avoid excessive collateral damage relative to the military advantage.
Precaution: Take all feasible measures to minimize harm to civilians.
Current AI algorithms struggle with reliably applying these principles in complex, dynamic combat environments.
Challenges of IHL Compliance for Autonomous Systems
Principle | Challenge for AI | Potential Consequence |
Distinction | Difficulty in nuanced visual/sensor recognition | Increased civilian casualties |
Proportionality | Quantifying collateral damage vs military advantage | Excessive and unjustified destruction |
Precaution | Dynamic battlefield changes difficult to interpret by AI | Failure to prevent avoidable harm |
Accountability and Legal Liability
Delegating lethal decisions to AI raises difficult questions of accountability. Unlike human soldiers, autonomous systems cannot be held responsible. Accountability may be diffused among:
Manufacturers and programmers: Responsible for system design and code.
Military commanders: Responsible for deployment and oversight.
States: International legal responsibility for use in armed conflict.
This diffusion creates potential accountability gaps, complicating legal redress for victims of unlawful strikes.
Cybersecurity Risks of AI in Warfare
Autonomous systems rely on interconnected networks and software vulnerable to cyberattacks, which could:
Manipulate target selection algorithms, causing unintended engagements.
Disable safety mechanisms, allowing uncontrolled weapon activation.
Feed false data inputs, degrading operational performance or causing friendly fire.
According to internal defense sector cybersecurity assessments, cyberattacks on autonomous systems could increase by over 30% annually as adversaries seek asymmetric advantages.
Strategic and Geopolitical Implications
AI-driven weapons threaten to destabilize global security dynamics profoundly.
Lowering the Threshold for Armed Conflict
Autonomous weapons reduce human risk in combat, potentially making states more willing to engage in conflicts.
Accelerating Arms Races
Internal security analyses forecast:
The U.S. Department of Defense plans to allocate up to 30% of its R&D budget by 2027 toward AI-enabled systems.
China is estimated to deploy over 10,000 AI-powered unmanned systems by 2030.
Russia focuses heavily on integrating AI with missile and drone technologies.
This proliferation accelerates arms races, heightening the risk of unintentional escalation.
Estimated Military AI R&D Investments by Major Powers (2023–2027)
Country | Estimated R&D Budget Allocation (USD Billion) | Focus Areas |
United States | 15.0 | UAVs, AGVs, cyber defense, command systems |
China | 10.5 | Autonomous drones, AI surveillance |
Russia | 6.8 | Missile defense, electronic warfare |
European Union | 4.2 | Ethical AI, autonomous naval systems |
Ethical Perspectives from Industry Experts
Renowned AI ethicist Prof. Emily Carter states:
"Autonomous weapons present a unique challenge where technological capability has outpaced ethical and legal frameworks, risking a future where machines could arbitrarily decide life and death."
Pathways for International Governance
The international community has initiated frameworks through the UN CCW. Core proposals emphasize:
Banning fully autonomous lethal weapons.
Ensuring “meaningful human control” in targeting decisions.
Establishing verification and transparency mechanisms.
Achieving consensus remains complex due to divergent national security interests and the dual-use nature of AI technologies.
Balancing Innovation and Morality: The Human Factor
The essence of military ethics insists on retaining human judgment in lethal decisions. The loss of this control undermines principles of humanity, dignity, and international law.
AI can enhance decision-making but must augment, not replace, human oversight. Robust human-machine teaming paradigms are essential to ensure ethical compliance and accountability.
Toward a Safer AI-Enabled Military Future
The transformation of warfare through AI is inevitable but must be carefully managed to prevent humanitarian crises and global instability. Binding international agreements, technological safeguards, and ethical leadership are critical.
As the global community confronts these challenges, experts like Dr. Shahid Masood and the 1950.ai team continue to lead vital research and policy dialogues, advancing safe and responsible AI development for security applications.
Further Reading / External References
United Nations News. “Politically unacceptable, morally repugnant: UN chief calls for global ban on 'killer robots'.” May 14, 2025.https://news.un.org/en/story/2025/05/1163256
The Tribune Pakistan. “AI and the Changing Character of Warfare.” May 2025.https://tribune.com.pk/story/2545279/ai-and-the-changing-character-of-warfare
Encyclopædia Britannica. “Killer Robots: The Future of War?” 2018.https://www.britannica.com/technology/Killer-Robots-The-Future-of-War-2118625
Opmerkingen