“This technology is our future threat,” warns Serhiy Beskrestnov as he inspects a recently captured Russian drone. Unlike traditional weapons, it uses artificial intelligence to locate and strike targets without human guidance.
Beskrestnov, a consultant for Ukraine’s defence forces, has analysed countless drones since the invasion began. This model stands apart. It neither sends nor receives signals, making it impossible to jam or detect.
Both Russian and Ukrainian forces now deploy AI in combat. They use it to track enemy positions, process intelligence, and clear mines faster than ever before.
AI becomes a strategic advantage
Artificial intelligence has become essential for Ukraine’s military. “Our forces receive over 50,000 video streams from the front every month,” says Deputy Defence Minister Yuriy Myronenko. “AI analyses the footage, identifies threats, and maps them for commanders.”
The technology accelerates decision-making, optimises resources, and helps reduce casualties. Its most visible impact is in unmanned systems. Ukrainian troops now operate drones that lock onto targets and complete the final attack phase autonomously.
These drones are nearly impossible to jam and extremely difficult to shoot down. Experts predict they will evolve into fully autonomous weapons capable of finding and eliminating targets independently.
Drones that operate on their own
“All a soldier needs to do is press a button on a smartphone,” explains Yaroslav Azhnyuk, CEO of Ukrainian tech company The Fourth Law. “The drone finds its target, drops explosives, assesses the damage, and returns to base. No piloting skills are required.”
Azhnyuk says these drones could significantly strengthen Ukraine’s air defences against Russian long-range drones like the Shaheds. “A computer-guided system can outperform humans,” he adds. “It reacts faster, sees more clearly, and moves more precisely.”
Myronenko admits fully autonomous systems are still in development but says Ukraine is close. “We have partly integrated the technology into some devices,” he notes. Azhnyuk predicts thousands of these drones could be operational by the end of 2026.
Progress carries major risks
Full automation brings serious dangers. “AI might not distinguish a Ukrainian soldier from a Russian one,” warns Vadym, a defence engineer who requested anonymity. “Their uniforms can look identical.”
Vadym’s company, DevDroid, produces remotely controlled machine guns that use AI to detect and track people. Automatic firing is disabled to prevent friendly fire. “We could enable it,” he says, “but we need more field experience and feedback before trusting it fully.”
Ethical and legal questions remain. Can AI obey the laws of war? Will it recognise civilians or surrendering soldiers? Myronenko stresses humans must make the final decision, even if AI assists. Yet he warns not all militaries will act responsibly.
The global arms race intensifies
AI is driving a new kind of arms race. Traditional defences—jamming, missiles, or tanks—struggle against swarms of intelligent drones.
Ukraine’s “Spider Web” operation last June, when 100 drones struck Russian air bases, reportedly relied on AI coordination. Many fear Moscow could replicate the tactic, both on the front and deep inside Ukraine.
President Volodymyr Zelensky told the United Nations that AI is fuelling “the most destructive arms race in human history.” He called for urgent global rules on AI weapons, stressing the threat is “as urgent as preventing the spread of nuclear arms.”
