“This technology is our future threat,” warns Serhiy Beskrestnov as he inspects a newly captured Russian drone. It is unlike anything he has seen. Controlled by artificial intelligence, it can locate and strike targets without human input.
Beskrestnov, a consultant to Ukraine’s defence forces, has analysed countless drones since the invasion began. But this one is different. It neither transmits nor receives signals, making it impossible to jam or detect.
Both Ukrainian and Russian forces now test AI-driven tools on the battlefield. They use it to identify enemy positions, analyse intelligence, and clear mines faster than ever before.
AI becomes Ukraine’s invisible force
For Ukraine’s military, artificial intelligence is essential. “Our army receives over 50,000 video streams from the front each month,” says Deputy Defence Minister Yuriy Myronenko. “AI analyses the footage, identifies threats, and marks them on a map.”
The technology accelerates decision-making, optimises resources, and helps protect lives. Its most profound effect appears in unmanned systems. Ukrainian troops now operate drones that lock onto targets and complete the final attack path autonomously.
These drones cannot be jammed and are extremely difficult to shoot down. Experts predict they will soon evolve into fully autonomous weapons capable of identifying and eliminating targets without human oversight.
Drones that act on their own
“All a soldier needs to do is press a button on a smartphone,” says Yaroslav Azhnyuk, CEO of Ukrainian tech company The Fourth Law. “The drone will find its target, drop explosives, assess the damage, and return to base. Piloting skills are unnecessary.”
Azhnyuk believes these drones could greatly strengthen Ukraine’s air defences against Russian long-range drones like the Shaheds. “A computer-guided system can outperform humans,” he says. “It reacts faster, sees better, and moves more precisely.”
Myronenko admits the technology is still in development but says Ukraine is close to finishing it. “We have partially integrated it into some devices,” he notes. Azhnyuk predicts thousands of autonomous drones could be deployed by the end of 2026.
Innovation and the risks of autonomy
Full automation carries serious risks. “AI might not distinguish a Ukrainian from a Russian soldier,” warns Vadym, a defence engineer who requested anonymity. “They often wear identical uniforms.”
Vadym’s company, DevDroid, produces remotely controlled machine guns that use AI to detect and track movement. Automatic firing is disabled to prevent friendly fire. “We could enable it,” he says, “but we need more experience and field feedback to ensure it is safe.”
Legal and ethical concerns remain. Can AI follow the laws of war? Will it recognise civilians or surrendering soldiers? Myronenko stresses that humans must make the final call, even if AI assists. Yet he acknowledges that not every military will act responsibly.
The new global arms race
AI is driving a dangerous new arms race. Traditional defences like jamming, tanks, or missiles struggle against intelligent drone swarms.
Ukraine’s “Spider Web” operation last June, when 100 drones hit Russian air bases, reportedly relied on AI coordination. Many fear Moscow could replicate the tactic, both on the front lines and deeper inside Ukraine.
President Volodymyr Zelensky told the United Nations that AI is fuelling “the most destructive arms race in human history.” He urged world leaders to establish global rules for AI in weapons, calling the issue “as urgent as preventing nuclear proliferation.”
