Advancing UK Aerospace, Defence, Security & Space Solutions Worldwide

Features

Boosting trust in AI

Dr Nira Chamberlain, Professional Head of Discipline for Data Science, Atkins, asks how can we be sure artificial intelligence (AI) can be relied upon to drive automated military vehicles?

Above: Dr Nira Chamberlain.
Courtesy Atkins

The defence sector has been using automated vehicles for decades and a new generation of automated vehicles, powered by AI, is on the increase as we put more trust in AI algorithms. Yet can we expect to have full confidence in these vehicles when we do not fully understand in any detail how AI works?

Advertisement
ODU RT

AI-supported automated vehicles will soon be operating widely in the defence space. You only need to look at R&D, such as AIMM, the US Army’s flagship AI research programme for manoeuvrability, or the UK Government’s assertion that deploying AI will provide an advantage over adversaries in military planning.

However the fact remains that it is quite hard for most of us humans to understand the actual rationale behind predictive AI decision-making. How does it work behind the scenes? And what happens if these robots make the wrong decision? Won’t that have an impact from a legal, ethical and safety critical point of view?

Surely, we need to start interrogating their responses and not settle for the computer saying 'yes' or 'no' – and us not really understanding why.

When it comes to automated vehicles, many of these are in the prototype phase and can be prone to making unusual - and erratic - decisions. To understand why, we must look into what is driving this decision-making.

While any good mathematician is expected to show their calculations, when it comes to AI – which is powered by systems generating predictions based on billions of calculations every minute – it is not so straightforward. One response can be to develop a reverse engineering algorithm, which, when wrapped around the black box that drives the automated vehicle, can approximate the logic and rationale for every prediction made.

In a defence context, this will prove to be a highly valuable asset. By reverse engineering the AI process, we can reveal with more granularity how decisions are arrived at, such as why automated vehicles are instructed by the algorithm to turn left, right, stop, start, or go straight ahead.

Reverse engineering can help us understand how tasks are accomplished through deductive reasoning. Take a washing machine, as an example: we know you put dirty laundry in it at point A, and at point B clean laundry comes out. To understand better how that process works, you can work backwards, stripping away at each stage of the process, until you are right back to the blueprint. Once there, you have revealed in reverse the chain of events that, going forward, result in clean laundry.

Advertisement
ODU RT

So, similarly, with AI algorithms we can reverse engineer to provide a picture of the logic of each decision made in that process. Also, while we will never have a totally full picture – using a 1,000-piece jigsaw as a comparative example, reverse engineering algorithms give us the means to see where around 750 of the pieces go – it does mean we can pinpoint a higher degree of certainty, around 75%, than ever before.

In effect, the black box driving the vehicle is a mathematical model: a simplified representation of a complex system. So effectively, what we have done is produce a mathematical model of a mathematical model. This allows us more trust in deployment of AI-driven automated vehicles but more confidence in our investment decisions as we can be assured that AI is safe and reliable in this context.

It is not enough for major manufacturers to claim that AI algorithms cannot be challenged. They can. Also, by gaining a better understanding of how those AI-driven black boxes reach the decisions they do, we can put protocols and processes in place to mitigate against any perceived risk. This will be a gamechanger to us harnessing the power of AI where any critical decision needs to be made; which, within the defence context, is a matter of course.

Dr Nira Chamberlain, PhD HonDSc, is the Professional Head of Discipline for Data Science for Atkins, a member of the SNC-Lavalin Group. He is the current President of the Institute of Mathematics and its Application (IMA) and is a Visiting Fellow of Loughborough University Mathematical Sciences Department. In 2019 the Inclusive Tech Alliance named Nira as one of the Top 100 Most Influential Black, Asian and Minority Ethnic leaders in the UK’s Tech. Nira has two mathematical doctorates and has been listed by the PowerList Top 100 Most Influential people of African or African Caribbean heritage in UK for four years running (2018-2021).

Advertisement
Babcock LB
Building steps towards CNI resilience

Features

Building steps towards CNI resilience

10 January 2025

Peter Lenk, Technical Lead at Goldilock, sets out the key steps for building a timeline of resilience for the protection of Critical National Infrastructure (CNI).

Cybersecurity trends in 2025

Features

Cybersecurity trends in 2025

7 January 2025

Usman Choudhary, Chief Product & Technology Officer, VIPRE Security Group, shares his thoughts on security trends that will likely dominate in 2025.

Public and private sector collaboration shaping modern defence

Features

Public and private sector collaboration shaping modern defence

9 December 2024

Tristan Wood, founder of Livewire Digital, explains how the UK is leveraging its advanced manufacturing sector and partnerships with industry, government and international allies, to transform military capabilities, with COTS solutions in the vanguard.

Collaboration key to managing supply shocks

Features

Collaboration key to managing supply shocks

20 November 2024

Neil McManus, a Partner and aerospace and defence specialist at Vendigital, encourages manufacturers to collaborate closely with suppliers to mitigate the impact of shocks and supply chain disruptions.

Advertisement
Cranfield
Harnessing AI for aerospace cybersecurity

Features

Harnessing AI for aerospace cybersecurity

28 October 2024

Graham Younger, VP of Aerospace at Expleo, looks at how the use of AI technologies can enhance cybersecurity across the aerospace sector.

Winning with data on the battlefield

Features

Winning with data on the battlefield

14 October 2024

As real-time data streaming on kit, soldiers, assets and even from space transform the battlefield, Rob Mather, VP, aerospace and defence industries, IFS, examines why interconnected technologies and data streams are now critical to mission success.

Advertisement
Cranfield