UK can set the standard in ethical Artificial Intelligence

Going for gold: Study by The Chartered Institute for IT (BCS) finds the UK can set the “gold standard” in ethical AI

The UK is home to companies including DeepMind, Graphcore, Oxbotica, Darktrace, BenevolentAI and others and is Europe’s leader in AI. However, the country is unable to match the funding and support available to similar companies based in countries like the US and China.

Because of this, many experts have suggested that the UK should tap its strengths in leading universities and institutions, diplomacy, and democratic values, in order to become a world leader in creating AI that ‘cares about humanity’.

The importance of Artificial Intelligence on our lives going forward

Dr Bill Mitchell OBE, Director of Policy at BCS, The Chartered Institute for IT and a lead author of the report, said creating the gold standard  would be a critical part of the UK’s economic recovery. He added that everyone deserved to have confidence in AI as it will affect our lives during the coming years and the technology should reflect the needs of everyone the developers are engineering the software for. He gave examples of credit scoring apps, cancer diagnosis and software which could decide if you get a job or not.

It is commonly feared that current biases in many AI systems could lead to increasing existing societal problems, including the wealth gap and discrimination based on race, sex, age, and more. It is thought access to digital skills and training across the board could combat these tendencies.

Some high profile AI mismodelling

Public trust in AI has been damaged through high-profile missteps including the crisis last summer when an algorithm was used to estimate the grades of students. A follow-up survey from YouGov – commissioned by BCS – found that 53 percent of UK adults had no faith in any organisation to make judgements about them.

In May last year, the national press reported that code written by Professor Neil Ferguson and his team at Imperial College London that informed the decision to enter a lockdown was “totally unreliable” and also damaged public trust in software. 

The report also found a large disparity in the competence and ethical practices of organisations using AI. And in the UK government’s National Data Strategy, it states: “Used badly, data could harm people or communities, or have its overwhelming benefits overshadowed by public mistrust.


Featured Articles

Securing the future of IoT with AI biometric technology

The world needs an IoT future, so It’s time to forget your password – for good. A new breed of AI-powered biometric security measures is required

EU networks plan to build a foundation for trustworthy AI

Artificial intelligence technologies are in their infancy, but commercial stakeholders must build them on foundations of trust, say research experts

ICYMI: Power users distrust AI and new National Robotarium

A week is a long time in artificial intelligence, so here’s a round-up of AI Magazine articles that have been starting conversations around the world

Reducing the impact of ecommerce with AI and ML

AI Applications

Now is the time for intelligent products and services

AI Strategy

The eyes of a machine: automation with vision

Machine Learning