UK Centre Launches Probe on Ethics of AI Algorithms in 'Pre-crime', Recruitment

The UK Centre for Data Ethics and Innovation (CDEI) will launch an investigation on the possibility of bias in algorithms used to tackle crime. Whilst algorithms have massive potential for preventing crime, they also pose risk from human influences in their recommendations.
Sputnik

Algorithms can be used to "assess the likelihood of re-offending and inform decisions about policing, probation and parole", the UK government in a Wednesday press release, adding that the CDEI aimed to ensure that the technologies will be used ethically.

Programmes such as the Harm Assessment Risk Tool (HART) in Durham are already in place, which helps police to decide if persons on parole are likely to commit future offences.

The CDEI supports the government's 2-Year Industry Strategy and was set up to ensure that data-driven technologies and AI were "used for the benefit of society", including partnering with the Race Disparity Unit to probe potential racial bias in decision-making in the criminal and justice systems. 

Facebook Set to Combat ‘Revenge Porn’ Through New AI Technology

Digital secretary Jeremy Wright said at a Downing Street event: "Technology is a force for good which has improved people's lives but we must make sure it is developed in a safe and secure way. 

The Centre for Data Ethics and Innovation was set up to help achieve that aim and "keep Britain at the forefront of technological development," Mr Wright added.  

— Ed Vaizey (@edvaizey) March 20, 2019

Roger Taylor, Chair of the Centre for Data Ethics and Innovation, said: "The Centre is focused on addressing the greatest challenges and opportunities posed by data driven technology. These are complex issues and we will need to take advantage of the expertise that exists across the UK and beyond. If we get this right, the UK can be the global leader in responsible innovation. 

READ MORE: ‘No One Is Learning': Hacker Offers New Cache of Data From 26 Million People

Other applications for such technologies include screening CVs and shortlisting candidates in job recruitment processes to avoid racial or gender bias, as well as boosting innovation in the digital economy. Financial services can also benefit from AI machine-learning by deciding who can receive loans, but such technologies have increased scrutiny on transparency and fairness of outcomes, the CDEI said. 

Science & TechNewsfeedSocietyUnited Kingdom (UK)Centre for Data Ethics and Innovation (CDEI)reportartificial intelligenceHarm Assessment Risk Toolethicsbiasalgorithmsinvestigation reportmachine learningalgorithm change
Discuss