Connect with us

Instagram

AI technologies — like police facial recognition — discriminate against people of colour

Published

on

Experts suggest that Williams is not alone, and that others have been subjected to similar injustices. The ongoing controversy about police use of Clearview AI certainly underscores the privacy risks posed by facial recognition technology. But it’s important to realize that not all of us bear those risks equally. (Pixabay photo)

Detroit police wrongfully arrested Robert Julian-Borchak Williams in January 2020 for a shoplifting incident that had taken place two years earlier. Even though Williams had nothing to do with the incident, facial recognition technology used by Michigan State Police “matched” his face with a grainy image obtained from an in-store surveillance video showing another African American man taking US$3,800 worth of watches.

Two weeks later, the case was dismissed at the prosecution’s request. However, relying on the faulty match, police had already handcuffed and arrested Williams in front of his family, forced him to provide a mug shot, fingerprints and a sample of his DNA, interrogated him and imprisoned him overnight.

Experts suggest that Williams is not alone, and that others have been subjected to similar injustices. The ongoing controversy about police use of Clearview AI certainly underscores the privacy risks posed by facial recognition technology. But it’s important to realize that not all of us bear those risks equally.




Read more:
The coronavirus pandemic highlights the need for a surveillance debate beyond ‘privacy’


Training racist algorithms

Facial recognition technology that is trained on and tuned to Caucasian faces systematically misidentifies and mislabels racialized individuals: numerous studies report that facial recognition technology is “flawed and biased, with significantly higher error rates when used against people of colour.”

This undermines the individuality and humanity of racialized persons who are more likely to be misidentified as criminal. The technology — and the identification errors it makes — reflects and further entrenches long-standing social divisions that are deeply entangled with racism, sexism, homophobia, settler-colonialism and other intersecting oppressions.

A France24 investigation into racial bias in facial recognition technology.

How technology categorizes users

In his game-changing 1993 book, The Panoptic Sort, scholar Oscar Gandy warned that “complex technology [that] involves the collection, processing and sharing of information about individuals and groups that is generated through their daily lives … is used to coordinate and control their access to the goods and services that define life in the modern capitalist economy.” Law enforcement uses it to pluck suspects from the general public, and private organizations use it to determine whether we have access to things like banking and employment.

Gandy prophetically warned that, if left unchecked, this form of “cybernetic triage” would exponentially disadvantage members of equality-seeking communities — for example, groups that are racialized or socio-economically disadvantaged — both in terms of what would be allocated to them and how they might come to understand themselves.

Some 25 years later, we’re now living with the panoptic sort on steroids. And examples of its negative effects on equality-seeking communities abound, such as the false identification of Williams.

Pre-existing bias

This sorting using algorithms infiltrates the most fundamental aspects of everyday life, occasioning both direct and structural violence in its wake.

The direct violence experienced by Williams is immediately evident in the events surrounding his arrest and detention, and the individual harms he experienced are obvious and can be traced to the actions of police who chose to rely on the technology’s “match” to make an arrest. More insidious is the structural violence perpetrated through facial recognition technology and other digital technologies that rate, match, categorize and sort individuals in ways that magnify pre-existing discriminatory patterns.

Structural violence harms are less obvious and less direct, and cause injury to equality-seeking groups through systematic denial to power, resources and opportunity. Simultaneously, it increases direct risk and harm to individual members of those groups.

Predictive policing uses algorithmic processing of historical data to predict when and where new crimes are likely to occur, assigns police resources accordingly and embeds enhanced police surveillance into communities, usually in lower-income and racialized neighbourhoods. This increases the chances that any criminal activity — including less serious criminal activity that might otherwise prompt no police response — will be detected and punished, ultimately limiting the life chances of the people who live within that environment.

And the evidence of inequities in other sectors continues to mount. Hundreds of students in the United Kingdom protested on Aug. 16 against the disastrous results of Ofqual, a flawed algorithm the U.K. government used to determine which students would qualify for university. In 2019, Facebook’s microtargeting ad service helped dozens of public and private sector employers exclude people from receiving job ads on the basis of age and gender. Research conducted by ProPublica has documented race-based price discrimination for online products. And search engines regularly produce racist and sexist results.




Read more:
Google’s algorithms discriminate against women and people of colour


Perpetuating oppression

These outcomes matter because they perpetuate and deepen pre-existing inequalities based on characteristics like race, gender and age. They also matter because they deeply affect how we come to know ourselves and the world around us, sometimes by pre-selecting the information we receive in ways that reinforce stereotypical perceptions. Even technology companies themselves acknowledge the urgency of stopping algorithms from perpetuating discrimination.

To date the success of ad hoc investigations, conducted by the tech companies themselves, has been inconsistent. Occasionally, corporations involved in producing discriminatory systems withdraw them from the market, such as when Clearview AI announced it would no longer offer facial recognition technology in Canada. But often such decisions result from regulatory scrutiny or public outcry only after members of equality-seeking communities have already been harmed.

It’s time to give our regulatory institutions the tools they need to address the problem. Simple privacy protections that hinge on obtaining individual consent to enable data to be captured and repurposed by companies cannot be separated from the discriminatory outcomes of that use. This is especially true in an era when most of us (including technology companies themselves) cannot fully understand what algorithms do or why they produce specific results.

Privacy is a human right

Part of the solution entails breaking down the current regulatory silos that treat privacy and human rights as separate issues. Relying on a consent-based data protection model flies in the face of the basic principle that privacy and equality are both human rights that cannot be contracted away.

Even Canada’s Digital Charter — the federal government’s latest attempt to respond to the shortcomings of the current state of the digital environment — maintains these conceptual distinctions. It treats hate and extremism, control and consent, and strong democracy as separate categories.

To address algorithmic discrimination, we must recognize and frame both privacy and equality as human rights. And we must create an infrastructure that is equally attentive to and expert in both. Without such efforts, the glossy sheen of math and science will continue to camouflage AI’s discriminatory biases, and travesties such as that inflicted on Williams can be expected to multiply.The Conversation

Jane Bailey, Professor of Law and Co-Leader of The eQuality Project, L’Université d’Ottawa/University of Ottawa; Jacquelyn Burkell, Associate Vice-President, Research, Western University, and Valerie Steeves, Full Professor, L’Université d’Ottawa/University of Ottawa

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Latest

silhouette of voter ballot silhouette of voter ballot
News1 min ago

Vietnam names new president

ISTANBUL – Vietnam has appointed Minister for Public Security To Lam as the state president, the country’s Communist Party announced Saturday....

PBBM PBBM
Headline6 mins ago

Marcos seeks probe on China’s wiretapping of AFP official

MANILA – President Ferdinand R. Marcos Jr. has vowed to investigate the alleged wiretapping of the Chinese Embassy in Manila against...

News43 mins ago

Gov’t imposes electronic invoicing of imported goods

MANILA – President Ferdinand R. Marcos Jr. has ordered the implementation of a digital and integrated pre-border technical verification and cross-border...

AFFORDABLE RICE AFFORDABLE RICE
News1 hour ago

KADIWA Centers sell rice at P29 per kilo in Metro Manila

MANILA – As part of the solution to bring down retail rice prices in the country, the Department of Agriculture (DA)...

Headline8 hours ago

Marcos: China policy vs ‘trespassers’ in South China Sea unacceptable

MANILA – President Ferdinand R. Marcos Jr. on Saturday said China’s policy to detain alleged “trespassers” in the South China Sea,...

News9 hours ago

Marcos’ PFP forges alliance with Sotto’s NPC

MANILA – President Ferdinand Marcos Jr.’s Partido Federal ng Pilipinas (PFP) officially signed an alliance with the Nationalist People’s Coalition (NPC)...

test tube bloods test tube bloods
Health1 day ago

Infected blood scandal – what you need to know

The infected blood scandal has been hailed the worst treatment disaster in the history of the NHS. Over 3,000 people...

hands holding pregnancy test hands holding pregnancy test
Health1 day ago

Britain’s abortion laws are still in the Victorian era, and women are the collateral damage

A vote on ending prosecutions for abortion appears to have been delayed again. MPs have been expecting to vote on...

sleeping woman and electric fan sleeping woman and electric fan
Environment & Nature1 day ago

Extreme heatwaves in south and south-east Asia are a sign of things to come

Since April 2024, wide areas of south and south-east Asia, from Pakistan to the Philippines, have experienced prolonged extreme heat....

News1 day ago

Beijing is walking a fine line between support for Russia and not angering the west too much

Russia’s Vladimir Putin and China’s Xi Jinping have announced they will work together more closely to offset US pressure as...

WordPress Ads