Menu Close

Articles on Predictive policing

Displaying all articles

Facial recognition algorithms are usually tested using white faces, which results in the technology being unable to differentiate between racialized individuals. (Shutterstock)

AI technologies — like police facial recognition — discriminate against people of colour

Technology is not neutral, as facial recognition algorithms and predictive policing have shown us. Algorithms discriminate by design, reflecting and reinforcing pre-existing biases.
A vigil in memory of Alton Sterling, who was shot dead by police. Baton Rouge, Louisiana. REUTERS/Jeffrey Dubinsky

Why is it so hard to improve American policing?

For 50 years, we have worked to make U.S. police more diverse and less intrusive. Why haven't we made more progress?

Top contributors