All people are created equal, but in the eyes of the algorithm, not all faces are just yet.
A new study from MIT and Microsoft researchers (pdf) is providing more evidence of exactly how bad facial-recognition software is at accurately identifying darker faces, especially when those faces belong to women. In a test to identify the sex of people from their faces, software was able to do so with more than 99% accuracy for light-skinned men. For darker-skinned women, the software could be wrong as frequently as one-third of the time.
The results shed more light on a known problem—how limited data sets can impact the effectiveness of artificial intelligence, which might in turn heighten bias against individuals as AI becomes more widespread.