In this course, you will be introduced to a number of data science topics from coding with Python, statistics, databases, natural language processing, and more. Data science is a relatively new field and every day data scientists are challenged with ethical questions for which there are little regulations and precedent to dictate their course of action.
When you are choosing a topic for a machine learning project, you have a lot of options, particularly when you are a student. As information technology professionals, it is crucial that you use good science to ensure that the insights you report to your stakeholders are based on sound evidence and not pseudo-science.
This lesson details some kinds of targets you should not try to build a project around predicting, in order to avoid performing "AI phrenology".
- Define phrenology
- Describe examples of AI projects that resemble phrenology
- List the types of machine learning targets to avoid
Phrenology is a type of pseudoscience first introduced in the the late 1700s. Practitioners of phrenology believed that the shapes of bumps on the human skull could tell you information about the person's personality, character, and psychology. Phrenology was used to uphold racist and sexist beliefs that were common at the time, and also to determine what sort of punishment or rehabilitation a person accused of a crime might receive.
Phrenology was already discredited by the end of the late 1800s and is no longer widely in use. However it continues to be a useful point of comparison for modern-day pseudoscience. In particular, people tend to make comparisons to phrenology any time someone tries to predict or analyze some attribute using data that is unrelated to that attribute, the way that bumps on a skull are unrelated to the attributes that phrenologists claimed to be measuring.
Image data (e.g. photos of faces) and text data (e.g. names) tend to be the most susceptible to modern phrenology attempts. This kind of data tends to be widely available, and both businesses and governments might be interested in inferring information about people based on these pieces of data.
Trying to predict personality traits based on a face or a name is the AI phrenology that most closely resembles historical phrenology. There is no evidence whatsoever that appearance can predict these things, but that has not stopped some people from trying to predict trustworthiness, criminality, and jobworthiness.
If you're interested in reading more about the issues with predicting these attributes, check out the links above. The main problem here is not that personality traits are impossible to predict, but that the data being used here is inappropriate for the intended target.
Regardless of the type of input data being used, predicting sensitive features can be an ethical minefield and should only be done in very rare circumstances, if at all. Attempting to predict things like sexual orientation, ethnicity, and disability status can bring real-world harm to people regardless of the accuracy of the models being used.
One example of when it might be appropriate to try to predict a sensitive feature is if you are specifically trying to measure disparities or discrimination. For example, researchers studying disparities in citation counts based on race and gender needed to predict the race and gender of most authors in their database in order to determine if they were being discriminated against.
If you decide to build a model that predicts a sensitive feature because you believe that you have a strong enough reason, try to make sure you consult with the people who might be marginalized on the basis of that feature in order to avoid any unintended harms.
The one other type of feature to watch out for is a feature where the "ground truth" is (nearly) impossible to determine. A classic example of this is emotion detection.
There are two main ways that a set of photographs can be turned into an emotion detection database:
- The people being photographed are asked to make a facial expression that represents a particular emotion
- Data labelers are asked to look at photos and infer emotions from facial expressions
In both cases, what is being measured is a proxy for the person's emotion, not their actual emotion. You can read more about the ethical challenges of different proxies for emotion here.
- Phrenology's History and Influence (web article)
- Citation inequities in the social sciences: The case of Communication studies - Dr Deen Freelon (YouTube video)
- Physiognomic Artificial Intelligence (research paper calling for legislative action)
Out of all the options for building a machine learning model, make sure you avoid dabbling in AI phrenology. This means not predicting personality traits based on images or names, not predicting sensitive features unless there is a very good reason, and not claiming you can predict an attribute while actually predicting a proxy for it.