as race, gender, economic background, demography,
geography etc. Furthermore, ‘Even if the perfect
notion of fairness is found, how should it be
enforced?’
An obvious question that comes to mind is - why
do standard machine learning techniques, when
deployed directly, lead to outcomes which are unfair?
While there are multiple explanations for the same,
there are some which are widely known.
Chouldechova et.al discuss several causes of
unfairness in their work (Chouldechova and Roth,
2020). Firstly, bias could be encoded in the data.
Consider an AI model designed to diagnose skin
diseases from medical images, such as photographs
of rashes or lesions. The model is trained on a dataset
containing images of patients from various sources,
including hospitals and clinics. In this scenario, an
example of bias being encoded in the data could be
the overrepresentation of lighter skin tones in the
training data, leading to a situation where the AI
model's predictions are unfair and less accurate for
individuals with darker skin tones.
Secondly, different groups can have significantly
different distributions. Next, it is possible that
features are less predictive on some groups as
opposed to other groups. Consider an AI model
designed to predict heart disease risk in patients based
on various health indicators. The model is trained on
a diverse dataset that includes individuals from
different demographic groups, including both men
and women. Imagine a scenario in which a specific
health indicator ‘X’ is more strongly correlated with
heart disease risk in men compared to women. Due to
the stronger correlation between health indicator ‘X’
and heart disease in men, the model might assign a
higher risk score to a woman with elevated levels of
‘X’, even if other risk factors for heart disease are less
significant in women. Therefore, the unequal
predictive strength of certain features for different
groups—stronger for men compared to women—has
led to a situation where the AI model's predictions are
less accurate and fair for female patients.
Lastly, it is possible that some groups are
inherently less predictable. Consider an AI model
trained on a diverse dataset, designed to assist in
diagnosing mental health disorders. It is commonly
known that individuals from culturally distinct groups
may express symptoms of mental health disorders in
ways that are not well-captured by standardized
assessments. Cultural norms, beliefs, and
communication styles can significantly influence
how symptoms manifest and are reported. Thus, the
inherent unpredictability of symptom expression
among culturally distinct groups can lead to a
situation where the AI model's predictions are less
reliable and equitable.
Some of these scenarios can have relatively
‘simple’ solutions such as collecting more
representative data or including features which are
more predictive on all groups, which in itself is a
challenging and expensive process. While some
algorithms such as bolt-on postprocessing methods
(introducing randomization to ensure fairness) have
been proposed by researchers, other scenarios are more
complex to solve and are still open areas of research.
Since it is evident that there is a need to augment
standard principles of AI algorithm deployment to
account for algorithmic fairness, we revisit the
process of defining ‘fairness’. Kearns et.al state that,
based on the vast majority of work done on fairness
in machine learning, various definitions of fairness
can be divided into two broad categories: statistical
definitions and individual definitions (Kearns, Neel,
Roth and Wu, 2019). Statistical definitions focus on
fixing a small number of protected groups (such as
race) and defining fairness as the equality of a
statistical measure across all the subgroups (Asian,
Hispanics, African-American - not an exhaustive list)
in the identified group. An example of such a
statistical measure in healthcare could be the False
Positive Rate of a medical diagnosis. Fairness, in this
case, would mean that the probability of
mispredicting the presence of a disease should be
approximately equal across all subgroups.
Individual definitions of fairness focus on
satisfying each person’s perspective of fairness.
Algorithmically, this can be viewed as a constraint
satisfaction problem in which each person’s
perspective of fairness is a constraint which must be
satisfied while we improve the performance of our AI
model. Satisfying individual definitions of fairness is
an open research question because it does not scale
well. This means that the feasibility of simultaneously
solving each of these fairness constraint satisfaction
problems reduces as the number of individuals
involved increases.
Till date, more research has been done on the
statistical definitions of fairness due to its
comparatively lower complexity and simpler
validation. The first step is to identify which groups
or attributes we wish to ‘protect’ when we deploy our
algorithm. By protect, we mean that we want to
identify which are the vulnerable or minority groups
in our dataset. The next step focuses on defining what
constitutes ‘harm’ in a system. As an example, in case
of medical diagnosis, harm with respect to fairness
could be a higher misprediction of the absence of a
disease in a certain group in a population (referred to