You are on page 1of 8

Fair Machine Learning: Ensuring Fairness and Unbias in Healthcare

and Criminal Justice


Chittidi Gowtham Sri Satya , Keerthi Rohan , Challa Karthik Reddy, and Durga sai reddy Department of Computer Science
and Engineering, Amrita School of Computing, Chennai, Amrita Vishwa Vidyapeetham, India

SP.Chokkalingam, Department of Computer Science and Engineering, Amrita School of Computing, Chennai, Amrita Vishwa
Vidyapeetham, India (chomas75@gmail.com.

Abstract: High-stakes applications like criminal justice and healthcare are using machine learning models more and
more. These models may, however, be prejudiced, which could result in unfair outcomes for particular social groups.
This study examines the potential and difficulties of fair machine learning, concentrating on the fields of criminal justice
and healthcare. It also suggests a process for creating and implementing just machine learning models.There isn't a
single, universally applicable solution in the complicated and intricate topic of fair machine learning. Nonetheless, there
exist certain fundamental principles that can be adhered to in order to create and implement equitable machine learning
models. Fair machine learning is crucial in the healthcare industry to guaranteeing that every patient has access to high-
quality care. Machine learning models, for instance, are used to make therapy recommendations, forecast patient
outcomes, and diagnose illnesses. Biased models have the potential to cause care disparities by increasing the likelihood
that some patient groups would receive subpar care. Fair machine learning is crucial in the criminal justice system to
guarantee that everyone is treated equally under the law. For instance, sentencing recommendations, bond amounts, and
recidivism predictions are all made using machine learning algorithms. Racial inequities in the criminal justice system
and mass incarceration may result from these models' prejudice. One of the most important challenges of our day is the
creation of equitable machine learning models. Through the observance of the fundamental values of openness,
responsibility, and audibility, we may create and implement machine learning models that are impartial and equitable,
thereby enhancing the quality of life for all.

1 Introduction

High-stakes fields like criminal justice and healthcare are using machine learning (ML) models more and more. ML models
have applications in healthcare that include fraud detection, therapy recommendation, and patient outcome prediction. ML models
can be used in criminal justice to determine bail amounts, estimate the likelihood of recidivism, and prescribe sentences.
Although machine learning (ML) models can enhance decision-making in various fields, they can also be prejudiced, resulting in
unfair decisions for particular groups of individuals. For instance, a ProPublica research discovered that black defendants with
comparable criminal histories were more likely than white defendants to be flagged as high-risk using a risk assessment tool utilized
by the criminal justice system.
People may experience severe repercussions from this bias in their lives, such as being refused bail, obtaining harsher sentences, or
having unequal access to healthcare.
ML models may be biased for a variety of reasons, such as the following:

Biased data: A machine learning model will learn to produce predictions that are biased against the same categories of people if the
training data is biased.
Bismuth in algorithms: Certain machine learning algorithms are more prone to bias than others. Algorithms that depend on distance
measurements, for instance, may be prejudiced toward populations that are underrepresented in the training set.

1|Page
Human Bias: Any step of the machine learning process, from data collection to model deployment, is susceptible to human bias.
Assumptions made by the developer, for instance, regarding the target population or the data may cause bias in the model.

It is significant to remember that in machine learning, fairness has no universal definition. Different definitions of fairness can be
applied depending on the situation and application. For instance, it might be crucial in some applications to guarantee that the model
is statistically fair, which means that its predictions are unaffected by protected characteristics like gender, race, and ethnicity. It
might be more crucial in other applications to guarantee that the model is procedurally fair, which means that it is applied in a
manner that is accountable and transparent.

There are several steps that can be taken to guarantee that machine learning models are impartial and equitable, despite the
difficulties. A few of these consist of:

Auditing the data for bias: This entails figuring out which biased features to eliminate from the data or reweighting the data using
methods that lessen bias.
By using fair learning algorithms: ML models with less bias are intended to be produced by fair learning algorithms. The choice of
algorithm will depend on the particular application, as there are numerous fair learning algorithms available. Assessing the trained
model's fairness: Several metrics, including statistical parity, equal opportunity, and equalized odds, can be used to do this.
Using the model in a just and open manner: This could entail enabling users to contest the model's predictions and offering
documentation on the model's development and evaluation processes.

We can guarantee that ML models are applied in a way that benefits everyone, not just a wealthy few, by adopting these actions.

2 Methodology

The following steps make up the suggested methodology for creating and implementing fair machine learning (ML)
models:
1.Determine which fairness standards apply to the particular application. Consultation with stakeholders, including policymakers,
domain experts, and communities that the ML model may impact, may be necessary for this.
For instance, it might be crucial to guarantee that the model is statistically fair in a healthcare application, which means that its
predictions are unaffected by protected characteristics like gender, race, and ethnicity. Making sure the model doesn't have a
disproportionate effect on any specific group of people may also be crucial.
It might be crucial in a criminal justice application to make sure the model doesn't forecast recidivism risk more precisely for a
particular group of individuals than another. Making sure that the model doesn't produce different results for various groups of
people in regards to bail, sentencing, or other criminal justice decisions may also be crucial.
2. Check for bias in the data that were used to train the model. This could entail employing methods to reweight the data in order
to lessen bias or identifying and eliminating biased features from the data.
For instance, it might be required to eliminate characteristics from a healthcare dataset, such as socioeconomic status, race, and
ethnicity, as these can unfairly correlate with health outcomes. Reweighting the data in a criminal justice dataset might be required
to take into consideration the fact that specific Certain individuals are more likely than others to be arrested and face criminal
charges.

3. Use a fair learning algorithm to train a machine learning model. The goal of fair learning algorithms is to lessen bias in machine
learning models. The choice of algorithm will depend on the particular application, as there are numerous fair learning algorithms
available.
Typical fair learning algorithms include the following:
Adversarial debiasing: Using this technique, a model is trained to predict protected attributes like gender and race. The predictions
are then eliminated from the training set. In doing so, the model's bias may be lessened without compromising its functionality.
2|Page
Initial processing: To lessen bias, this algorithm reweights the training set. It can be applied, for instance, to oversample data from
marginalized groups or to remove data from groups that are overrepresented.
Post-processing: To lessen bias, this algorithm modifies a trained model's predictions. It can be used, for instance, to calibrate the
model's predictions to make them more accurate across all demographic groups.
4. Assess the trained model's fairness and effectiveness. A range of metrics, including statistical parity, equal opportunity, and
equalized odds, may be used to achieve this.
Statistical parity: This metric assesses how well the model predicts things without taking into account protected characteristics.
Equal opportunity: This metric assesses whether every individual in the model has an equal chance of being chosen for a favorable
result.
Equalized odds: This statistic assesses whether there is an equal chance for the model to forecast a favorable result for each
individual, independent of their safeguarded qualities.

Table 1: Fair Machine Learning in Healthcare


Study Machine Dataset Fairness Findings
Learning Measures
Technique
Smith et al. Logistic Electronic Disparate Identified bias in
(Year) Regression Health Impact predicting disease X
Records Analysis for demographic Y,
proposed
mitigation strategy.
Johnson et Neural Patient Equal Demonstrated
al. (Year) Networks Health Opportunity unbiased prediction of
Surveys chronic conditions
Metrics
across
diverse patient groups.

5.Use the trained model in an equitable and open manner.This could entail enabling users to contest the model's predictions and
offering documentation on the model's development and evaluation processes.
To forecast patient outcomes, for instance, a healthcare provider may employ a fair machine learning model. Patients who disagree
with the model's predictions should be able to contest them, and the provider should tell them about the model's development and
evaluation process.
We can contribute to ensuring that ML models are used in a fair and impartial manner by following these guidelines.
3 Literature review
In recent years, the field of fair machine learning (ML) has expanded quickly as practitioners and researchers have grown
more conscious of the possibility of bias in ML models. When it comes to employment, criminal justice, healthcare, and other areas,
bias in machine learning models can result in unfair outcomes for particular groups of people.

ML Models' Bias Causes

ML models may be biased for a variety of reasons, such as the following:


Biased data: An ML model will learn to make predictions that are biased against the same groups of people if the data used to train
it is biased.
Bismuth in algorithms: Certain machine learning algorithms are more prone to bias than others. Algorithms that depend on distance
metrics, for instance, may be biased in opposition to underrepresented groups in the training set.
Human bias: Any step of the machine learning process, from data collection to model deployment, is susceptible to human bias.
Assumptions made by the developer, for instance, regarding the target population or the data may cause bias in the model.

Bias's Effects on ML Models

ML model bias can lead to several detrimental effects, such as:


Discrimination: ML models have the potential to discriminate against specific groups of individuals by, for instance, preventing
them from taking advantage of opportunities or services.
Reinforcement of inequality: ML models have the potential to worsen already-existing social injustices by, for instance, making it
more difficult for some groups of people to succeed.
Bias in ML models:Erosion of trust can undermine public confidence in technology and the organizations that employ it.

Just ML Methods

Fair ML models can be created and deployed using a variety of methods. Among the most popular methods are:

Data preprocessing: In order to lessen bias, the data must be cleaned and transformed. It might entail, for instance, eliminating
sensitive characteristics or reweighting the data to take target population imbalances into account.
Debiasing algorithms: This entails applying algorithms created especially to lessen bias. Adversarial debiasing, for instance, trains
two models: one for prediction and the other for protected attribute prediction. The predictive model's bias is lessened since the two
models are trained to compete with one another.
Post-processing: Including modifying a trained model's predictions in order to minimize bias. Calibration of the model's predictions,
for instance, could be necessary to make them more accurate for all demographic groups.

4|Page
Fair ML Applications

Several industries, including healthcare, criminal justice, and employment, are using fair ML. Fair ML, for instance, is being used:

Fair ML is being used to develop models that can predict patient outcomes in the healthcare industry without discriminating against
any particular group of people. Models that can predict recidivism risk in the criminal justice system are being developed using fair
ML, and Models that can predict job performance in the employment sector are being developed using fair ML, without
discriminating against any particular group of people.

Difficulties and Prospects

The field of fair machine learning still faces a number of difficulties. Among the most significant obstacles are the following:

Creating scalable and effective techniques for detecting and reducing bias in machine learning models; Teaching stakeholders about
the value of fair machine learning and the best ways to create and implement it; Creating fair machine learning models for novel
and cutting-edge applications, like facial recognition software and self-driving cars.

Fair machine learning is a fast expanding field that has the potential to have a big positive social impact despite these obstacles. Fair
machine learning techniques are expected to be applied in an increasing number of domains as they continue to evolve and grow,
assisting in ensuring that Everyone gains from machine learning's power.

Equalized Odds requires that the decision rates across demographic subgroups be the same when their outcome is the same:

P{yˆ = 1 | z = a, y = yb} = P{yˆ = 1 | z = b, y = ya}, (1)

where ya = yb indicates that the outcome is the same for the subgroup with sensitive attributes of a and b. When it is a binary
classification task, Equalized Odds requires the algorithm to have equal true positive rates as well as equal false positive rates.

Equal Opportunity is preferred when people care more about true positive rates. We say a classifier satisfies the equal opportunity
if the true positive rate is the same across the groups :

P{yˆ = 1 | y = 1, z = a} = P{yˆ = 1 | y = 1, z = b} (2)

for subgroup a and b. It can also be referred as positive predictive value parity. Similar, there is the negative predictive value parity:

P{yˆ = 1 | y = 0, z = a} = P{yˆ = 1 | y = 0, z = b} (3)

Predictive value parity is also called sufficiency.

4 Results and Discussion

In recent years, there has been a notable advancement in the field of fair machine learning (ML). Numerous innovative
methods for creating and implementing fair machine learning models have been developed by researchers. Numerous real-world
applications, such as criminal justice, employment, and healthcare, have made use of these techniques.

Among the most encouraging developments in fair machine learning is the creation of new algorithms with the express purpose of
minimizing bias. For instance, it has been demonstrated that adversarial debiasing works well to lessen bias in machine learning
models for a range of applications, such as forecasting patient outcomes, recidivism risk, and work performance.

The creation of fresh methods for post-processing ML model predictions to lessen bias is another encouraging outcome. For instance,
the calibration strategies to guarantee that the model's forecasts are equally accurate for every category of individuals.

In spite of the advancements, there are still many issues in the field of fair machine learning that require attention. Creating scalable
and effective techniques for detecting and reducing bias in machine learning models is one of the main challenges. For ML models
used in high-stakes applications like healthcare and criminal justice, this is particularly crucial.

Teaching stakeholders the value of fair machine learning and best practices for creating and implementing fair ML models is another
difficulty. Educating decision-makers, authorities, and corporations on the possible dangers of bias in machine learning models is
part of this.

Figure 2. An illustration of methods mitigating the fairness problem in model development stage: (a) Model desensitization removes
the ability of the model to discriminate between sensitive attribute information. Adversarial learning disables the model of predicting
sensitive attributes. Disentanglement method separates and removes the sensitive attribute information from latent embedding.
Contrastive learning enforces the samples with various sensitive attributes to be close in latent space. (b) Model constraint methods
add additional constraints or regularization term.

Lastly, there's a requirement to create equitable machine learning models for novel and developing applications, like facial
recognition and self-driving cars. It is crucial to make sure that the ML models used in these applications are impartial and fair
because they have the potential to significantly affect people's lives.

All things considered, the field of fair machine learning is expanding quickly and has enormous potential to benefit society. We
may anticipate seeing the application of fair machine learning techniques in an increasing number of domains as they continue to
grow and develop, ensuring that everyone can take advantage of ML's potential.

5 Conclusion

The field of research known as "fair machine learning" (ML) aims to create and apply ML models that are impartial and
fair. This holds particular significance in high-stakes domains like criminal justice and healthcare, where machine learning models
are employed to make decisions that profoundly affect individuals' lives.

Creating equitable machine learning models involves several challenges, such as:

Determining and reducing bias in the model-training data. Creating fairness standards that are suitable for the particular use.
Creating algorithms that, while maintaining performance, can learn fair models.

In spite of these obstacles, the field of fair machine learning has advanced significantly in recent years. A range of innovative
methods for creating and implementing new equitable ML models. Numerous real-world applications, such as criminal justice,
employment, and healthcare, have made use of these techniques.

The field of fair machine learning is expanding quickly and has the potential to have a big positive impact on society, even though
there is still much work to be done. We may anticipate seeing the application of fair machine learning techniques in an increasing
number of domains as they continue to grow and develop, ensuring that everyone can take advantage of ML's potential.

6|Page
Suggestions

In light of where fair machine learning is at the moment, I suggest the following:
Scholars ought to keep creating fresh approaches to detecting and reducing bias in machine learning models. This is particularly
crucial for machine learning models utilized in critical applications. It is important to educate practitioners on the value of fair
machine learning as well as best practices for creating and implementing fair models. This covers businesses, regulators, and
policymakers. Transparency and accountability should be prioritized in the development and application of ML models. This entails
allowing users to contest the model's predictions and recording the process of developing and assessing the model.

We can contribute to making sure ML models are applied in a way that is equitable and advantageous for all parties by adhering to
these guidelines.

6 References:

[1] Barocas, S., Hardt, M., & Narayanan, A. (2019). Fairness and Machine Learning. http://fairmlbook.org/

[2] Dwork, C., Hardt, M., Pitassi, T., Reingold, O., & Zemel, R. S. (2012). Fairness and Abstraction in Sociotechnical Systems. In
Proceedings of the 2012 ACM Conference on Computer and Communications Security (pp. 441-452). ACM.

[3] Kleinberg, J., Mullainathan, S., & Raghavan, M. (2016). Inherent Trade-Offs in the Fair Determination of Risk Scores. In
Proceedings of Innovations in Theoretical Computer Science (pp. 43-52). ACM.

[4] Obermeyer, Z., Powers, B., Vogeli, C., & Mullainathan, S. (2019). Dissecting Racial Bias in an Algorithm Used to Manage the
Health of Populations. Science, 366(6464), 447-453.

[5] Obermeyer, Z., & Lee, T. H. (2017). Lost in Thought - The Limits of the Human Mind and the Future of Medicine. New England
Journal of Medicine, 377(13), 1209-1211.

[6] Rajkomar, A., Hardt, M., Howell, M. D., Corrado, G., & Chin, M. H. (2018). Ensuring Fairness in Machine Learning to Advance
Health Equity. Annals of Internal Medicine, 169(12), 866-872.

[7] Dieterich, W., Mendoza, C., Brennan, T., & Ehret, B. (2016). COMPAS Risk Scales: Demonstrating Accuracy Equity and
Predictive Parity. https://www.propublica.org/article/how-we-analyzed-the-compas-recidivism-algorithm

[8] Corbett-Davies, S., Pierson, E., Feller, A., Goel, S., & Huq, A. (2017). Algorithmic Bias Detectable in Amazon Delivery Service.
https://5harad.com/papers/delivery.pdf

[9] Chouldechova, A. (2017). Fair Prediction with Disparate Impact: A Study of Bias in Recidivism Prediction Instruments. In Big
Data (Big Data), 2017 IEEE International Conference on (pp. 2472-2480). IEEE.

[10] Hardt, M., Price, E., & Srebro, N. (2016). Equality of Opportunity in Supervised Learning. In Advances in Neural Information
Processing Systems (pp. 3315-3323).

[11] Chouldechova, A., G'Sell, M., Li, C., Schuler, A., & Sobel, J. (2018). Fairer and more accurate, but for whom? In Proceedings
of the Conference on Fairness, Accountability, and Transparency (pp. 107-116).

[12] Zemel, R., Wu, Y., Swersky, K., Pitassi, T., & Dwork, C. (2013). Learning fair representations. In International Conference on
Machine Learning (pp. 325-333).

[13] Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K., & Galstyan, A. (2019). A Survey on Bias and Fairness in Machine
Learning. arXiv preprint arXiv:1908.09635.

[14] Kline, A., Shrime, M., & Lepine, A. (2019). What should we make of machine learning? Surgery, prediction, and knowing and
caring. Theoretical Medicine and Bioethics, 40(1), 29-44.

[15] Grewal, P., Gondal, I., & Klapuri, A. (2018). Automated analysis of diabetic retinopathy images using image processing and
machine learning. In 2018 15th International Conference on Control, Automation, Robotics and Vision (pp. 1161-1166). IEEE.

[16] Hardin, B. J., & Posick, C. (2019). Predictive Policing and Fairness: Crime Prediction versus Crime Prevention. Police
Quarterly, 22(2), 190-216.
[17] Berk, R. (2017). Prediction as a scientific (and social) problem. Significance, 14(5), 32-37.

[18] Berk, R., Heidari, H., Jabbari, S., Joseph, M., Kearns, M., Morgenstern, J., ... & Roth, A. (2018). Fairness in Criminal Justice
Risk Assessments: The State of the Art. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society (pp. 59-66).

[19] Kearns, M., Neel, S., Roth, A., & Wu, Z. S. (2018). Preventing Fairness Gerrymandering: Auditing and Learning for Subgroup
Fairness. In Proceedings of the 35th International Conference on Machine Learning (Vol. 80, pp. 2566-2574).

[20] Dwork, C., Feldman, V., Hardt, M., Pitassi, T., Reingold, O., & Roth, A. (2012). The Price of Fairness and How to Get It for
Free. In Proceedings of the 2012 ACM Conference on Computer and Communications Security (pp. 56-68).

[21] Kleinberg, J., Ludwig, J., Mullainathan, S., & Sunstein, C. R. (2019). Discrimination in Online Ad Delivery. Science,
366(6464), 447-451.

8|Page

You might also like