{"id":9857,"date":"2020-09-25T08:07:55","date_gmt":"2020-09-25T08:07:55","guid":{"rendered":"https:\/\/www.experfy.com\/blog\/?p=9857"},"modified":"2023-10-30T13:32:57","modified_gmt":"2023-10-30T13:32:57","slug":"programming-fairness-in-algorithms","status":"publish","type":"post","link":"https:\/\/www.experfy.com\/blog\/ai-ml\/programming-fairness-in-algorithms\/","title":{"rendered":"Programming Fairness in Algorithms"},"content":{"rendered":"\t\t<div data-elementor-type=\"wp-post\" data-elementor-id=\"9857\" class=\"elementor elementor-9857\" data-elementor-post-type=\"post\">\n\t\t\t\t\t\t<section class=\"has_eae_slider elementor-section elementor-top-section elementor-element elementor-element-4a567f8f elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"4a567f8f\" data-element_type=\"section\" data-e-type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"has_eae_slider elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-ed0b470\" data-id=\"ed0b470\" data-element_type=\"column\" data-e-type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-32676528 elementor-widget elementor-widget-text-editor\" data-id=\"32676528\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\n<p class=\"has-medium-font-size\">Understanding and combating issues of fairness in supervised learning.<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>\u201cBeing good is easy, what is difficult is being just.\u201d\u00a0<em>\u2015\u00a0<\/em><strong><em>Victor Hugo<\/em><\/strong><\/p>\n<p>\u201cWe need to defend the interests of those whom we\u2019ve never met and never will.\u201d\u00a0<em>\u2015\u00a0<\/em><strong><em>Jeffrey D. Sachs<\/em><\/strong><\/p>\n<\/blockquote>\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-082ebea elementor-widget elementor-widget-text-editor\" data-id=\"082ebea\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\n<p id=\"515a\"><strong>Note:\u00a0<\/strong>This article is intended for a general audience to try and elucidate the complicated nature of unfairness in machine learning algorithms. As such, I have tried to explain concepts in an accessible way with minimal use of mathematics, in the hope that everyone can get something out of reading this.<\/p>\n\n\n\n<p id=\"8756\">Supervised machine learning algorithms are inherently discriminatory. They are discriminatory in the sense that they use information embedded in the features of data to separate instances into distinct categories \u2014 indeed, this is their designated purpose in life. This is reflected in the name for these algorithms which are often referred to as discriminative algorithms (splitting data into categories), in contrast to generative algorithms (generating data from a given category). When we use supervised machine learning, this \u201cdiscrimination\u201d is used as an aid to help us categorize our data into distinct categories within the data distribution, as illustrated below.<\/p>\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-f6caacb elementor-widget elementor-widget-image\" data-id=\"f6caacb\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/09\/Home-448x253_9.png\" alt=\"\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-9d2ce6c elementor-widget elementor-widget-text-editor\" data-id=\"9d2ce6c\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\n<p id=\"083b\">Whilst this occurs when we apply discriminative algorithms \u2014 such as support vector machines, forms of parametric regression (e.g. vanilla linear regression), and non-parametric regression (e.g. random forest, neural networks, boosting) \u2014 to any dataset, the outcomes may not necessarily have any moral implications. For example, using last week\u2019s weather data to try and predict the weather tomorrow has no moral valence attached to it. However, when our dataset is based on information that describes people \u2014 individuals, either directly or indirectly, this can inadvertently result in discrimination on the basis of group affiliation.<\/p>\n\n\n\n<p id=\"ed68\">Clearly then, supervised learning is a dual-use technology. It can be used to our benefits, such as for information (e.g. predicting the weather) and protection (e.g. analyzing computer networks to detect attacks and malware). On the other hand, it has the potential to be weaponized to discriminate at essentially any level. This is not to say that the algorithms are evil for doing this, they are merely learning the representations present in the data, which may themselves have embedded within them the manifestations of historical injustices, as well as individual biases and proclivities. A common adage in data science is \u201cgarbage in = garbage out\u201d to refer to models being highly dependent on the quality of the data supplied to them. This can be stated analogously in the context of algorithmic fairness as \u201cbias in = bias out\u201d.<\/p>\n\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-9889db7 elementor-widget elementor-widget-heading\" data-id=\"9889db7\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\"><strong>Data Fundamentalism<\/strong><\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-5f5a607 elementor-widget elementor-widget-text-editor\" data-id=\"5f5a607\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\n<p id=\"8b37\">Some proponents believe in\u00a0<a href=\"https:\/\/hbr.org\/2013\/04\/the-hidden-biases-in-big-data#:~:text=The%20hype%20becomes%20problematic%20when,analytics%20always%20reflect%20objective%20truth.\" target=\"_blank\" rel=\"noreferrer noopener\"><strong>data fundamentalism<\/strong><\/a>, that is to say, that the data reflects the objective truth of the world through empirical observations.<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>\u201cwith enough data, the numbers speak for themselves.\u201d\u00a0<strong><em>\u2014 Former Wired editor-in-chief Chris Anderson (a data fundamentalist)<\/em><\/strong><\/p>\n<p>Data and data sets are not objective; they are creations of human design. We give numbers their voice, draw inferences from them, and define their meaning through our interpretations. Hidden biases in both the collection and analysis stages present considerable risks, and are as important to the big-data equation as the numbers themselves.\u00a0<strong><em>\u2014 Kate Crawford, principal researcher at Microsoft Research Social Media Collective<\/em><\/strong><\/p>\n<\/blockquote>\n\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-59e4313 elementor-widget elementor-widget-text-editor\" data-id=\"59e4313\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\n\n<p id=\"5f8f\">Superficially, this seems like a reasonable hypothesis, but Kate Crawford provides a good counterargument in a\u00a0<a href=\"https:\/\/hbr.org\/2013\/04\/the-hidden-biases-in-big-data#:~:text=The%20hype%20becomes%20problematic%20when,analytics%20always%20reflect%20objective%20truth.\" target=\"_blank\" rel=\"noreferrer noopener\">Harvard Business Review article<\/a>:<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>Boston has a problem with potholes, patching approximately 20,000 every year. To help allocate its resources efficiently, the City of Boston released the excellent\u00a0<a href=\"http:\/\/streetbump.org\/\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">StreetBump smartphone app<\/a>, which draws on accelerometer and GPS data to help passively detect potholes, instantly reporting them to the city. While certainly a clever approach, StreetBump has a signal problem. People in lower income groups in the US are less likely to have smartphones, and this is particularly true of older residents, where smartphone penetration can be as low as 16%. For cities like Boston, this means that smartphone data sets are missing inputs from significant parts of the population \u2014 often those who have the fewest resources. \u2014\u00a0<strong><em>Kate Crawford, principal researcher at Microsoft Research<\/em><\/strong><\/p>\n<\/blockquote>\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-785b120 elementor-widget elementor-widget-text-editor\" data-id=\"785b120\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\n<p id=\"8c53\">Essentially, the StreetBump app picked up a preponderance of data from wealthy neighborhoods and relatively little from poorer neighborhoods. Naturally, the first conclusion you might draw from this is that the wealthier neighborhoods had more potholes, but in reality, there was just a lack of data from poorer neighborhoods because these people were less likely to have smartphones and thus have downloaded the SmartBump app. Often, it is data that we do not have in our dataset that can have the biggest impact on our results. This example illustrates a subtle form of discrimination on the basis of income. As a result, we should be cautious when drawing conclusions such as these from data that may suffer from a \u2018signal problem\u2019. This signal problem is often characterized as\u00a0<a href=\"https:\/\/en.wikipedia.org\/wiki\/Sampling_bias\" target=\"_blank\" rel=\"noreferrer noopener\"><strong>sampling bias<\/strong><\/a>.<\/p>\n\n\n\n<p id=\"bf00\">Another notable example is the \u201cCorrectional Offender Management Profiling for Alternative Sanctions\u201d algorithm or COMPAS for short. This algorithm is used by a number of states across the United States to predict recidivism \u2014 the likelihood that a former criminal will re-offend. Analysis of this algorithm by ProPublica, an investigative journalism organization,\u00a0<a href=\"https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing\" target=\"_blank\" rel=\"noreferrer noopener\">sparked controversy<\/a>\u00a0when it seemed to suggest that the algorithm was discriminating on the basis of race \u2014 a protected class in the United States. To give us a better idea of what is going on, the algorithm used to predict recidivism looks something like this:<\/p>\n\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-8003e53 elementor-widget elementor-widget-heading\" data-id=\"8003e53\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h3 class=\"elementor-heading-title elementor-size-default\"><strong>Recidivism Risk Score<\/strong>\u00a0= (age*\u2212w)+(age-at-first-arrest*\u2212w)+(history of violence*w) + (vocation education * w) + (history of noncompliance * w)<\/h3>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<section class=\"has_eae_slider elementor-section elementor-top-section elementor-element elementor-element-436cbe3 elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"436cbe3\" data-element_type=\"section\" data-e-type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"has_eae_slider elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-5cddf56\" data-id=\"5cddf56\" data-element_type=\"column\" data-e-type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-feba88a elementor-widget elementor-widget-text-editor\" data-id=\"feba88a\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p>Do you have a feeling that deep learning on graphs is a bunch of heuristics that work sometimes and nobody has a clue why? In this post, I discuss the graph isomorphism problem, the Weisfeiler-Lehman heuristic for graph isomorphism testing, and how it can be used to analyse the expressive power of graph neural networks.<\/p>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-5206e11 elementor-widget elementor-widget-text-editor\" data-id=\"5206e11\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\n<p id=\"8637\">It should be clear that race is not one of the variables used as a predictor. However, the data distribution between two given races may be significantly different for some of these variables, such as the \u2018history of violence\u2019 and \u2018vocation education\u2019 factors, based on historical injustices in the United States as well as demographic, social, and law enforcement statistics (which are often another target for criticism since they often use algorithms to determine which neighborhoods to patrol). The mismatch between these data distributions can be leveraged by an algorithm, leading to disparities between races and thus to some extent a result that is moderately biased towards or against certain races. These entrenched biases will then be operationalized by the algorithm and continue to persist as a result, leading to further injustices. This loop is essentially a\u00a0<a href=\"https:\/\/en.wikipedia.org\/wiki\/Self-fulfilling_prophecy\" rel=\"noopener\">self-fulfilling prophecy<\/a>.<\/p>\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-a68d5cb elementor-widget elementor-widget-heading\" data-id=\"a68d5cb\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\"><strong>Historical Injustices \u2192 Training Data \u2192 Algorithmic Bias in Production<\/strong><\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-0b114bf elementor-widget elementor-widget-text-editor\" data-id=\"0b114bf\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\n<p id=\"9e28\">This leads to some difficult questions \u2014 do we remove these problematic variables? How do we determine whether a feature will lead to discriminatory results? Do we need to engineer a metric that provides a threshold for \u2018discrimination\u2019? One could take this to the extreme and remove almost all variables, but then the algorithm would be of no use. This paints a bleak picture, but fortunately, there are ways to tackle these issues that will be discussed later in this article.<\/p>\n\n\n\n<p id=\"e887\">These examples are not isolated incidents. Even breast cancer prediction algorithms show a level of unfair discrimination. Deep learning algorithms to predict breast cancer from mammograms are much less accurate for black women than white women. This is partly because the dataset used to train these algorithms is predominantly based on mammograms of white women, but also because the data distribution for breast cancer between black women and white women likely has substantial differences.\u00a0<a href=\"https:\/\/www.cdc.gov\/cancer\/dcpc\/research\/articles\/breast_cancer_rates_women.htm\" class=\"broken_link\" rel=\"noopener\">According to the Center for Disease Control (CDC)<\/a>\u00a0\u201c<em>Black women and white women get breast cancer at about the same rate, but black women die from breast cancer at a higher rate than white women<\/em><strong>\u201d<\/strong>.<\/p>\n\n<hr class=\"wp-block-separator\" \/>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-a81affe elementor-widget elementor-widget-text-editor\" data-id=\"a81affe\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\n<p id=\"8765\">These issues raise questions about the motives of algorithmic developers \u2014 did the individuals that designed these models do so knowingly? Do they have an agenda they are trying to push and trying to hide it inside gray box machine learning models?<\/p>\n\n\n\n<p id=\"b6eb\">Although these questions are impossible to answer with certainty, it is useful to consider Hanlon\u2019s razor when asking such questions:<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>Never attribute to malice that which is adequately explained by stupidity \u2014\u00a0<strong><em>Robert J. Hanlon<\/em><\/strong><\/p>\n<\/blockquote>\n\n\n\n<p id=\"a16f\">In other words, there are not that many evil people in the world (thankfully), and there are certainly less evil people in the world than there are incompetent people. On average, we should assume that when things go wrong it is more likely attributable to incompetence, naivety, or oversight than to outright malice. Whilst there are likely some malicious actors who would like to push discriminative agendas, these are likely a minority.<\/p>\n\n\n\n<p id=\"7b27\">Based on this assumption, what could have gone wrong? One could argue that statisticians, machine learning practitioners, data scientists, and computer scientists are not adequately taught how to develop supervised learning algorithms that control and correct for prejudicial proclivities.<\/p>\n\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-235b268 elementor-widget elementor-widget-heading\" data-id=\"235b268\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Why is this the case?<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-88cb1e0 elementor-widget elementor-widget-text-editor\" data-id=\"88cb1e0\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p id=\"98cf\">In truth, techniques that achieve this do not exist. Machine learning fairness is a young subfield of machine learning that has been growing in popularity over the last few years in response to the rapid integration of machine learning into social realms. Computer scientists, unlike doctors, are not necessarily trained to consider the ethical implications of their actions. It is only relatively recently (one could argue since the advent of social media) that the designs or inventions of computer scientists were able to take on an ethical dimension.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"f75c\">This is demonstrated in the fact that most computer science journals do not require ethical statements or considerations for submitted manuscripts. If you take an image database full of millions of images of real people, this can without a doubt have ethical implications. By virtue of physical distance and the size of the dataset, computer scientists are so far removed from the data subjects that the implications on any one individual may be perceived as negligible and thus disregarded. In contrast, if a sociologist or psychologist performs a test on a small group of individuals, an entire ethical review board is set up to review and approve the experiment to ensure it does not transgress across any ethical boundaries.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"7fff\">On the bright side, this is slowly beginning to change. More data science and computer science programs are starting to require students to take classes on data ethics and critical thinking, and journals are beginning to recognize that ethical reviews through IRBs and ethical statements in manuscripts may be a necessary addition to the peer-review process. The rising interest in the topic of <a href=\"https:\/\/www.experfy.com\/blog\/what-machine-learning-data-poisoning\/\" target=\"_blank\" rel=\"noreferrer noopener\">machine learning<\/a> fairness is only strengthening this position.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:separator --><hr class=\"wp-block-separator\" \/><!-- \/wp:separator -->\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-884ebe4 elementor-widget elementor-widget-heading\" data-id=\"884ebe4\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\"><strong>Fairness in Machine Learning<\/strong><\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-09132e1 elementor-widget elementor-widget-image\" data-id=\"09132e1\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/09\/Home-448x253_9.png\" alt=\"\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-c502222 elementor-widget elementor-widget-text-editor\" data-id=\"c502222\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"40dc\">As mentioned previously, widespread adoption of supervised machine learning algorithms has raised concerns about algorithmic fairness. The more these algorithms are adopted, and the increasing control they have on our lives will only exacerbate these concerns. The machine learning community is well aware of these challenges and algorithmic fairness is now a rapidly developing subfield of machine learning with many excellent researchers such as Moritz Hardt, Cynthia Dwork, Solon Barocas, and Michael Feldman.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"2e26\">That being said, there are still major hurdles to overcome before we can achieve truly fair algorithms. It is fairly easy to prevent\u00a0<strong>disparate treatment<\/strong>\u00a0in algorithms \u2014 the explicit differential treatment of one group over another, such as by removing variables that correspond to these attributes from the dataset (e.g. race, gender). However, it is much less easy to prevent\u00a0<strong>disparate impact<\/strong>\u00a0\u2014implicit differential treatment of one group over another, usually caused by something called\u00a0<em>redundant encodings<\/em>\u00a0in the data.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-255774f elementor-widget elementor-widget-image\" data-id=\"255774f\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/09\/Home-448x253_9.png\" alt=\"\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-a6233aa elementor-widget elementor-widget-text-editor\" data-id=\"a6233aa\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"7fd8\">A\u00a0<strong>redundant encoding<\/strong>\u00a0tells us information about a protected attribute, such as race or gender, based on features present in our dataset that correlate with these attributes. For example, buying certain products online (such as makeup) may be highly correlated with gender, and certain zip codes may have different racial demographics that an algorithm might pick up on.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"3adc\">Although an algorithm is not trying to discriminate along these lines, it is inevitable that data-driven algorithms that supersede human performance on pattern recognition tasks might pick up on these associations embedded within data, however small they may be. Additionally, if these associations were non-informative (i.e. they do not increase the accuracy of the algorithm) then the algorithm would ignore them, meaning that some information is clearly embedded in these protected attributes. This raises many challenges to researchers, such as:<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:list -->\n<ul>\n<li>Is there a fundamental tradeoff between fairness and accuracy? Are we able to extract relevant information from protected features without them being used in a discriminatory way?<\/li>\n<li>What is the best statistical measure to embed the notion of \u2018fairness\u2019 within algorithms?<\/li>\n<li>How can we ensure that governments and companies produce algorithms that protect individual fairness?<\/li>\n<li>What biases are embedded in our training data and how can we mitigate their influence?<\/li>\n<\/ul>\n<!-- \/wp:list -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-714f290 elementor-widget elementor-widget-text-editor\" data-id=\"714f290\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"1738\">We will touch upon some of these questions in the remainder of the article.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:separator --><hr class=\"wp-block-separator\" \/><!-- \/wp:separator -->\n\n<!-- wp:heading -->\n<h2 id=\"47ed\">The Problem with Data<\/h2>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph -->\n<p id=\"37ca\">In the last section, it was mentioned that redundant encodings can lead to features correlating with protected attributes. As our data set scales in size, the likelihood of the presence of these correlations scales accordingly. In the age of big data, this presents a big problem:<strong>\u00a0the more data we have access to, the more information we have at our disposal to discriminate<\/strong>. This discrimination does not have to be purely race- or gender-based, it could manifest as discrimination against individuals with pink hair, against web developers, against Starbucks coffee drinkers, or a combination of all of these groups. In this section, several biases present in training data and algorithms are presented that complicate the creation of fair algorithms.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-cac49e3 elementor-widget elementor-widget-heading\" data-id=\"cac49e3\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\"><strong>The Majority Bias<\/strong><\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-3f4ff70 elementor-widget elementor-widget-text-editor\" data-id=\"3f4ff70\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"3df9\">Algorithms have no affinity to any particular group, however, they do have a proclivity for the majority group due to their statistical basis. As outlined by Professor Moritz Hardt in a Medium\u00a0<a href=\"https:\/\/medium.com\/@mrtz\/how-big-data-is-unfair-9aa544d739de\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">article<\/a>, classifiers generally improve with the number of data points used to train them since the error scales with the inverse square root of the number of samples, as shown below.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-91c2a8c elementor-widget elementor-widget-image\" data-id=\"91c2a8c\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/09\/Home-448x253_9.png\" alt=\"\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-3b6fc85 elementor-widget elementor-widget-text-editor\" data-id=\"3b6fc85\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"5191\">This leads to an unsettling reality that since there will, by definition, always be less data available about minorities, our models will tend to perform worse on those groups than on the majority. This assumption is only true if the majority and minority groups are drawn from separate distributions, if they are drawn from a single distribution then increasing sample size will be equally beneficial to both groups.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"ed3e\">An<a href=\"https:\/\/www.engadget.com\/2019-05-07-mit-ai-model-breast-cancer.html\" target=\"_blank\" rel=\"noreferrer noopener\">\u00a0example<\/a>\u00a0of this is the breast cancer detection algorithms we discussed previously. For this\u00a0<a href=\"https:\/\/www.csail.mit.edu\/news\/using-ai-predict-breast-cancer-and-personalize-care\" target=\"_blank\" rel=\"noreferrer noopener\">deep learning model<\/a>, developed by researchers at MIT, of the 60,000 mammogram images in the dataset used to train the neural network, only 5% were mammograms of black women, who are 43% more likely to die from breast cancer. As a result of this, the algorithm performed more poorly when tested on black women, and minority groups in general. This could partially be accounted for because\u00a0<a href=\"https:\/\/www.vice.com\/en_us\/article\/neqaww\/breast-cancer-screening-age-women-of-color\" target=\"_blank\" rel=\"noreferrer noopener\">breast cancer often manifests at an earlier age among women of color<\/a>, which indicates a disparate impact because the probability distribution of women of color was underrepresented.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"7d87\">This also presents another important question. Is accuracy a suitable proxy for fairness? In the above example, we assumed that a lower classification accuracy on a minority group corresponds to unfairness. However, due to the widely differing definitions and the somewhat ambiguous nature of fairness, it can sometimes be difficult to ensure that the variable we are measuring is a good proxy for fairness. For example, our algorithm may have 50% accuracy for both black and white women, but if there 30% false positives for white women and 30% false negatives for black women, this would also be indicative of disparate impact.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"1d81\">From this example, it seems almost intuitive that this is a form of discrimination since there is differential treatment on the basis of group affiliation. However, there are times when this group affiliation is informative to our prediction. For example, for an e-commerce website trying to decide what content to show its users, having an idea of the individual\u2019s gender, age, or socioeconomic status is incredibly helpful. This implies that if we merely remove protected fields from our data, we will decrease the accuracy (or some other performance metric) of our model. Similarly, if we had sufficient data on both black and white women for the breast cancer model, we could develop an algorithm that used race as one of the inputs. Due to the differences in data distributions between the races, it is likely that the accuracy would have increased for both groups.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"d50a\"><mark><strong>Thus, the ideal case would be to have an algorithm that contains these protected features and uses them to make algorithmic generalizations but is constrained by fairness metrics to prevent the algorithm from discriminating.<\/strong><\/mark><\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"dae7\">This is an idea proposed by Moritz Hardt and Eric Price in \u2018<a href=\"https:\/\/arxiv.org\/abs\/1610.02413\" target=\"_blank\" rel=\"noreferrer noopener\">Equality of Opportunity in Supervised Learning<\/a>\u2019. This has several advantages over other metrics, such as statistical parity and equalized odds, but we will discuss all three of these methods in the next section.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:separator --><hr class=\"wp-block-separator\" \/><!-- \/wp:separator -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-0138cbe elementor-widget elementor-widget-heading\" data-id=\"0138cbe\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Definitions of Fairness<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-286d142 elementor-widget elementor-widget-text-editor\" data-id=\"286d142\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"f2e6\">In this section we analyze some of the notions of fairness that have been proposed by machine learning fairness researchers. Namely, statistical parity, and then nuances of statistical parity such as equality of opportunity and equalized odds.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-5f46bce elementor-widget elementor-widget-heading\" data-id=\"5f46bce\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h3 class=\"elementor-heading-title elementor-size-default\">Statistical Parity<\/h3>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-cc2eeb7 elementor-widget elementor-widget-text-editor\" data-id=\"cc2eeb7\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"d9c6\">Statistical parity is the oldest and simplest method of enforcing fairness. It is expanded upon greatly in the arXiv article<em>\u00a0\u201c<\/em><a href=\"https:\/\/arxiv.org\/pdf\/1701.08230.pdf\" rel=\"noopener\"><em>Algorithmic decision making and the cost of fairness<\/em><\/a><em>\u201d<\/em>\u00a0The formula for statistical parity is shown below.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-55ac8b8 elementor-widget elementor-widget-image\" data-id=\"55ac8b8\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/09\/Home-448x253_9.png\" alt=\"\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-384f6c1 elementor-widget elementor-widget-text-editor\" data-id=\"384f6c1\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"936f\">For statistical parity, the outcome will be independent of my group affiliation. What does this mean intuitively? It means that the same proportion of each group will be classified as positive or negative. For this reason, we can also describe statistical parity as\u00a0<strong>demographic parity<\/strong>. For all demographic groups subsumed within\u00a0<em>p<\/em>, statistical parity will be enforced.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"9ec3\">For a dataset that has not had statistical parity applied, we can measure how far our predictions deviate from statistical parity by calculating the statistical parity distance shown below.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-fca12bc elementor-widget elementor-widget-image\" data-id=\"fca12bc\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/09\/Home-448x253_9.png\" alt=\"\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-f0cca24 elementor-widget elementor-widget-text-editor\" data-id=\"f0cca24\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"b9f9\">This distance can provide us with a metric for how fair or unfair a given dataset is based on the group affiliation\u00a0<em>p.<\/em><\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"a2e5\">What are the tradeoffs of using statistical parity?<\/p>\n<!-- \/wp:paragraph -->\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-e65f7c9 elementor-widget elementor-widget-heading\" data-id=\"e65f7c9\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\">Statistical parity doesn\u2019t ensure fairness.<\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-aad1337 elementor-widget elementor-widget-text-editor\" data-id=\"aad1337\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"baa7\">As you may have noticed though, statistical parity says nothing about the accuracy of these predictions. One group may be much more likely to be predicted as positive than another, and hence we might obtain large disparities between the false positive and true positive rates for each group. This itself can cause a disparate impact as qualified individuals from one group (<em>p=0<\/em>)<em>\u00a0may\u00a0<\/em>be missed out in favor of unqualified individuals from another group (<em>p=1<\/em>). In this sense, statistical parity is more akin to\u00a0<a href=\"https:\/\/en.wikipedia.org\/wiki\/Equality_of_outcome\" target=\"_blank\" rel=\"noreferrer noopener\">equality of outcome<\/a>.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"5dd0\">The figures below illustrate this nicely. If we have two groups \u2014 one with 10 people (group A=1), and one with 5 people (group A=0) \u2014 and we determine that 8 people (80%) in group A=1 achieved a score of Y=1, then 4 people (80%) in group A=0 would also have to be given a score of Y=1, regardless of other factors.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-4861ca0 elementor-widget elementor-widget-image\" data-id=\"4861ca0\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/09\/Home-448x253_9.png\" alt=\"\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-83c80fa elementor-widget elementor-widget-image\" data-id=\"83c80fa\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/09\/Home-448x253_9.png\" alt=\"\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-dbb3768 elementor-widget elementor-widget-heading\" data-id=\"dbb3768\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\"><strong>Statistical parity reduces algorithmic accuracy<\/strong><\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-81137e8 elementor-widget elementor-widget-text-editor\" data-id=\"81137e8\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"5485\">The second problem with statistical parity is that a protected class may provide some information that would be useful for a prediction, but we are unable to leverage that information because of the strict rule imposed by statistical parity. Gender might be very informative for making predictions about items that people might buy, but if we are prevented from using it, our model becomes weaker and accuracy is impacted. A better method would allow us to account for the differences between these groups without generating disparate impact. Clearly, statistical parity is misaligned with the fundamental goal of accuracy in machine learning \u2014 the perfect classifier may not ensure demographic parity.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"f9ff\">For these reasons, statistical parity is no longer considered a credible option by several machine learning fairness researchers. However, statistical parity is a simple and useful starting point that other definitions of fairness have built upon.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"43ab\">There are slightly more nuanced versions of statistical parity, such as true positive parity, false positive parity, and positive rate parity.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-cf796f2 elementor-widget elementor-widget-heading\" data-id=\"cf796f2\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h3 class=\"elementor-heading-title elementor-size-default\"><strong>True Positive Parity (Equality of Opportunity)<\/strong><\/h3>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-d845692 elementor-widget elementor-widget-text-editor\" data-id=\"d845692\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"c3c2\">This is only possible for binary predictions and performs statistical parity on true positives (the prediction output was 1 and the true output was also 1).<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-41901ea elementor-widget elementor-widget-image\" data-id=\"41901ea\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/09\/Home-448x253_9.png\" alt=\"\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-52dfcb4 elementor-widget elementor-widget-text-editor\" data-id=\"52dfcb4\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"1e69\">It ensures that in both groups, of all those who qualified (Y=1), an equal proportion of individuals will be classified as qualified (C=1). This is useful when we are only interested in parity over the positive outcome.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<section class=\"has_eae_slider elementor-section elementor-top-section elementor-element elementor-element-b0abdca elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"b0abdca\" data-element_type=\"section\" data-e-type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"has_eae_slider elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-0a10ab0\" data-id=\"0a10ab0\" data-element_type=\"column\" data-e-type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-4ccee64 elementor-widget elementor-widget-image\" data-id=\"4ccee64\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/09\/Home-448x253_9.png\" alt=\"\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-02f27f0 elementor-widget elementor-widget-heading\" data-id=\"02f27f0\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h3 class=\"elementor-heading-title elementor-size-default\">False Positive Parity<\/h3>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-38246ba elementor-widget elementor-widget-text-editor\" data-id=\"38246ba\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"c02a\">This is also only applicable to binary predictions and focuses on false positives (the prediction output was 1 but the true output was 0). This is analogous to the true positive rate but provides parity across false positive results instead.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-27b5bc3 elementor-widget elementor-widget-heading\" data-id=\"27b5bc3\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h3 class=\"elementor-heading-title elementor-size-default\"><strong>Positive Rate Parity (Equalized Odds)<\/strong><\/h3>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-9e41bb0 elementor-widget elementor-widget-text-editor\" data-id=\"9e41bb0\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"c300\">This is a combination of statistical parity for true positives and false positives simultaneously and is also know as\u00a0<strong>equalized odds<\/strong>.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-3499c31 elementor-widget elementor-widget-image\" data-id=\"3499c31\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/09\/Home-448x253_9.png\" alt=\"\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-1b1fb3a elementor-widget elementor-widget-text-editor\" data-id=\"1b1fb3a\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"2f70\">Notice that for equal opportunity, we relax the condition of equalized odds that odds must be equal in the case that Y=0. Equalized odds and equality of opportunity are also more flexible and able to incorporate some of the information from the protected variable without resulting in disparate impact.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"34cc\">Notice that whilst all of these provide some form of a solution that can be argued to be fair, none of these are particularly satisfying. One reason for this is that there are many conflicting definitions of what fairness entails, and it is difficult to capture these in algorithmic form. These are good starting points but there is still much room for improvement.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:separator --><hr class=\"wp-block-separator\" \/><!-- \/wp:separator -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-6b31077 elementor-widget elementor-widget-heading\" data-id=\"6b31077\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Other Methods to Increase Fairness<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-f35d7f8 elementor-widget elementor-widget-text-editor\" data-id=\"f35d7f8\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"7af2\">Statistical parity, equalized odds, and equality of opportunity are all great starting points, but there are other things we can do to ensure that algorithms are not used to unduly discriminate individuals. Two such solutions which have been proposed are human-in-the-loop and algorithmic transparency.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-f30b9ed elementor-widget elementor-widget-heading\" data-id=\"f30b9ed\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Human-in-the-Loop<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-a2b5734 elementor-widget elementor-widget-text-editor\" data-id=\"a2b5734\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"9c98\">This sounds like some kind of rollercoaster ride, but it merely refers to a paradigm whereby a human oversees the algorithmic process.\u00a0<a href=\"https:\/\/en.wikipedia.org\/wiki\/Human-in-the-loop\" target=\"_blank\" rel=\"noreferrer noopener\">Human-in-the-loop<\/a>\u00a0is often implemented in situations that have high risks if the algorithm makes a mistake. For example, missile detection systems that inform the military when a missile is detected allow individuals to review the situation and decide how to respond \u2014 the algorithm does not respond without human interaction. Just imagine the catastrophic consequences of running nuclear weapon systems with AI that had permission to fire when they detected a threat \u2014 one false positive and the entire world would be doomed.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"abbd\">Another example of this is the COMPAS system for recividism \u2014 the system does not categorize you as a recidivist and make a legal judgment. Instead, the judge reviews the COMPAS score and uses this as a factor in their evaluation of the circumstance. This raises new questions such as how humans interact with the algorithmic system. Studies using Amazon Mechanical Turk have shown that some individuals will follow the algorithm\u2019s judgment wholeheartedly, as they perceive it to have greater knowledge than a human is likely to, other individuals take its output with a pinch of salt, and some ignore it completely. Research into human-in-the-loop is relatively novel but we are likely to see more of it as machine learning becomes more pervasive in our society.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"9d96\">Another important and similar concept is human-on-the-loop. This is similar to human-in-the-loop, but instead of the human being actively involved in the process, they are passively involved in the algorithm\u2019s oversight. For example, a data analyst might be in charge of monitoring sections of an oil and gas pipeline to ensure that all of the sensors and processes are running appropriately and there are no concerning signals or errors. This analyst is in an oversight position but is not actively involved in the process. Human-on-the-loop is inherently more scalable than human-in-the-loop since it requires less manpower, but it may be untenable in certain circumstances \u2014 such as looking after those nuclear missiles!<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-5c96ca1 elementor-widget elementor-widget-heading\" data-id=\"5c96ca1\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h3 class=\"elementor-heading-title elementor-size-default\">Algorithmic Transparency<\/h3>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-e5f6543 elementor-widget elementor-widget-text-editor\" data-id=\"e5f6543\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"e6ad\">The dominant position in the legal literature for fairness is through algorithmic interpretability and explainability via transparency. The argument is that if an algorithm is able to be viewed publicly and analyzed with scrutiny, then it can be ensured with a high level of confidence that there is no disparate impact built into the model. Whilst this is clearly desirable on many levels, there are some downsides to algorithmic transparency.<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-bca0bc0 elementor-widget elementor-widget-heading\" data-id=\"bca0bc0\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\"><strong>Proprietary algorithms by definition cannot be transparent<\/strong>.<\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-a92591e elementor-widget elementor-widget-text-editor\" data-id=\"a92591e\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"73ad\">From a commercial standpoint, this idea is untenable in most circumstances \u2014 trade secrets or proprietary information may be leaked if algorithms and business processes are provided for all to see. Imagine Facebook or Twitter being asked to release their algorithms to the world so they can be scrutinized to ensure there are no biasing issues. Most likely I could download their code and go and start my own version of Twitter or Facebook pretty easily. Full transparency is only really an option for algorithms used in public services, such as by the government (to some extent), healthcare, the legal system, etc. Since legal scholars are predominantly concerned with the legal system, it makes sense that this remains the consensus at the current time.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"773d\">In the future, perhaps regulations on algorithmic fairness may be a more tenable solution than algorithmic transparency for private companies that have a vested interest to keep their algorithms from the public eye. Andrew Tutt discusses this idea in his paper \u201c<a href=\"https:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2747994\" class=\"broken_link\" rel=\"noopener\">An FDA For Algorithms<\/a>\u201d, which focused on the development of a regulatory body similar to the FDA to regulate algorithms. Algorithms could be submitted to the regulatory body, or perhaps third party auditing services, and analyzed to ensure they are suitable to be used without resulting in disparate impact.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"632a\">Clearly, such an idea would require large amounts of discussion, money, and expertise to implement, but this seems like a potentially workable solution from my perspective. There is still a long way to go to ensure our algorithms are free of both disparate treatment and disparate impact. With a combination of regulations, transparency, human-in-the-loop, human-on-the-loop, and new and improved variations of statistical parity, we are part of the way there, but this field is still young and there is much work to be done \u2014 watch this space.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:separator --><hr class=\"wp-block-separator\" \/><!-- \/wp:separator -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-9f851cc elementor-widget elementor-widget-heading\" data-id=\"9f851cc\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Final Comments<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-ad705a6 elementor-widget elementor-widget-text-editor\" data-id=\"ad705a6\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"e187\">In this article, we have discussed at length multiple biases present within training data due to the way in which it is collected and analyzed. We have also discussed several ways in which to mitigate the impact of these biases and to help ensure that algorithms remain non-discriminatory towards minority groups and protected classes.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"4b60\">Although machine learning, by its very nature, is always a form of statistical discrimination, the discrimination becomes objectionable when it places certain privileged groups at a systematic advantage and certain unprivileged groups at a systematic disadvantage. Biases in training data, due to either prejudice in labels or under-\/over-sampling, yields models with unwanted bias.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"b301\">Some might say that these decisions were made on less information and by humans, which can have many implicit and cognitive biases influencing their decision. Automating these decisions provides more accurate results and to a large degree limits the extent of these biases. The algorithms do not need to be perfect, just better than what previously existed. The arc of history curves towards justice.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"75f5\">Some might say that algorithms are being given free rein to allow inequalities to be systematically instantiated, or that data itself is inherently biased. That variables related to protected attributes should be removed from data to help mitigate these issues, and any variable correlated with the variables removed or restricted.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"0ada\">Both groups would be partially correct. However, we should not remain satisfied with unfair algorithms, there is also room for improvement. Similarly, we should not waste all of this data we have and remove all variables, as this would make systems perform much worse and would render them much less useful. That being said, at the end of the day, it is up to the creators of these algorithms and oversight bodies, as well as those in charge of collecting data, to try to ensure that these biases are handled appropriately.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"2dda\">Data collection and sampling procedures are often glazed over in statistics classes, and not understood well by the general public. Until such a time as a regulatory body appears, it is up to machine learning engineers, statisticians, and data scientists to ensure the equality of opportunity is embedded in our machine learning practices. We must be mindful of where our data comes from and what we do with it. Who knows who our decisions might impact in the future?<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:quote -->\n<blockquote class=\"wp-block-quote\">\n<p>\u201cThe world isn\u2019t fair, Calvin.\u201d<br \/>\u201cI know Dad, but why isn\u2019t it ever unfair in my favor?\u201d<br \/><em>\u2015\u00a0<\/em><strong><em>Bill Watterson,\u00a0<\/em><\/strong><a href=\"https:\/\/www.goodreads.com\/work\/quotes\/2244438\" target=\"_blank\" rel=\"noreferrer noopener\"><strong><em>The Essential Calvin and Hobbes: A Calvin and Hobbes Treasury<\/em><\/strong><\/a><\/p>\n<\/blockquote>\n<!-- \/wp:quote -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-9f864b5 elementor-widget elementor-widget-heading\" data-id=\"9f864b5\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h2 class=\"elementor-heading-title elementor-size-default\">Further Reading<\/h2>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-c438c9e elementor-widget elementor-widget-text-editor\" data-id=\"c438c9e\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<!-- wp:paragraph -->\n<p id=\"664d\"><strong>[1]<\/strong>\u00a0<a href=\"https:\/\/obamawhitehouse.archives.gov\/sites\/default\/files\/microsites\/ostp\/2016_0504_data_discrimination.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Big Data: A Report on Algorithmic Systems, Opportunity, and Civil Rights<\/a>. The White House. 2016.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"d6fe\"><strong>[2]<\/strong>\u00a0<a href=\"http:\/\/dl.acm.org\/citation.cfm?id=230561\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">Bias in computer systems<\/a>. Batya Friedman, Helen Nissenbaum. 1996<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"8e2f\"><strong>[3]\u00a0<\/strong><a href=\"https:\/\/hbr.org\/2013\/04\/the-hidden-biases-in-big-data\" target=\"_blank\" rel=\"noreferrer noopener\">The Hidden Biases in Big Data<\/a>. Kate Crawford. 2013.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"fe49\"><strong>[4]\u00a0<\/strong><a href=\"https:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2477899\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">Big Data\u2019s Disparate Impact<\/a>. Solon Barocas, Andrew Selbst. 2014.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"688f\"><strong>[5]\u00a0<\/strong>Blog post:\u00a0<a href=\"https:\/\/medium.com\/@mrtz\/how-big-data-is-unfair-9aa544d739de\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">How big data is unfair<\/a>. Moritz Hardt. 2014<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"1fc4\"><strong>[6]\u00a0<\/strong><a href=\"http:\/\/science.sciencemag.org\/content\/356\/6334\/183\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">Semantics derived automatically from language corpora contain human-like biases<\/a>. Aylin Caliskan, Joanna J. Bryson, Arvind Narayanan<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"5982\"><strong>[7]<\/strong>\u00a0<a href=\"http:\/\/science.sciencemag.org\/content\/187\/4175\/398\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">Sex Bias in Graduate Admissions: Data from Berkeley<\/a>. P. J. Bickel, E. A. Hammel, J. W. O\u2019Connell. 1975.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"d1ca\"><strong>[8]\u00a0<\/strong>Simpson\u2019s paradox. Pearl (Chapter 6).\u00a0<a href=\"http:\/\/bayes.cs.ucla.edu\/R264.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Tech report<\/a><\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"2649\"><strong>[9]<\/strong>\u00a0<a href=\"https:\/\/arxiv.org\/abs\/1412.3756\" target=\"_blank\" rel=\"noreferrer noopener\">Certifying and removing disparate impact<\/a>. Michael Feldman, Sorelle Friedler, John Moeller, Carlos Scheidegger, Suresh Venkatasubramanian<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"0260\"><strong>[10]\u00a0<\/strong><a href=\"https:\/\/arxiv.org\/abs\/1610.02413\" target=\"_blank\" rel=\"noreferrer noopener\">Equality of Opportunity in Supervised Learning<\/a>. Moritz Hardt, Eric Price, Nathan Srebro. 2016.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"73d4\"><strong>[11]<\/strong>\u00a0Blog post:\u00a0<a href=\"http:\/\/blog.mrtz.org\/2016\/09\/06\/approaching-fairness.html\" target=\"_blank\" rel=\"noreferrer noopener\">Approaching fairness in machine learning<\/a>. Moritz Hardt. 2016.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"e352\"><strong>[12]\u00a0<\/strong><a href=\"https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing\" target=\"_blank\" rel=\"noreferrer noopener\">Machine Bias<\/a>. Julia Angwin, Jeff Larson, Surya Mattu and Lauren Kirchner, ProPublica. Code review:\u00a0<a href=\"https:\/\/github.com\/propublica\/compas-analysis\" target=\"_blank\" rel=\"noreferrer noopener\">github.com\/probublica\/compas-analysis<\/a>,\u00a0<a href=\"https:\/\/github.com\/adebayoj\/fairml\" target=\"_blank\" rel=\"noreferrer noopener\">github.com\/adebayoj\/fairml<\/a><\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"5233\"><strong>[13]\u00a0<\/strong><a href=\"https:\/\/www.documentcloud.org\/documents\/2998391-ProPublica-Commentary-Final-070616.html\" target=\"_blank\" rel=\"noreferrer noopener\">COMPAS Risk Scales: Demonstrating Accuracy Equity and Predictive Parity<\/a>. Northpointe Inc.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"58fc\"><strong>[14]<\/strong>\u00a0<a href=\"https:\/\/arxiv.org\/abs\/1703.09207\" target=\"_blank\" rel=\"noreferrer noopener\">Fairness in Criminal Justice Risk Assessments: The State of the Art<\/a><br \/>Richard Berk, Hoda Heidari, Shahin Jabbari, Michael Kearns, Aaron Roth. 2017.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"15cc\"><strong>[15]<\/strong>\u00a0<a href=\"http:\/\/www.datacivilrights.org\/pubs\/2015-1027\/Courts_and_Predictive_Algorithms.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Courts and Predictive Algorithms<\/a>. Ang\u00e8le Christin, Alex Rosenblat, and danah boyd. 2015.\u00a0<a href=\"http:\/\/www.datacivilrights.org\/pubs\/2015-1027\/WDN-Courts_and_Predictive_Algorithms.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Discussion paper<\/a><\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"fd69\"><strong>[16]\u00a0<\/strong><a href=\"https:\/\/www.nature.com\/articles\/s41562-017-0141\" target=\"_blank\" rel=\"noreferrer noopener\">Limitations of mitigating judicial bias with machine learning<\/a>. Kristian Lum. 2017.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"1aa9\"><strong>[17]<\/strong>\u00a0<a href=\"http:\/\/citeseer.ist.psu.edu\/viewdoc\/summary?doi=10.1.1.41.1639\" target=\"_blank\" rel=\"noreferrer noopener\">Probabilistic Outputs for Support Vector Machines and Comparisons to Regularized Likelihood Methods<\/a>. John C. Platt. 1999.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"2ba8\"><strong>[18]<\/strong>\u00a0<a href=\"https:\/\/arxiv.org\/abs\/1609.05807\" target=\"_blank\" rel=\"noreferrer noopener\">Inherent Trade-Offs in the Fair Determination of Risk Scores<\/a>. Jon Kleinberg, Sendhil Mullainathan, Manish Raghavan. 2016.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"9f2e\"><strong>[19]\u00a0<\/strong><a href=\"https:\/\/arxiv.org\/abs\/1610.07524\" target=\"_blank\" rel=\"noreferrer noopener\">Fair prediction with disparate impact: A study of bias in recidivism prediction instruments<\/a>. Alexandra Chouldechova. 2016.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"6b90\"><strong>[20]<\/strong>\u00a0<a href=\"https:\/\/research.google.com\/bigpicture\/attacking-discrimination-in-ml\/\" target=\"_blank\" rel=\"noreferrer noopener\">Attacking discrimination with smarter machine learning<\/a>. An interactive visualization by Martin Wattenberg, Fernanda Vi\u00e9gas, and Moritz Hardt. 2016.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"6701\"><strong>[21]<\/strong>\u00a0<a href=\"https:\/\/arxiv.org\/abs\/1701.08230\" target=\"_blank\" rel=\"noreferrer noopener\">Algorithmic decision making and the cost of fairness<\/a>. Sam Corbett-Davies, Emma Pierson, Avi Feller, Sharad Goel, Aziz Huq. 2017.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"e677\"><strong>[22]\u00a0<\/strong><a href=\"https:\/\/5harad.com\/papers\/threshold-test.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">The problem of Infra-marginality in Outcome Tests for Discrimination<\/a>. Camelia Simoiu, Sam Corbett-Davies, Sharad Goel. 2017.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"b0c8\"><strong>[23]<\/strong>\u00a0<a href=\"https:\/\/arxiv.org\/abs\/1610.02413\" target=\"_blank\" rel=\"noreferrer noopener\">Equality of Opportunity in Supervised Learning<\/a>. Moritz Hardt, Eric Price, Nathan Srebro. 2016.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"1d6b\"><strong>[24]\u00a0<\/strong><a href=\"http:\/\/www.math.ku.dk\/~peters\/elements.html\" target=\"_blank\" rel=\"noreferrer noopener\">Elements of Causal Inference<\/a>. Peters, Janzing, Sch\u00f6lkopf<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"3517\"><strong>[25]<\/strong>\u00a0<a href=\"https:\/\/www.ncbi.nlm.nih.gov\/pmc\/articles\/PMC4125322\/\" target=\"_blank\" rel=\"noreferrer noopener\">On causal interpretation of race in regressions adjusting for confounding and mediating variables<\/a>. Tyler J. VanderWeele and Whitney R. Robinson. 2014.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"1502\"><strong>[26]\u00a0<\/strong><a href=\"https:\/\/arxiv.org\/abs\/1703.06856\" target=\"_blank\" rel=\"noreferrer noopener\">Counterfactual Fairness<\/a>. Matt J. Kusner, Joshua R. Loftus, Chris Russell, Ricardo Silva. 2017.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"f9d7\"><strong>[27]<\/strong>\u00a0<a href=\"https:\/\/arxiv.org\/abs\/1706.02744\" target=\"_blank\" rel=\"noreferrer noopener\">Avoiding Discrimination through Causal Reasoning<\/a>. Niki Kilbertus, Mateo Rojas-Carulla, Giambattista Parascandolo, Moritz Hardt, Dominik Janzing, Bernhard Sch\u00f6lkopf. 2017.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"dd5a\"><strong>[28]<\/strong>\u00a0<a href=\"https:\/\/arxiv.org\/abs\/1705.10378\" target=\"_blank\" rel=\"noreferrer noopener\">Fair Inference on Outcomes<\/a>. Razieh Nabi, Ilya Shpitser<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"bbba\"><strong>[29]\u00a0<\/strong><a href=\"https:\/\/arxiv.org\/abs\/1104.3913\" target=\"_blank\" rel=\"noreferrer noopener\">Fairness Through Awareness<\/a>. Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, Rich Zemel. 2012.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"ec19\"><strong>[30]\u00a0<\/strong><a href=\"https:\/\/arxiv.org\/abs\/1609.07236\" target=\"_blank\" rel=\"noreferrer noopener\">On the (im)possibility of fairness<\/a>. Sorelle A. Friedler, Carlos Scheidegger, Suresh Venkatasubramanian. 2016.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"59a4\"><strong>[31]<\/strong>\u00a0<a href=\"https:\/\/gking.harvard.edu\/files\/gking\/files\/psnot.pdf\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">Why propensity scores should not be used<\/a>. Gary King, Richard Nielson. 2016.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"517d\"><strong>[32]<\/strong>\u00a0<a href=\"https:\/\/mitpress.mit.edu\/books\/raw-data-oxymoron\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">Raw Data is an Oxymoron<\/a>. Edited by Lisa Gitelman. 2013.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"0d3b\"><strong>[33]<\/strong>\u00a0Blog post:\u00a0<a href=\"http:\/\/andrewgelman.com\/2015\/04\/28\/whats-important-thing-statistics-thats-not-textbooks\/\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">What\u2019s the most important thing in Statistics that\u2019s not in the textbooks<\/a>. Andrew Gelman. 2015.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"e9bf\"><strong>[34]<\/strong>\u00a0<a href=\"http:\/\/statlab.bio5.org\/sites\/default\/files\/fall2014\/hand-deconstructin.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Deconstructing Statistical Questions<\/a>. David J. Hand. 1994.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"13c4\"><strong>[35]<\/strong>\u00a0<a href=\"http:\/\/www.lps.uci.edu\/~johnsonk\/CLASSES\/MeasurementTheory\/Hand1996.StatisticsAndTheTheoryOfMeasurement.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Statistics and the Theory of Measurement<\/a>. David J. Hand. 1996.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"4b72\"><strong>[36]<\/strong>\u00a0<a href=\"http:\/\/www.wiley.com\/WileyCDA\/WileyTitle\/productCd-0470685670.html\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">Measurement Theory and Practice: The World Through Quantification<\/a>. David J. Hand. 2010<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"24ad\"><strong>[37]<\/strong>\u00a0<a href=\"http:\/\/www.wiley.com\/WileyCDA\/WileyTitle\/productCd-0470465468.html\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">Survey Methodology, 2nd Edition<\/a>. Robert M. Groves, Floyd J. Fowler, Jr., Mick P. Couper, James M. Lepkowski, Eleanor Singer, Roger Tourangeau. 2009<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"f01c\"><strong>[38]<\/strong>\u00a0<a href=\"https:\/\/arxiv.org\/abs\/1607.06520\" target=\"_blank\" rel=\"noreferrer noopener\">Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings<\/a>. Tolga Bolukbasi, Kai-Wei Chang, James Zou, Venkatesh Saligrama, Adam Kalai. 2016.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"f6bf\"><strong>[39]<\/strong>\u00a0<a href=\"https:\/\/arxiv.org\/abs\/1707.09457\" target=\"_blank\" rel=\"noreferrer noopener\">Men Also Like Shopping: Reducing Gender Bias Amplification using Corpus-level Constraints<\/a>. Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, Kai-Wei Chang. 2017.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"da97\"><strong>[40]<\/strong>\u00a0<a href=\"https:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2477899\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">Big Data\u2019s Disparate Impact<\/a>. Solon Barocas, Andrew Selbst. 2014.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"fd67\"><strong>[41]\u00a0<\/strong><a href=\"https:\/\/www.stanfordlawreview.org\/online\/privacy-and-big-data-its-not-privacy-and-its-not-fair\" target=\"_blank\" rel=\"noreferrer noopener\">It\u2019s Not Privacy, and It\u2019s Not Fair<\/a>. Cynthia Dwork, Deirdre K. Mulligan. 2013.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"145e\"><strong>[42]<\/strong>\u00a0<a href=\"http:\/\/journals.sagepub.com\/doi\/abs\/10.1177\/0162243915605575\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">The Trouble with Algorithmic Decisions<\/a>. Tal Zarsky. 2016.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"08d3\"><strong>[43]\u00a0<\/strong><a href=\"https:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=3024938\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">How Copyright Law Can Fix Artificial Intelligence\u2019s Implicit Bias Problem<\/a>. Amanda Levendowski. 2017.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p id=\"ed44\"><strong>[44]<\/strong>\u00a0<a href=\"https:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2747994\" target=\"_blank\" rel=\"noreferrer noopener\" class=\"broken_link\">An FDA for Algorithms<\/a>. Andrew Tutt. 2016<\/p>\n<!-- \/wp:paragraph -->\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<\/div>\n\t\t","protected":false},"excerpt":{"rendered":"<p>Understanding and combating issues of fairness in supervised learning. \u201cBeing good is easy, what is difficult is being just.\u201d\u00a0\u2015\u00a0Victor Hugo \u201cWe need to defend the interests of those whom we\u2019ve never met and never will.\u201d\u00a0\u2015\u00a0Jeffrey D. Sachs Note:\u00a0This article is intended for a general audience to try and elucidate the complicated nature of unfairness in<\/p>\n","protected":false},"author":682,"featured_media":24262,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"content-type":"","footnotes":""},"categories":[183],"tags":[563,92,664],"ppma_author":[3471],"class_list":["post-9857","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai-ml","tag-algorithms","tag-machine-learning","tag-supervised-learning"],"authors":[{"term_id":3471,"user_id":682,"is_guest":0,"slug":"matthew-stewart","display_name":"Matthew Stewart","avatar_url":"https:\/\/www.experfy.com\/blog\/wp-content\/uploads\/2020\/04\/medium_c57055f3-5301-4262-af65-4cc7d40cbf3d-150x150.jpg","user_url":"https:\/\/criticalfutureglobal.com\/","last_name":"Stewart","first_name":"Matthew","job_title":"","description":"Matthew Stewart is a Machine Learning consultant on AI for\u00a0<a href=\"https:\/\/www.criticalfutureglobal.com\/\" target=\"_blank\" rel=\"noopener\">Critical Future<\/a>, and machine learning engineer at Scalable Magic, an AI-based digital media startup. He is also a Graduate Teaching Assistant and a Ph.D. Candidate at Harvard University."}],"_links":{"self":[{"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/posts\/9857","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/users\/682"}],"replies":[{"embeddable":true,"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/comments?post=9857"}],"version-history":[{"count":6,"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/posts\/9857\/revisions"}],"predecessor-version":[{"id":33828,"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/posts\/9857\/revisions\/33828"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/media\/24262"}],"wp:attachment":[{"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/media?parent=9857"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/categories?post=9857"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/tags?post=9857"},{"taxonomy":"author","embeddable":true,"href":"https:\/\/www.experfy.com\/blog\/wp-json\/wp\/v2\/ppma_author?post=9857"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}