Repository logo

Enhancing Fairness in Supervised Machine Learning

dc.contributor.authorOmidi, Bita
dc.contributor.supervisorSokolova, Marina
dc.contributor.supervisorYeap, Tet
dc.date.accessioned2021-04-23T19:45:43Z
dc.date.available2021-04-23T19:45:43Z
dc.date.issued2021-04-23en_US
dc.description.abstractThe increasing influence of machine learning algorithms and artificial intelligence on the high-impact domains of decision-making has led to an increasing concern for the ethical and legal challenges posed by sensitive data-driven systems. Machine learning can identify the statistical patterns in the historically collected big data generated by a huge number of instances that might be affected by human and structural biases. ML algorithms have the potential to amplify these inequities. Lately, there have been several attempts to reduce bias in artificial intelligence in order to maintain fairness in machine learning projects. These methods fall under three categories of pre-processing, in-processing, and post-processing techniques. There are at least 21 notations of fairness in the recent literature, which not only provide different measurement methods of fairness but also lead to completely different concepts. It is worth mentioning that, it is impossible to satisfy all of the definitions of fairness at the same time and some of them are incompatible with each other. As a result, it is important to choose a fairness definition that need to be satisfied according to the context that we are working on. The current study investigates some of the most common definitions and metrics for fairness introduced by researchers to compare three of the proposed de-biasing techniques regarding their effects on the performance and fairness measures through empirical experiments on four different datasets. The de-biasing methods include the “Reweighing Algorithm”, “Adversarial De-biasing Method”, the “Reject Option Classification Method” performed on the classification tasks of “Survival of patients with heart failure”(Heart Failure Dataset), “Prediction of hospital readmission among diabetes patients” (Diabetes Dataset), “Credit classification of bank account holders” (German Credit Dataset), and “The COVID19 related anxiety level classification of Canadians” (CAMH Dataset). Findings show that the adversarial de-biasing in-processing method can be the best technique for mitigating bias working with the deep learning classifiers when we are capable of changing the classification process. This method has not led to a considerable reduction of accuracy except for the CAMH dataset. The “Reject Option Classification” which is a post-processing method, causes the most deterioration of prediction accuracy in all datasets. On the other hand, this method has the best performance in alleviating the bias generated through the classification process. The “Reweighing Algorithm” as a pre-processing technique does not cause a considerable reduction in the accuracy and is capable of reducing bias in classification tasks, although its performance is not as strong as the Reject Option classifier.en_US
dc.identifier.urihttp://hdl.handle.net/10393/42036
dc.identifier.urihttp://dx.doi.org/10.20381/ruor-26258
dc.language.isoenen_US
dc.publisherUniversité d'Ottawa / University of Ottawaen_US
dc.subjectMachine Learningen_US
dc.subjectClassification algorithmsen_US
dc.subjectFairness metricsen_US
dc.subjectDe-biasingen_US
dc.titleEnhancing Fairness in Supervised Machine Learningen_US
dc.typeThesisen_US
thesis.degree.disciplineGénie / Engineeringen_US
thesis.degree.levelMastersen_US
thesis.degree.nameMScen_US
uottawa.departmentScience informatique et génie électrique / Electrical Engineering and Computer Scienceen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail ImageThumbnail Image
Name:
Omidi_Bita_2021_thesis.pdf
Size:
1.31 MB
Format:
Adobe Portable Document Format
Description:

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail ImageThumbnail Image
Name:
license.txt
Size:
6.65 KB
Format:
Item-specific license agreed upon to submission
Description: