keyboard_arrow_up
Mitigation Techniques to Overcome Data Harm in Model Building for ML

Authors

Ayse Arslan, Oxford Alumni of Northern California, USA

Abstract

Given the impact of Machine Learning (ML) on individuals and the society, understanding how harm might be occur throughout the ML life cycle becomes critical more than ever. By offering a framework to determine distinct potential sources of downstream harm in ML pipeline, the paper demonstrates the importance of choices throughout distinct phases of data collection, development, and deployment that extend far beyond just model training. Relevant mitigation techniques are also suggested for being used instead of merely relying on generic notions of what counts as fairness.

Keywords

Fairness in machine learning, societal implications of machine learning, algorithmic bias, AI ethics, allocative harm, representational harm.

Full Text  Volume 11, Number 19