Volume 11, Number 19, November 2021
Mitigation Techniques to Overcome Data Harm in Model Building for ML
Authors
Ayse Arslan, Oxford Alumni of Northern California, USA
Abstract
Given the impact of Machine Learning (ML) on individuals and the society, understanding how harm might be occur throughout the ML life cycle becomes critical more than ever. By offering a framework to determine distinct potential sources of downstream harm in ML pipeline, the paper demonstrates the importance of choices throughout distinct phases of data collection, development, and deployment that extend far beyond just model training. Relevant mitigation techniques are also suggested for being used instead of merely relying on generic notions of what counts as fairness.
Keywords
Fairness in machine learning, societal implications of machine learning, algorithmic bias, AI ethics, allocative harm, representational harm.