Taming Technical Bias in Machine Learning Pipelines

Open Access
Authors
Publication date 12-2020
Journal Bulletin of the Technical Committee on Data Engineering
Volume | Issue number 43 | 4
Pages (from-to) 39-50
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Machine Learning (ML) is commonly used to automate decisions in domains as varied as credit and lending, medical diagnosis, and hiring. These decisions are consequential, imploring us to carefully balance the benefits of efficiency with the potential risks. Much of the conversation about the risks centers around bias — a term that is used by the technical community ever more frequently but that is still poorly understood. In this paper we focus on technical bias — a type of bias that has so far received limited attention and that the data engineering community is well-equipped to address. We discuss dimensions of technical bias that can arise through the ML lifecycle, particularly when it’s due to preprocessing decisions or post-deployment issues. We present results of our recent work, and discuss future research directions. Our over-all goal is to support the development of systems that expose the knobs of responsibility to data scientists, allowing them to detect instances of technical bias and to mitigate it when possible.
Document type Article
Language English
Published at http://sites.computer.org/debull/A20dec/p39.pdf
Other links http://sites.computer.org/debull/A20dec/issue1.htm
Downloads
p39 (Final published version)
Permalink to this page
Back