MLOps Conference

MLOps Conference

On DataOps, productionizing ML models, and running experiments at scale.

lavanya TS

@lavanyats

Fairness in ML: How do we build unbiased ML workflows?

Submitted Jun 29, 2021

Biases often arise in automated workflows based on MachineLearning models due to erroneous assumptions made in the learning process. Examples of such biases involve societal biases such as gender bias, racial bias, age bias and so on.

In this 15 minute talk, we hope to cover prominent sources of such biases that lead to ML models producing unwanted outcomes. We will also look at ways to detect and measure such biases in our production workflows.

Outline:

Introduction: Why do we care about biases? (1 min)

Sources of biases (3 minutes)
→ Specification Bias
→ Sampling Bias
→ Measurement Bias
→ Annotator/ Label Biases
→ Inherited Biases from other ML models

Metrics to measure biases(5 minutes)
(Will cover classification use-case only - since it might be hard to do more here)
→ TPR across Groups
→ FPR across Groups
→ Accuracy across groups
→ Demographic Parity

How to avoid biased ML workflows(3 mins)

→ Techniques for Debiasing Data going into model
→ Techniques involving Post Processing model outcome

If time permits: (3 mins)
Biases in word embeddings: Case study: Examples from word2vec embeddings

Closing Remarks:
What is fair and what is not is contextual.
Importance of Human inputs/judgement in designing debiasing techniques in ML workflows

Slides in Progress
https://docs.google.com/presentation/d/1KrZBZLvaEKOHp76Idp33I_9UEK_NisLEukJ8bQ0JACo/edit?usp=sharing

Comments

{{ gettext('Login to leave a comment') }}

{{ gettext('Post a comment…') }}
{{ gettext('New comment') }}
{{ formTitle }}

{{ errorMsg }}

{{ gettext('No comments posted yet') }}