Measuring Bias in a Model you own this product

prerequisites
beginner Python, Jupyter Notebook, and pandas • basic scikit-learn • basics of machine learning
skills learned
setting up a Google Colab environment • preprocessing a dataset using one-hot encoding • computing bias performance metrics for logistic regression
Adriano Soares Koshiyama, Umar Mohammed, Emre Kazim, and Catherine Inness
1 week · 5-9 hours per week · INTERMEDIATE

pro $24.99 per month

  • access to all Manning books, MEAPs, liveVideos, liveProjects, and audiobooks!
  • choose one free eBook per month to keep
  • exclusive 50% discount on all purchases

lite $19.99 per month

  • access to all Manning books, including MEAPs!

team

5, 10 or 20 seats+ for your team - learn more


Look inside
In this liveProject, you’ll investigate and report on whether your company’s mortgage application classifier is making fair decisions between male and female applicants. You’ll train a logistic regression classifier on the HMDA dataset, compute performance metrics with a logistic regression classifier, and then chart “equality of opportunity” bias metrics.
This project is designed for learning purposes and is not a complete, production-ready application or solution.

project authors

Adriano Koshiyama
Adriano Koshiyama is a Research Fellow in Computer Science at University College London, and the co-founder of Holistic AI, a start-up focused on providing advisory, auditing and assurance of AI systems. He helps manage TheAlgo Conferences (thealgo.co) and is the main investigator at UCL Algorithm Standards and Technology Lab. Academically, he has published more than 30 papers in international conferences and journals.
Catherine Inness
Catherine Inness is a senior manager in Accenture's Data Science practice in the UK. She has more than ten years of experience in technology, focused on the public sector. She holds an MSc in data science and machine learning from University College London, where her research focused on algorithm fairness.
Umar Mohammed
Umar Mohammed is the lead developer at Holistic AI. He has over 10 years of experience writing software and is currently focused on creating debiasing tools for AI systems. He holds an MSc in vision imaging and virtual environments from University College London where his research focused on face recognition. He has published papers on computer vision in international conferences.
Emre Kazim
Emre Kazim is a research fellow in the computer science department of University College London, working in the field of AI ethics. His current focus is on governance, policy, and auditing of AI systems, including algorithm interpretability and certification. Emre has a PhD in philosophy.

prerequisites

The liveProject is for beginner data scientists and software engineers looking to tackle the basic principles of measuring and mitigating ML bias. To begin this liveProject, you will need to be familiar with:

TOOLS
  • Basic Python
  • Basic Jupyter Notebook
  • Basic pandas
  • Basic scikit-learn
  • Basic seaborn
TECHNIQUES
  • Basic machine learning

features

Self-paced
You choose the schedule and decide how much time to invest as you build your project.
Project roadmap
Each project is divided into several achievable steps.
Get Help
While within the liveProject platform, get help from other participants and our expert mentors.
Compare with others
For each step, compare your deliverable to the solutions by the author and other participants.
book resources
Get full access to select books for 90 days. Permanent access to excerpts from Manning products are also included, as well as references to other resources.

choose your plan

team

monthly
annual
$49.99
$499.99
only $41.67 per month
  • five seats for your team
  • access to all Manning books, MEAPs, liveVideos, liveProjects, and audiobooks!
  • choose another free product every time you renew
  • choose twelve free products per year
  • exclusive 50% discount on all purchases
  • Measuring Bias in a Model project for free