Systematic and unfair discrimination that occurs when algorithms used in recruiting and hiring decisions produce outcomes that disproportionately favor or disadvantage candidates based on protected characteristics like race, gender, age, or disability.
Key Takeaways
Algorithmic bias in hiring is what happens when the tools designed to make recruiting fairer actually make it worse. It sounds counterintuitive. The whole point of using algorithms is to remove human subjectivity. But algorithms learn from data, and historical hiring data is full of human biases. If a company has historically hired fewer women in engineering roles, an ML model trained on that data will learn to rate male candidates higher. It isn't making a conscious choice. It's doing exactly what it was trained to do: replicate past patterns. The problem is bigger than one bad data set. Bias can enter at every stage of the algorithmic pipeline. The features you choose to include in the model matter. The way you define "success" matters. The population of candidates in your training data matters. Even the way a job description is written can introduce bias before the algorithm ever sees a resume. What makes algorithmic bias particularly dangerous is scale. A biased human recruiter might review 200 resumes and unconsciously favor certain candidates. A biased algorithm processes 200,000 resumes and does the same thing, consistently, every time. It amplifies bias rather than reducing it.
Algorithmic bias isn't a single problem. It shows up in different forms depending on where in the model development process things go wrong.
The most common type. The training data reflects past discrimination. If women have been underrepresented in technical roles at your company for the past decade, the algorithm learns that being male is a predictor of getting hired. It doesn't know this is discrimination. It sees a statistical pattern and weights it accordingly. Historical bias exists in nearly every organization's hiring data because human biases have influenced hiring decisions for decades.
This occurs when the proxy you're using to measure "good performance" is itself biased. Many companies train hiring algorithms on performance review scores. But performance reviews are subjective and influenced by manager bias. If your performance reviews systematically underrate women or minorities (which research shows they often do), your algorithm will learn that those groups are worse performers. The bias isn't in the algorithm. It's in the label you're asking it to predict.
Your training data doesn't reflect the actual candidate population. If your historical applicant pool is 80% from three universities and 20% from everywhere else, the model will be better at evaluating candidates who resemble your historical applicants. It doesn't know what to do with non-traditional backgrounds because it hasn't seen enough examples. This particularly disadvantages candidates from underrepresented backgrounds who may have taken non-traditional career paths.
Even when protected characteristics are removed from the model, other features can serve as proxies. Zip code correlates with race. Name patterns correlate with ethnicity. College attended correlates with socioeconomic status. Graduation year correlates with age. An algorithm that's technically "blind" to protected characteristics can still discriminate through these proxies. This is why simply removing demographic fields from the model doesn't solve the bias problem.
These cases illustrate how algorithmic bias manifests in practice and the consequences for organizations.
| Case | What Happened | Bias Type | Outcome |
|---|---|---|---|
| Amazon resume screener (2018) | ML model trained on 10 years of hiring data penalized resumes with "women's" and female-associated terms | Historical bias + feature bias | Project scrapped; never used in production hiring decisions |
| HireVue video analysis (2019-2021) | Video interview AI scored candidates on facial expressions and vocal tone, raising concerns about disability and racial bias | Measurement bias | HireVue discontinued facial analysis in 2021 after pressure from civil rights groups |
| Facebook job ad targeting (2019) | Ad algorithm showed job ads to different demographic groups based on engagement patterns, excluding older workers from some roles | Representation bias + feature bias | DOJ settlement requiring changes to ad targeting practices |
| iTutorGroup age filter (2023) | AI tool automatically rejected applicants over 55 for tutoring positions | Direct discrimination (coded rules, not ML) | EEOC settled for $365,000; first EEOC AI bias case |
| Workday class action (2023) | Lawsuit alleging screening algorithms disproportionately rejected Black, disabled, and older applicants | Historical bias + proxy discrimination | Ongoing federal litigation; case allowed to proceed past motion to dismiss |
Bias detection isn't a one-time check. It's an ongoing process that should run every time your model is updated or your hiring demographics shift.
The EEOC's four-fifths rule is the starting point. Calculate the selection rate for each demographic group at every stage of your funnel (application to screen, screen to interview, interview to offer). If any group's selection rate is less than 80% of the highest group's rate, you have a prima facie case of adverse impact. This isn't definitive proof of illegal discrimination, but it triggers an obligation to investigate. Run this analysis monthly or quarterly, not annually.
Go beyond the four-fifths rule and use statistical significance tests. The Fisher's exact test or chi-square test can tell you whether observed differences in selection rates are statistically significant or could be explained by chance. For small sample sizes (fewer than 50 candidates per group), the four-fifths rule is unreliable and statistical testing becomes essential.
Examine which features your model weights most heavily. If zip code, university name, or years of experience are top features, investigate whether they're serving as proxies for protected characteristics. Use SHAP values or LIME (model explainability tools) to understand why the model makes specific predictions. If the model gives a candidate a low score, you should be able to explain which features drove that score.
Create pairs of identical resumes that differ only on a single dimension: male vs female name, traditionally Black vs white name, young vs old graduation year. Run them through your algorithm and compare scores. If identical qualifications produce different scores based on these variations, your model has bias. This technique is borrowed from audit studies in employment discrimination research and translates well to algorithmic testing.
Eliminating bias entirely isn't realistic. The goal is to minimize it to a level that's legally defensible and ethically acceptable, then monitor continuously.
The legal environment around algorithmic hiring bias is evolving quickly. Here's what HR teams need to track.
| Regulation | Jurisdiction | Key Requirements | Effective Date |
|---|---|---|---|
| Title VII (EEOC guidance) | Federal (US) | Employers are liable for disparate impact from algorithms, same as any other selection tool | Existing law; EEOC guidance updated 2023 |
| NYC Local Law 144 | New York City | Annual bias audit by independent auditor; publish results; notify candidates; allow opt-out | July 2023 |
| Illinois AIPA | Illinois | Consent required before AI video interview analysis; data deletion rights | January 2020 |
| Colorado AI Act (SB 205) | Colorado | Developers and deployers of high-risk AI must assess and mitigate bias; notify consumers | February 2026 |
| EU AI Act | European Union | High-risk classification for employment AI; conformity assessments; transparency; human oversight | Phased enforcement 2024-2026 |
| Maryland HB 1202 | Maryland | Employers must get consent before using facial recognition in interviews | October 2020 |
A structured audit program protects your organization legally and ensures your tools work as intended.
Assign a specific person or team as the owner of algorithmic fairness. This is typically a senior HR analytics leader or a dedicated AI governance role. Without clear ownership, audits don't happen. The owner should have authority to pause or modify tools that fail bias tests, regardless of the vendor relationship or the tool's cost.
Audit every tool that influences hiring decisions: resume screeners, matching algorithms, chatbot pre-screening, video interview analysis, and assessment scoring. Run audits quarterly for high-volume tools and annually for lower-volume ones. Every major model update should trigger an audit, even if it falls outside the regular schedule.
Maintain records of every audit, including methodology, findings, remediation actions, and follow-up results. This documentation serves two purposes: it demonstrates good faith if you're ever challenged in court, and it creates institutional knowledge that makes future audits more efficient. The EEOC has explicitly stated that organizations that proactively audit and address algorithmic bias are in a stronger legal position than those who don't.
Current data on the prevalence and impact of algorithmic bias in recruiting.