How AI Resume Screening Actually Works
Understanding the mechanics of AI resume screening is essential for evaluating it critically. Most systems fall into one of two broad categories:
Rules-based screening applies hard criteria — required keywords, minimum years of experience, specific credentials — to filter applications. These systems are transparent but brittle; they miss qualified candidates whose resumes don't use the exact expected language and cannot capture context or potential.
Machine learning-based screening trains models on historical hiring data — past resumes, recruiter scores, interview progression, and hire decisions — to predict which new candidates resemble successful past hires. These systems can identify non-obvious signals, but they inherit every bias present in historical hiring decisions. If your past hires skewed toward candidates from certain schools, backgrounds, or demographics, the model will perpetuate that skew.
A third category — large language model (LLM) based screening — has emerged rapidly. These systems can process natural language context more flexibly, but their decision logic is even less transparent and their behavior on novel inputs less predictable.
The Accuracy Disclosure Problem
No major AI resume screening vendor publishes standardized, independently verified accuracy benchmarks that allow apples-to-apples comparison. Vendors quote precision rates — the share of top-scored candidates who were genuinely qualified — but rarely disclose recall rates — the share of genuinely qualified candidates who scored highly. A system with 95% precision and 40% recall is quietly rejecting most of your qualified candidate pool. Always ask for both metrics before any procurement decision.
The False Negative Problem HR Doesn't Talk About
The most significant accuracy failure in AI resume screening is the false negative: the qualified candidate the system rejects who never enters your pipeline. False negatives are invisible — you never see the people you didn't advance — which is why they persist undetected for years.
Research has identified consistent false negative patterns in AI screening systems trained on historical data:
- Career changers: Professionals transitioning from adjacent industries often have highly relevant skills packaged in non-standard career narratives. Systems trained on conventional career paths systematically underscore these candidates.
- Non-traditional education: Candidates with bootcamp credentials, self-taught skills, or international degrees frequently score lower than their actual capability warrants, because training data over-represents traditional four-year US degree holders.
- Resume format variation: Candidates who format resumes in non-standard ways — functional resumes, skills-first formats, non-English-origin naming conventions — trigger parsing errors that degrade their scores.
- Women in technical fields: Multiple published studies have found that AI systems trained on historical tech hiring data systematically underprioritize female candidates, reproducing the historical gender imbalance in the training data.
The Legal Landscape: NYC Law 144 and EEOC Guidance
The regulatory environment for AI hiring tools is evolving rapidly in the US, and employers — not just vendors — bear legal responsibility for outcomes:
- New York City Local Law 144: Effective July 2023, NYC employers using automated employment decision tools must conduct annual bias audits by independent auditors and provide advance notice to candidates. Violations carry civil penalties. This is the most specific municipal AI hiring regulation in the US and is setting a template for other jurisdictions.
- EEOC Guidance (2023): The EEOC has made clear that Title VII's disparate impact standard applies to AI hiring tools. Employers are responsible for ensuring their screening systems do not produce adverse impact on protected classes, regardless of whether the tool was built by a third-party vendor.
- Illinois AI Video Interview Act: Requires employers using AI to analyze video interviews to notify candidates, explain how AI is used, and obtain consent.
- State-level expansion: Maryland, New Jersey, and Washington have introduced or are considering similar AI hiring transparency requirements. HR teams should monitor these developments proactively.
Employer Liability Does Not Transfer to Vendors
A common misconception: that using a third-party AI vendor shifts legal responsibility to the vendor. Under current EEOC interpretation, this is incorrect. The employer who deploys the tool is legally responsible for its employment outcomes. Vendor contracts that indemnify employers against discrimination claims are not reliable protection. Before deployment, require vendors to provide bias audit results, and conduct your own adverse impact analysis on a sample of screening decisions.
Questions Every HR Team Should Ask AI Screening Vendors
Before purchasing or renewing any AI screening tool, HR should require clear answers to:
- What was the training dataset — size, recency, demographic composition, and industry coverage?
- Has the system undergone independent bias auditing? Can we see the full audit report, not a summary?
- What is the adverse impact ratio for the screening output across gender, race, and age? (Selection rate for protected groups divided by selection rate for the most-selected group — EEOC's 4/5ths rule applies.)
- What human review is required before a rejection decision is finalized? Can the system make rejections autonomously?
- How are screening criteria updated when our job requirements change? Is retraining required?
- What data from our candidate pipeline is used to retrain the model, and how is that consent obtained?
AI as Prioritization, Not Gatekeeping
Treegarden integrates AI to help recruiters prioritize and organize candidates — surfacing relevant matches and flagging completeness — while keeping human reviewers in the decision loop at every stage. This approach delivers the efficiency benefits of AI screening without the black-box rejection risk that creates compliance exposure.
Using AI Screening Responsibly: A Practical Framework
HR teams that use AI screening responsibly treat it as a prioritization tool, not a gatekeeper:
- Never use AI as the sole rejection mechanism. AI scores should surface candidates for faster human review, not automatically close applications. Every rejection decision should have a human in the loop who can override the model.
- Audit your screened-out population quarterly. Pull a random sample of AI-rejected applications and have a recruiter assess whether the rejections were appropriate. Look for demographic patterns in the rejected pool.
- Recalibrate criteria regularly. Job requirements evolve, and a model trained 18 months ago may be optimizing for the wrong profile. Review and update screening criteria with each major hiring wave.
- Document your human review process. When a retaliation or discrimination claim arises, your ability to demonstrate that human judgment was applied at every rejection decision is your primary defense.
Frequently Asked Questions
How accurate is AI resume screening?
AI resume screening accuracy varies significantly by vendor, role type, and dataset quality. Well-implemented systems in structured roles with large historical data sets can reach 80–90% precision in identifying technically qualified candidates. However, accuracy degrades for novel roles, unconventional career paths, and candidates whose resumes don't match the format the model was trained on. No vendor publishes standardized accuracy benchmarks, making comparison nearly impossible without independent testing.
Can AI resume screening create legal liability for employers?
Yes. Under EEOC guidance and emerging state laws — particularly New York City Local Law 144 — employers that use automated employment decision tools bear legal responsibility for ensuring those tools do not produce disparate impact on protected classes. An AI screening system that systematically rejects a higher proportion of candidates from a protected group can expose the employer to discrimination claims, even if no bias was intentional.
What is a false negative in AI resume screening?
A false negative occurs when an AI screening system rejects a candidate who would have been a qualified hire. False negatives are harder to detect than false positives because the rejected candidate never enters the pipeline. Research suggests AI screening systems trained on past hiring data tend to systematically under-select candidates whose profiles differ from the historical majority — affecting women in tech, career changers, and candidates from non-traditional backgrounds.
What questions should HR ask AI screening vendors?
Key questions include: What was the training dataset, and how recent is it? Has the system undergone independent bias auditing, and can you share the results? What is the false negative rate for protected class candidates specifically? Does the system comply with New York City Local Law 144 and EEOC guidance? What level of human review is required before rejection decisions are finalized? How are screening criteria updated when your job requirements change?
How can HR reduce bias risk when using AI screening?
Require vendors to provide bias audit results before purchasing. Conduct your own adverse impact analysis on screening outputs by protected class. Implement mandatory human review before any rejection is final. Regularly audit your screened-out population for unexplained demographic patterns. Avoid using AI screening as the sole determinant of advancement — treat it as a prioritization tool rather than a binary pass-fail gate.