Two approaches to the same problem

Every recruiter's core challenge is the same: given a stack of applicants, identify who deserves a closer look. Manual CV review and AI match scoring are two fundamentally different approaches to solving this problem, and understanding where each excels and where each falls short is essential for building an effective screening process.

Manual review relies on human judgement. The recruiter reads each CV, draws on their experience to assess fit, and makes a gut-informed decision about whether the candidate warrants further evaluation. This approach leverages the recruiter's contextual understanding, intuition about cultural fit, and ability to read between the lines of a CV.

AI match scoring relies on systematic analysis. The system evaluates each CV against defined criteria — skills, experience, education, keyword relevance — and produces a numerical score. This approach leverages consistency, speed, and the ability to apply the same criteria identically to every application regardless of when it arrived or how many CVs came before it.

Neither approach is perfect in isolation. The question is not which one to use, but how to combine them effectively.

Time: the most obvious difference

The time comparison is stark. Consider a role that attracts 150 applications — a fairly typical number for a mid-level professional position at a well-known company.

Manual review: A thorough manual review takes 3-5 minutes per CV. At the lower end, reviewing all 150 candidates takes 7.5 hours. At the upper end, 12.5 hours. Realistically, no recruiter reviews 150 CVs at a consistent pace. After 60-90 minutes, attention degrades. The review stretches across multiple sessions, often across multiple days. Total elapsed time from first review to completed shortlist: typically 2-4 working days.

AI match scoring: Scoring 150 candidates takes minutes. The recruiter then spends detailed review time on the top 25-30 candidates (those with green and blue score badges), investing 5-7 minutes per candidate for a thorough assessment. Total review time: 2-3.5 hours. Elapsed time to completed shortlist: typically half a day.

The net saving is 5-9 hours per role. For a recruiter managing eight open roles simultaneously, that is 40-72 hours recovered per hiring cycle — effectively an entire working week.

Time Savings Scale With Volume

The time savings from AI scoring increase as application volume grows. For a role with 50 applications, manual review might take 3-4 hours. For a role with 300 applications, it takes 15-25 hours. AI scoring time remains roughly constant regardless of volume — the recruiter's review time scales only with the number of top candidates, not total applicants.

Consistency: where AI has a structural advantage

Consistency is perhaps the dimension where AI scoring has the clearest advantage over manual review. When a human reviews CVs, their assessment is influenced by factors that have nothing to do with the candidate:

Fatigue effects. The 100th CV in a review session does not receive the same quality of attention as the 10th. Research on sequential decision-making consistently shows that decision quality degrades over extended sessions. Candidates whose CVs are reviewed later in a batch receive systematically less thorough evaluation.

Order effects. A candidate reviewed immediately after a very strong applicant tends to be assessed more harshly than the same candidate would be if reviewed after a weak one. This contrast effect is well-documented in hiring research and is completely unrelated to the candidate's actual qualifications.

Inter-reviewer variability. When multiple recruiters review the same candidates, their assessments often diverge significantly. One reviewer might prioritise technical depth while another weights communication skills more heavily. Without explicit scoring criteria, each reviewer is effectively using a different rubric.

Mood and context. A recruiter's assessment can be influenced by how their day is going, whether they recently had a positive or negative interaction, and even factors as mundane as whether they have eaten recently. These influences are unconscious and unmeasurable, but they affect decisions.

AI scoring eliminates all of these variables. The 150th CV receives exactly the same quality of analysis as the 1st. The score is not influenced by which CVs came before it. Two different recruiters initiating scoring on the same candidate pool will get identical results. The assessment does not vary based on time of day or the scorer's emotional state.

This does not mean AI scoring is more accurate — that depends on how well the scoring criteria match what actually matters for the role. But it is more consistent, which is a prerequisite for fairness.

Bias: different risks, different solutions

Both manual review and AI scoring carry bias risks, but they are different types of bias with different characteristics.

Human bias in manual review is well-documented. Name-based bias (studies consistently show that identical CVs receive different callback rates depending on the candidate's name), age bias (inferred from graduation dates and career length), education prestige bias (over-weighting the institution rather than the qualification), and affinity bias (favouring candidates who resemble the reviewer) are all present in manual screening to varying degrees.

AI bias typically reflects the criteria it is configured to use. If the scoring weights overemphasise keywords that correlate with a particular demographic (for example, specific university names or industry jargon that varies by region), the scores may inadvertently disadvantage certain groups. However, AI bias has a crucial characteristic that human bias lacks: it is systematic and auditable.

Because AI scoring is consistent, you can test for bias. Run the scoring on a diverse candidate pool and check whether scores correlate with protected characteristics. If they do, you can identify which scoring dimension is causing the disparity and adjust it. You cannot do this with manual review because the bias is inconsistent, unconscious and distributed across multiple reviewers making thousands of micro-decisions.

Consistent Scoring in Treegarden

Treegarden's Edera AI applies the same scoring criteria to every candidate, eliminating fatigue effects, order bias and inter-reviewer inconsistency. Colour-coded badges (green, blue, yellow, red) make prioritisation instant. All scoring is user-initiated and GDPR-compliant. Try it free.

Shortlist quality: where the methods converge

The ultimate measure of any screening method is the quality of the resulting shortlist. Do the candidates who make it through the screening actually perform well in interviews and, ultimately, in the job?

Manual review produces shortlists that are shaped by the reviewer's experience and intuition. An experienced recruiter who deeply understands the role, the team and the company culture can produce excellent shortlists through manual review alone. However, the quality depends heavily on the individual recruiter's expertise and the time they have available. An overloaded recruiter working through a high-volume role will produce a worse shortlist than the same recruiter would with fewer applications and more time.

AI scoring produces shortlists that are shaped by the configured criteria. When the weights accurately reflect what the role requires, the resulting shortlist concentrates the strongest technical matches at the top. However, AI scoring may miss candidates whose value comes from factors that are hard to quantify: an unusual career path that brings diverse perspective, a personal connection to the company's mission, or soft skills that are evident in how the CV is written but not in the keywords it contains.

The best shortlists come from combining both methods. AI scoring provides the initial prioritisation, ensuring that no strong candidate is overlooked due to reviewer fatigue or volume. Human review then evaluates the top-scoring candidates with the contextual judgement that AI cannot replicate. This two-layer approach consistently outperforms either method used alone.

When manual review still wins

AI scoring is not superior in every situation. There are specific scenarios where manual review produces better results:

Very senior hires. When hiring for a C-suite or VP-level position, the applicant pool is small (often under 20), and each candidate's unique trajectory and narrative matters more than keyword matching. Manual deep-dive review of each candidate is both feasible and appropriate.

Creative roles. For positions where the CV itself is part of the assessment — design portfolios, creative writing, or roles where communication style matters — AI scoring of structured criteria misses the most important signal.

Culture-critical hires. When team dynamics or cultural contribution is the primary concern rather than technical qualification, human judgement about how a candidate's background and personality might integrate with the existing team cannot be replicated by scoring.

Extremely niche roles. When a role requires a combination of skills so specific that only a handful of candidates globally would qualify, AI scoring adds little value. The recruiter already knows what to look for and the pool is small enough to review manually.

When AI scoring is clearly better

Conversely, there are scenarios where AI scoring dramatically outperforms manual review:

High-volume roles. Any role attracting more than 80-100 applications is a strong candidate for AI scoring. The time savings are significant and the consistency advantage is most valuable when the volume exceeds what a human can reasonably process with sustained attention.

Multiple similar roles. When recruiting for the same position across multiple locations or teams, AI scoring ensures that candidates for all instances of the role are assessed against the same criteria — preventing the inconsistency that arises when different recruiters screen for what is nominally the same job.

Technical roles with clear requirements. Positions where the required skills, tools and certifications can be clearly specified produce the best AI scoring results. The more objectively the requirements can be defined, the more effective AI scoring becomes.

Speed-critical hiring. When time-to-hire is a competitive factor — particularly in tight markets where strong candidates accept offers within days — the speed advantage of AI scoring directly translates to hiring outcomes.

The optimal approach: AI first, human second

The evidence points clearly toward a combined approach. Use AI scoring as the first layer to prioritise the candidate pool, then apply human review to the top candidates for nuanced assessment. This workflow captures the strengths of both methods:

Step 1: Configure scoring weights based on what matters for this specific role.

Step 2: Run AI scoring across the full candidate pool.

Step 3: Review green-badge candidates (70%+) in detail. These are your most likely interviews.

Step 4: Scan blue-badge candidates (40-69%) for hidden gems — candidates with transferable skills or unusual backgrounds that the AI may have under-scored.

Step 5: Make interview decisions based on human assessment of the AI-prioritised shortlist.

This approach typically reduces screening time by 60-75% while maintaining or improving shortlist quality compared to pure manual review.

Frequently asked questions

Is AI scoring more accurate than manual CV review?

AI scoring is more consistent than manual review. It produces the same score for the same CV every time. Whether it is more accurate depends on how well the scoring weights are configured for the specific role.

Does AI candidate scoring eliminate bias in recruitment?

AI scoring reduces certain types of human bias — fatigue bias, order effects, and inter-reviewer inconsistency. However, AI can introduce its own biases based on how it is configured. The key advantage is that AI bias is systematic and auditable, while human bias is inconsistent and much harder to detect.

Should AI scoring completely replace manual CV review?

No. The most effective approach uses AI scoring as a first-pass prioritisation layer, followed by human review of the top candidates. AI identifies who to look at first; the recruiter decides who to interview.

How do I know if AI scoring is working correctly for my roles?

Compare your AI-generated shortlists to your interview outcomes. If the candidates you advance are consistently the ones the AI scored highest, the scoring is well-calibrated. If you frequently find strong candidates among low-scoring applicants, adjust your weight configuration.