The headcount planning problem nobody talks about
Picture a VP of Talent at a mid-market SaaS company. Every quarter, she presents hiring forecasts to the leadership team. How many people will we hire? How long will it take? What will it cost? And every quarter, her forecasts miss the mark by 25 to 35 percent. Not because she lacks experience — she has fifteen years of it — but because she is building her projections on gut feel, anecdotal memory, and rough averages that ignore the variables that actually determine hiring outcomes.
She remembers that the last senior engineer took three months to fill. But she does not account for the fact that it was posted in January (historically the slowest month for engineering talent), that the salary was 12 percent below market at the time, or that the hiring manager rejected seven qualified candidates before the eighth was approved. When she forecasts "60 days" for the next engineering hire, she is essentially guessing — educated guessing, but guessing nonetheless.
This is the problem predictive hiring analytics solves. Instead of building forecasts on memory and instinct, you build them on data: hundreds or thousands of completed hires, each with measurable attributes and outcomes. The difference between forecasting with gut feel and forecasting with data is the same as the difference between weather prediction in 1950 and weather prediction today. The underlying system has not changed, but our ability to model it has improved beyond recognition.
What predictive hiring analytics actually is
Predictive hiring analytics is the practice of using historical recruitment data — combined with statistical modelling or machine learning — to forecast future hiring outcomes. It takes what has already happened in your recruitment process and uses those patterns to predict what will happen next.
The distinction from traditional recruitment metrics is important. Traditional analytics is descriptive: it tells you that your average time-to-fill last quarter was 47 days, that your offer acceptance rate was 74 percent, that LinkedIn produced 38 percent of your hires. This information is useful, but it looks backward. It tells you what happened, not what will happen.
Predictive analytics looks forward. It uses the same historical data but applies mathematical models to generate forecasts: this specific engineering role will likely take 52 to 61 days to fill given current market conditions and your historical patterns. This candidate has a 78 percent probability of accepting an offer at the proposed compensation level. Employee referrals are expected to produce the highest-quality candidates for this particular role category based on 18 months of outcome data.
The practical value is that you stop reacting to recruitment outcomes and start anticipating them. Budget conversations become grounded in evidence. Headcount planning becomes a modelling exercise rather than a negotiation over competing intuitions. Pipeline decisions are informed by probability rather than hope.
According to SHRM research, organisations that adopt data-driven hiring practices see measurable improvements in quality of hire and reductions in time-to-fill — not because the data itself improves hiring, but because it replaces the cognitive biases that degrade human decision-making under uncertainty.
Five types of predictions your hiring data can make
Not all predictions are created equal. Some require modest data and produce reliable results quickly. Others demand years of outcome tracking before the models become trustworthy. Here are the five most practical prediction types for recruitment teams, ordered from easiest to hardest to implement.
1. Time-to-fill forecasting
This is the entry point for most organisations because the data is readily available and the business value is immediately obvious. Time-to-fill forecasting uses your historical hiring data to predict how long a new role will take to fill based on its attributes: department, seniority level, location, salary range, and the time of year it is posted.
A simple model might reveal that your engineering roles take 40 percent longer to fill than sales roles, that Q1 postings fill 15 percent slower than Q3 postings, and that roles priced below the 50th percentile of market rate take twice as long as those at the 75th percentile. With enough data, these patterns become surprisingly precise.
The business application is direct: when a hiring manager submits a requisition, you can immediately provide a data-backed estimate of when the role will be filled rather than defaulting to a generic "30 to 60 days" that means nothing.
2. Candidate success probability
This prediction answers the question every recruiter cares about most: will this candidate actually succeed in the role? Success can be defined in multiple ways — meeting performance goals at 6 months, passing probation, achieving tenure beyond 12 months — but the modelling approach is consistent. You look at the attributes of past candidates who succeeded and those who did not, and you build a model that scores new candidates based on how closely they resemble each group.
Attributes that typically correlate with success include years of relevant experience (not just total experience), the size and stage of previous employers, tenure patterns in past roles, the specificity of skills overlap with the job requirements, and assessment scores where available. Note that AI-based candidate scoring can feed directly into these models by providing structured, consistent evaluation data across all applicants.
The critical caveat is that candidate success models require outcome data — you need to know which past hires actually performed well and which did not. This means you need at least 12 to 18 months of performance tracking for a meaningful dataset.
3. Offer acceptance likelihood
Losing candidates at the offer stage is expensive. You have invested interview hours, assessment time, and pipeline management effort into a candidate who ultimately declines. Offer acceptance prediction estimates the probability that a given candidate will accept an offer based on factors you can observe during the process.
Key predictors include the gap between offered compensation and the candidate's stated expectations, the number of days between final interview and offer extension, the candidate's current employment status (employed candidates accept at different rates than unemployed ones), the number of other processes they have disclosed, and the length of the overall recruitment timeline. Research from LinkedIn Talent Insights consistently shows that time-to-offer is one of the strongest predictors of acceptance.
A practical application: if your model indicates that a candidate has only a 45 percent acceptance probability at the proposed salary, you can adjust the offer, accelerate the timeline, or prepare a backup candidate — all before extending the offer.
4. Source quality prediction
Not all candidate sources perform equally, and their performance varies by role type. LinkedIn might produce your best engineering candidates but mediocre sales candidates. Employee referrals might deliver high acceptance rates but lower diversity. Job boards might generate volume but low conversion to hire.
Source quality prediction goes beyond measuring historical source performance (descriptive analytics) and forecasts which sources will produce the best candidates for a specific upcoming role. The model considers the role type, seniority level, required skills, and location, then predicts which sources will yield the highest ratio of qualified candidates per application. You can learn more about tracking these patterns in our recruitment funnel analytics guide.
The budget implication is significant. Instead of spreading job advertising spend evenly across all channels, you allocate budget based on predicted performance for each specific role. This typically produces 20 to 40 percent better return on recruitment advertising spend.
5. Attrition risk scoring
The most advanced prediction type and the most valuable when done well. Attrition risk scoring estimates the probability that a new hire will leave within a defined period — typically 6 or 12 months. High attrition is not just a retention problem; it is a hiring problem. If you consistently hire candidates who leave within a year, something about your selection process, your offer calibration, or your onboarding experience is systematically failing.
Predictive attrition models examine factors like the candidate's historical job tenure pattern, the alignment between the role offered and the candidate's stated career goals, compensation positioning relative to market, the quality of the interview experience (as measured by candidate feedback scores), and the match between the candidate's working style preferences and the team they will join.
The practical value is intervention: if a candidate scores as high attrition risk, you can address the underlying factors before the hire is made or shortly after, rather than discovering the problem six months later when they resign.
The data you need (and the data quality that matters more)
Every predictive model is only as good as the data it is trained on. For predictive hiring analytics, the data requirements fall into two categories: the volume of records you need and the quality standards those records must meet.
Minimum data volumes
For time-to-fill forecasting, you need a minimum of 50 to 100 completed hires within a specific role category. Fewer than 50 hires means the model cannot distinguish genuine patterns from random noise. For candidate success models, the threshold is higher: 200 or more hires with documented outcome data spanning at least 12 months post-hire. Offer acceptance models require at least 100 offers with recorded accept/decline outcomes and the associated variables. Source quality models need 150 or more hires with tracked source attribution and quality-of-hire metrics.
If these numbers seem large, consider that many organisations have years of ATS data sitting unused. The information exists; it simply has not been structured for analysis. Treegarden's ATS features capture these data points automatically throughout the hiring process, building the dataset you will need for predictive analytics from day one.
Data quality considerations
Volume without quality produces confident but wrong predictions. The most common data quality problems in recruitment data are:
Inconsistent stage definitions. If "phone screen" means a 15-minute qualification call for one recruiter and a 45-minute technical assessment for another, any model that uses stage progression data will produce unreliable results. Standardise your stage definitions across the organisation before attempting predictive modelling.
Missing outcome data. Many organisations track whether a hire was made but not whether the hire succeeded. Without performance and retention data linked to hiring records, you cannot build candidate success or attrition models. Start capturing this data now, even if you are not ready to build models yet.
Survivorship bias. Your hiring data only contains information about candidates who progressed through your process. It tells you nothing about qualified candidates who were screened out early, applied to competitors instead, or never applied at all. Models trained exclusively on your internal data may miss entire categories of successful candidate profiles that your process systematically filters out.
Temporal drift. Hiring patterns change over time. A model trained on data from a tight labour market will produce inaccurate predictions during an economic downturn, and vice versa. Effective models weight recent data more heavily than older data and are retrained periodically as conditions change.
Building prediction models: three approaches
You do not need a PhD in statistics to build useful hiring predictions. The three primary modelling approaches range from accessible to advanced, and each serves different prediction types effectively.
Regression analysis
Regression is the workhorse of predictive analytics and the most accessible starting point. A regression model identifies the relationship between input variables (role attributes, market conditions, candidate characteristics) and a continuous outcome variable (days to fill, compensation required for acceptance).
For time-to-fill prediction, a multiple regression model might include variables like department, seniority level, salary percentile versus market, number of required skills, geographic location, and posting month. The model calculates a coefficient for each variable, telling you exactly how much each factor contributes to the predicted time-to-fill. For example: posting in Q1 adds 8 days, pricing below the 50th percentile adds 14 days, requiring more than 7 mandatory skills adds 11 days.
Regression models are interpretable — you can explain exactly why the model produced a given prediction — and they require relatively modest data volumes to produce useful results. They struggle with non-linear relationships and interactions between variables, but for most recruitment prediction tasks, they perform surprisingly well.
Classification models
Classification models predict categorical outcomes: will the candidate accept or decline? Will the hire succeed or fail? Will the new employee stay beyond 12 months or leave? The output is a probability between 0 and 1 (or 0% and 100%), where higher values indicate greater likelihood of the predicted outcome.
Common classification approaches include logistic regression (the simplest and most interpretable), decision trees (which model sequential decision points and handle non-linear relationships), random forests (which combine many decision trees for improved accuracy), and gradient boosting (which iteratively corrects prediction errors for maximum accuracy at the cost of interpretability).
For offer acceptance prediction, a logistic regression model might reveal that candidates whose offer is extended within 3 days of the final interview are 2.4 times more likely to accept than those who wait 10 or more days. For data-driven hiring decisions, this kind of quantified insight is far more actionable than the vague sense that "we should move faster on offers."
Clustering analysis
Clustering is an unsupervised technique — it does not predict a specific outcome but identifies natural groupings within your data that you might not have recognised. In recruitment, clustering can reveal candidate segments, role archetypes, or hiring patterns that inform your prediction strategy.
For example, clustering your historical hires by attributes and outcomes might reveal three distinct groups: "fast, successful hires" who share specific characteristics (referred by current employees, had fewer than two previous employers, responded to the posting within 48 hours), "slow but successful hires" who have a different profile, and "fast but unsuccessful hires" who share yet another set of attributes.
Once these clusters are identified, you can build targeted prediction models for each group rather than using a single model that averages across fundamentally different hiring patterns. Research published in the Harvard Business Review confirms that segmented hiring models consistently outperform one-size-fits-all approaches.
Predictive analytics use cases at a glance
| Prediction Type | Data Inputs | Model Type | Accuracy Range | Business Impact |
|---|---|---|---|---|
| Time-to-fill forecasting | Role type, seniority, location, salary percentile, posting month, hiring manager history | Multiple regression | ±5–10 days (MAE) | Accurate headcount planning, budget forecasting, recruiter workload balancing |
| Candidate success probability | Skills match score, experience depth, tenure history, assessment results, education alignment | Logistic regression / Random forest | 65–80% (AUC) | Reduced mis-hires, better shortlists, improved quality of hire |
| Offer acceptance likelihood | Compensation gap, time-to-offer, candidate employment status, competing offers, process length | Logistic regression | 70–85% (AUC) | Fewer declined offers, faster closes, better salary calibration |
| Source quality prediction | Historical source-to-hire ratios by role type, cost per source, candidate quality scores per channel | Decision tree / Regression | 60–75% (precision) | 20–40% better ROI on recruitment ad spend |
| Attrition risk scoring | Tenure pattern, compensation vs market, career goal alignment, onboarding feedback, interview quality scores | Gradient boosting / Random forest | 60–75% (AUC) | Reduced first-year turnover, better retention planning, proactive intervention |
Practical applications: where predictions meet real decisions
Headcount planning
The most immediate application of predictive hiring analytics is replacing guesswork in headcount planning with data-backed forecasts. Instead of telling the CFO "we expect to fill 15 engineering roles in Q2," you can say "based on our historical patterns, market conditions, and current pipeline, we predict filling 12 to 14 engineering roles in Q2 with 80 percent confidence, assuming no changes to compensation bands."
The specificity matters. It changes headcount planning from a wish list into a probability-weighted forecast that finance teams can actually build budgets around. When the prediction includes a confidence interval, leadership understands the range of outcomes rather than treating a single number as a guarantee.
Budget forecasting
Recruitment budgets are typically set using last year's spend plus an adjustment factor. Predictive analytics enables bottom-up budget construction: multiply the predicted time-to-fill for each planned role by the predicted cost-per-day of recruiting (recruiter time, advertising spend, tool costs), add predicted agency fees for roles likely to require external support, and sum the totals. The resulting budget is tied to specific hiring plans rather than historical averages.
Pipeline health scoring
At any point during a hiring process, predictive analytics can score the health of your candidate pipeline. Are you on track to make a hire within the predicted timeframe? Is your candidate flow rate sufficient to produce a shortlist? A pipeline health score combines current metrics (applications received, screen pass rates, interviews scheduled) with predicted outcomes to give a green/yellow/red assessment of each open role.
This is where an AI-powered ATS adds particular value. When your applicant tracking system captures structured data about every stage transition, the pipeline health model has real-time data to work with rather than manually updated spreadsheets. Treegarden tracks these pipeline metrics automatically, creating the data foundation that predictive models require.
Interviewer effectiveness
Not all interviewers evaluate candidates with equal accuracy. Some interviewers consistently approve candidates who go on to succeed. Others have high approval rates but poor downstream outcomes — the candidates they recommend are more likely to fail probation or leave within a year. Predictive analytics can identify these patterns by correlating interviewer decisions with subsequent hire outcomes.
The application is not punitive but instructive: interviewers whose predictions of candidate success are consistently inaccurate may benefit from calibration training, adjusted interview formats, or pairing with a more accurate co-interviewer. Over time, improving interviewer effectiveness improves the accuracy of your entire hiring process.
Implementation: a realistic six-step approach
Implementing predictive hiring analytics is a data maturity journey, not a software installation. Here is a practical sequence that accounts for where most recruitment teams actually start.
Step 1: Audit your existing data. Before building any models, understand what data you currently have. Export your ATS data and assess it for the variables listed above. How many completed hires do you have? Is stage progression data consistent? Do you have outcome data linked to hiring records? Most organisations discover they have more data than they expected but with significant quality gaps that need addressing.
Step 2: Standardise your data capture. Close the quality gaps identified in step one. Standardise stage definitions, implement mandatory fields for key variables (source, salary offered, days between stages), and begin capturing outcome data if you are not already doing so. This step often takes 2 to 3 months to implement properly across the organisation.
Step 3: Start with descriptive analytics. Before predicting the future, understand the past. Build dashboards that show your historical time-to-fill by role category, offer acceptance rates by compensation band, source performance by role type, and stage conversion rates by recruiter. These descriptive views will reveal the patterns your predictive models will later formalise. Our hiring analytics guide covers this foundational step in detail.
Step 4: Build your first prediction model. Start with time-to-fill forecasting — it requires the least outcome data and produces immediately useful results. Use a simple multiple regression approach with 3 to 5 input variables. Compare the model's predictions against actual outcomes for the most recent 20 percent of your data (which you held out of the training set). If the model's predictions are consistently within 10 days of actual outcomes, you have a useful tool.
Step 5: Expand to classification models. Once you have 12 or more months of outcome data, build your first candidate success or offer acceptance model. Start simple — logistic regression with 5 to 7 variables — and increase complexity only if the simple model's accuracy is insufficient.
Step 6: Operationalise and iterate. Integrate predictions into your daily recruitment workflow. This might mean displaying predicted time-to-fill on every new requisition, showing offer acceptance probability on candidate profiles, or flagging pipeline health warnings on your recruitment dashboard. Retrain models quarterly as new data accumulates and market conditions shift.
Analytics-Ready from Day One
Treegarden captures structured hiring data at every pipeline stage — source attribution, stage timestamps, candidate scores, and outcome tracking. This means the predictive analytics dataset builds automatically as you use the platform, so when you are ready to model, the data is already there. Start building your dataset today.
Three pitfalls that destroy prediction accuracy
Predictive analytics is powerful, but it is also easy to get wrong. These three pitfalls account for the majority of failed implementations.
Pitfall 1: Overfitting
Overfitting occurs when a model learns the noise and idiosyncrasies in your historical data rather than the genuine underlying patterns. An overfitted model will show impressive accuracy when tested against past data but perform poorly on new, unseen cases.
The classic sign of overfitting is a model with too many variables relative to the number of data points. If you have 80 completed hires and your model uses 25 input variables, it is almost certainly overfitted — it has enough degrees of freedom to memorise the training data rather than learning from it.
Prevention strategies: use a holdout validation set (train on 70 to 80 percent of data, test on the rest), apply regularisation techniques that penalise model complexity, and start with simple models before adding complexity. A simple model that is right 70 percent of the time on new data is more valuable than a complex model that is right 95 percent of the time on old data.
Pitfall 2: Bias in training data
This is the most ethically significant pitfall. If your historical hiring data reflects biased decisions — and most hiring data does to some degree — a prediction model trained on that data will reproduce and potentially amplify those biases.
Consider a candidate success model trained on five years of hiring data. If during those five years, hiring managers systematically gave lower performance ratings to employees from certain backgrounds, the model will learn that association and apply it to future predictions. The model is not "biased" in the human sense; it is faithfully reflecting the patterns in the data it was given. But the result is the same: biased predictions that perpetuate historical inequities.
Mitigation requires careful feature selection (never include protected characteristics or their proxies as model inputs), regular bias audits that test model predictions across demographic groups, and maintaining human oversight of all hiring decisions. A prediction model should inform decisions, never make them autonomously. This aligns with the principles of responsible AI in recruitment — a topic covered in the SHRM guidelines on AI and hiring bias prevention.
Pitfall 3: False precision
A model that predicts a candidate has a 73.4 percent chance of succeeding in a role is communicating false precision. The underlying data and methodology cannot support that level of specificity. Reporting "approximately 70 to 75 percent" is more honest and more useful, because it communicates the inherent uncertainty in the prediction.
False precision is dangerous because it encourages over-reliance on model outputs. When a prediction looks precise, decision-makers treat it as fact rather than as an informed estimate. This leads to under-weighting other important information — interview impressions, reference checks, cultural fit assessments — that the model cannot capture.
Best practice: always report predictions with confidence intervals or ranges, clearly label predictions as estimates rather than certainties, and educate stakeholders about what the numbers mean and what they do not mean.
Measuring prediction accuracy: are your forecasts worth trusting?
A prediction model is only valuable if its predictions are more accurate than the method it replaces. Here is how to measure whether your models are actually working.
For continuous predictions (time-to-fill, cost)
Mean Absolute Error (MAE) is the most intuitive metric. It calculates the average absolute difference between predicted and actual values. If your time-to-fill model has an MAE of 7 days, it means predictions are off by an average of 7 days in either direction. For most recruitment teams, an MAE under 10 days represents a substantial improvement over gut-feel estimates.
Root Mean Squared Error (RMSE) penalises large errors more heavily than small ones. If occasional predictions are wildly off — predicting 30 days for a role that takes 90 — RMSE will flag that more aggressively than MAE. Both metrics should be tracked together.
For classification predictions (accept/decline, success/failure)
AUC (Area Under the Curve) measures the model's ability to distinguish between positive and negative outcomes. An AUC of 0.5 means the model is no better than random guessing. An AUC of 0.7 indicates moderate predictive power. An AUC of 0.8 or above indicates strong predictive power. For most recruitment classification tasks, an AUC between 0.65 and 0.80 is realistic and useful.
Precision and recall measure different aspects of accuracy. Precision answers "of the candidates the model predicted would succeed, how many actually did?" Recall answers "of the candidates who actually succeeded, how many did the model correctly identify?" The balance between these two depends on your priorities: if false positives are costly (promoting a candidate who will fail), optimise for precision. If false negatives are costly (screening out a candidate who would have succeeded), optimise for recall.
The baseline comparison
The most important accuracy test is not against a statistical benchmark but against your current forecasting method. If your VP of Talent's gut-feel predictions for time-to-fill have an MAE of 18 days, and your regression model achieves an MAE of 9 days, the model has halved the prediction error. That is the number that matters for business adoption — not whether the model achieves some abstract standard of statistical excellence.
Where to start if you have never done this before
If predictive hiring analytics feels overwhelming, start with the smallest useful step: export your last 12 months of completed hires from your ATS. For each hire, record the role category, the number of days from posting to accepted offer, the source, and the seniority level. Put it in a spreadsheet.
Now calculate your average time-to-fill by role category and by source. You have just done descriptive analytics. Next, look at whether certain combinations (engineering roles sourced from referrals versus job boards) have consistently different outcomes. You are now thinking predictively.
The gap between that spreadsheet exercise and a formal regression model is smaller than most people think. The hard part is not the maths — it is the discipline of capturing consistent, quality data over time. An ATS that structures this data automatically, like Treegarden, removes the biggest barrier to entry by making data capture a byproduct of your normal recruitment workflow rather than an additional administrative burden.
Frequently asked questions
What is predictive hiring analytics?
Predictive hiring analytics is the practice of using historical recruitment data and statistical or machine learning models to forecast future hiring outcomes. This includes predicting time-to-fill for open roles, estimating the probability that a candidate will succeed in a position, forecasting offer acceptance rates, evaluating source channel quality, and identifying attrition risk before a hire is made. The approach moves hiring decisions from gut-feel guesswork to evidence-based forecasting.
How much historical data do I need to build useful hiring predictions?
For basic time-to-fill predictions, you need at least 50 to 100 completed hires in a specific role category. For candidate success models, you need 200 or more hires with outcome data spanning at least 12 months. Offer acceptance models require 100 or more offers with accept/decline records. The more data you have, the more accurate the predictions become, but even modest datasets can produce directionally useful forecasts when the data quality is high.
What types of predictions can hiring analytics actually make?
The most common predictions include: time-to-fill forecasting (how long a role will take to fill based on historical patterns), candidate success probability (likelihood a hire will meet performance benchmarks), offer acceptance likelihood (whether a candidate will accept an offer based on compensation, timing, and other factors), source quality prediction (which channels will produce the best candidates for a specific role), and attrition risk scoring (probability a new hire will leave within 12 months).
Can predictive analytics introduce bias into hiring?
Yes, and this is one of the most significant risks. If your historical hiring data reflects biased decisions — for example, if certain demographic groups were systematically undervalued in past hiring — a prediction model trained on that data will reproduce and potentially amplify those biases. Mitigating this requires careful feature selection (excluding protected characteristics and their proxies), regular bias audits of model outputs, and maintaining human oversight of all hiring decisions.
What is overfitting and why does it matter in recruitment predictions?
Overfitting occurs when a prediction model learns the noise and quirks in your historical data rather than the genuine underlying patterns. An overfitted model will perform extremely well on past data but fail badly on new, unseen cases. In recruitment, this might mean a model that perfectly predicts past hiring outcomes but gives wildly inaccurate forecasts for new roles. You prevent overfitting by using separate training and validation datasets, keeping models simple, and testing predictions against real outcomes before relying on them.
How do I measure whether my hiring predictions are actually accurate?
Use a holdout validation approach: train your model on 70 to 80 percent of your historical data, then test its predictions against the remaining 20 to 30 percent that it has never seen. For time-to-fill predictions, measure mean absolute error (the average difference between predicted and actual days). For classification predictions like offer acceptance, track precision, recall, and the AUC score. Most importantly, compare your model's accuracy against your current method — even a model that is right 65 percent of the time may be a significant improvement over intuition-based forecasting that is right 40 percent of the time.
Do I need a data science team to implement predictive hiring analytics?
Not necessarily. Modern ATS platforms like Treegarden are building predictive analytics directly into their reporting and AI features, which means the statistical modelling happens behind the scenes. For basic predictions like time-to-fill forecasting and source quality analysis, the tools do the heavy lifting. However, if you want to build custom candidate success models or integrate hiring predictions with broader workforce planning, a data analyst or data scientist will add significant value to the implementation.
What is the difference between descriptive analytics and predictive analytics in recruitment?
Descriptive analytics tells you what happened: your average time-to-fill last quarter was 42 days, your offer acceptance rate was 78 percent, your top source was LinkedIn. Predictive analytics tells you what will likely happen: this engineering role will probably take 55 days to fill, this candidate has a 72 percent chance of accepting the offer, employee referrals will produce the best candidates for this specific position. Descriptive analytics looks backward at historical data; predictive analytics looks forward using that same data to forecast future outcomes.