The CRM contains the data that produces the forecast. The forecast is always wrong. The logical inference — that the data in the CRM might be the problem — should be obvious. It is not, apparently. Because the most common response to a bad forecast is to have a pipeline review meeting where sales managers pressure reps to update their close dates. This fixes nothing. The close dates were wrong before the meeting and they are wrong after it, just updated to reflect a different optimistic assumption.
CRM data quality is not a technology problem. Salesforce does not cause bad data. HubSpot does not cause bad data. The CRM is a container. What you put in it is what you get out of it. The cause of poor data quality is human: nobody enforces entry standards, there are no consequences for incomplete records, field design makes accurate entry difficult, and reps have learned that the system is used against them in pipeline reviews so they manage the data to manage the conversation rather than to reflect reality. Fix those four things and the data quality problem largely solves itself. Leave them unfixed and no amount of data hygiene campaigns will help.
Why CRM Data Degrades
No Enforcement
In the vast majority of sales organisations, CRM data entry is nominally mandatory and practically optional. Required fields exist. Reps populate them with whatever gets the record past the validation. Close date: last day of the quarter. Deal value: the number from the first conversation that was never confirmed. Stage: whatever avoids a follow-up question in the next one-on-one. When there is no downstream consequence for inaccurate data, the data will be as accurate as reps find convenient. That is usually not very accurate.
Enforcement requires a closed loop between data quality and process. Stage advancement should require specific fields to be populated with evidence-based values. Opportunities in late stages without documented next steps should trigger automatic alerts. Forecast inclusion should be conditional on data completeness. When the process depends on data quality, data quality improves because reps have a direct incentive to maintain it. When the process proceeds regardless of data quality, it does not.
No Consequence for Bad Data
Related but distinct: enforcement is about the system. Consequence is about management behaviour. If a rep's pipeline review goes ahead regardless of whether their opportunity data is complete, the rep learns that incomplete data has no cost. If the pipeline review starts with a discussion of data quality issues and the rep is asked to update records before the substantive conversation can proceed, they learn the opposite. Management behaviour is the primary driver of rep data behaviour. The CRM reflects the standards that managers actually enforce, not the standards the company nominally espouses.
Field Design That Makes Accuracy Hard
Some CRM implementations are designed to capture everything imaginable, which means reps are faced with screens full of fields they do not understand, do not have the information to complete, or both. The result is that everything gets filled with placeholder values or left blank. Good field design is minimalist. It captures the information that is actually used in decisions — the fields that feed the forecast, inform the territory plan, and enable segmentation analysis. Every additional field beyond that reduces the accuracy of the fields that matter because attention and effort are finite. A CRM with 20 required fields per opportunity will have worse data quality on the five fields that actually matter than a CRM with eight required fields designed around those five.
The Adversarial Dynamic
In organisations where CRM data is primarily used to manage performance — to identify underperformers, to pressure-test close dates, to justify pipeline reviews — reps learn to manage the data to manage the conversation. They move deals forward when they anticipate a push, move them back when they want to buy time, inflate values when they need to look credible, and deflate them when they do not want the scrutiny. The CRM becomes a negotiating tool rather than a record of commercial reality. This dynamic is almost impossible to reverse without changing how leadership uses the data. If the data is used punitively, it will be managed punitively.
The Data Fields That Matter Most for Forecasting
Not all CRM fields are equal. For forecasting purposes, six fields determine the quality of the output more than everything else combined.
Close date — the projected date of deal close. This field is the most commonly falsified in CRM systems. It should reflect the buyer's stated timeline, not the rep's quota period. A close date that has slipped three times is not a close date. It is a hope with a label.
Deal value — the projected contract value. This should reflect confirmed budget, not the rep's aspirational deal size. The gap between unconfirmed and confirmed deal value is usually significant and systematically optimistic.
Stage — where the deal is in the sales process. Stage definitions that are not tied to specific buyer actions are not stage definitions. They are rep sentiment fields. "Proposal sent" is a rep action. "Proposal reviewed and questions raised" is a buyer action and a meaningfully better stage criterion.
Next step — the specific, dated next action agreed with the buyer. An opportunity without a next step is not in a sales process. It is in a waiting state. Deals in waiting states do not close on the forecast date.
Stakeholder map — who is involved in the decision. In enterprise deals, absence of economic buyer contact is the single strongest predictor of deal loss or prolonged cycle. This field is frequently empty because it requires reps to admit they have not reached the right people yet.
Competitive situation — whether there are known competitors in the deal. This affects close probability and pricing behaviour. Its absence in the CRM record typically means the rep either does not know or does not want to record it. Both are problems. For a view on how competitive intelligence from CRM data connects to broader win/loss patterns, see Win/Loss Analysis: Run One That Actually Changes Behaviour.
THE FRAMEWORK
The full interrogation framework is Dispatch #001 — Pipeline & Forecast Framework. 38 questions across four sections that expose where your pipeline data is unreliable and why your forecast never lands. $97. Instant download.
See the full framework →How to Audit CRM Data Quality
A CRM data quality audit starts with a completeness assessment: for each required field across all active pipeline, what percentage of records are populated? Not populated with any value — populated with a plausible value. Close dates in the past are not populated. Deal values of zero are not populated. Stage fields that have not changed in 90 days on a deal that is supposedly progressing are not accurate.
Run completeness by field, by rep, and by manager. The distribution is usually illuminating. Some reps maintain near-perfect records. Others have persistent gaps. The pattern by manager tells you more than the pattern by rep — because data quality is a management standard, and managers who enforce it tend to produce teams with clean data. Managers who do not enforce it tend to produce teams with similar gaps regardless of individual rep variability.
Beyond completeness, audit consistency. Do stage definitions mean the same thing across reps? Pull a sample of deals at each stage and check whether the buyer actions that should have occurred to justify that stage placement actually occurred. In most organisations, stage definitions are interpreted loosely and consistently optimistically. A deal is in "Proposal" because the rep sent a proposal, not because the buyer has engaged with it. The gap between nominal stage and actual stage is where forecast error lives.
Finally, audit the velocity of records. Deals that have been in the same stage for longer than the median sales cycle at that stage are stalled. They may appear in the forecast. They should not. Stalled deals close at a fraction of the rate of progressing deals and at a fraction of the projected value. A pipeline full of stalled deals looks larger than it is and will produce a forecast that is consistently optimistic. The relationship between data accuracy and pipeline coverage ratios is explored at Pipeline Coverage Ratio: What It Hides From You.
The Governance Process That Maintains Quality Over Time
Auditing data quality once is useful. Maintaining it over time requires governance — the combination of rules, enforcement mechanisms, and accountability that makes data quality the path of least resistance rather than an additional burden.
Effective governance has four components. First, field-level validation rules that prevent advancement to the next stage without required fields being populated with non-placeholder values. This is the technical layer. Second, manager review cadences that include data quality checks as a standing agenda item — not a separate initiative, but embedded in the existing pipeline review process. Third, data quality reporting that is visible to sales leaders on the same dashboard as pipeline and forecast data. When data quality is invisible, it is ignored. When it sits next to the forecast number, the connection between them becomes obvious. Fourth, a clear owner — someone in sales ops or RevOps who is accountable for the overall quality of CRM data and who has the authority to flag issues to management without it being taken as a personal criticism of specific reps.
One governance mechanism that is underused is peer visibility. In organisations where reps can see each other's pipeline data quality scores, the social pressure to maintain clean records is significant. Top performers generally have clean CRM data. That correlation is not accidental, and making it visible has motivational value without requiring a formal enforcement mechanism. The broader metrics framework for tracking data governance health sits within the RevOps metrics layer — the full set of indicators is at RevOps Metrics: The 12 Numbers That Actually Matter.
The Connection Between CRM Data Quality and Forecast Accuracy
The mechanism is direct. Forecast accuracy is bounded by the accuracy of the underlying opportunity data. If close dates are systematically optimistic by three weeks, the forecast will be systematically optimistic by a comparable amount. If deal values are inflated by 20% on average at the proposal stage, the forecast will overstate expected revenue by 20% until late-stage correction occurs. If stage definitions are inconsistently applied, conversion rate assumptions — which are derived from historical stage data — will be wrong, and the statistical adjustments the forecasting model applies will compound the error rather than correct it.
You cannot build a reliable forecast on unreliable data. No forecasting methodology, no matter how sophisticated, compensates for systematic inaccuracy in the underlying records. The organisations that forecast accurately are not the ones with the most sophisticated models. They are the ones with the most accurate opportunity data, combined with a forecasting process that applies consistent logic to that data. Both are necessary. Clean data with a poor forecasting process produces an overconfident forecast. Good forecasting methodology applied to dirty data produces a sophisticated-looking wrong number. The specific ways forecast methodology breaks down when built on poor data are examined in detail at Why the Forecast Was Never Real.
The sales velocity calculation illustrates this concretely. Velocity depends on four inputs: number of opportunities, average deal value, win rate, and sales cycle length. Every one of those inputs is derived from CRM data. If the opportunity count includes stalled deals, velocity is overstated. If average deal value is based on unconfirmed figures, velocity is overstated. If win rate is calculated on the total opportunity population including deals that were never real, velocity is overstated. The resulting velocity number looks precise and is systematically wrong. The way this plays out in pipeline management is laid out at Sales Velocity: The Formula That Predicts Revenue.
The forecast does not lie. The data lies. The forecast just faithfully reports the lies you entered into the CRM last Tuesday.
Fix the data before you fix the forecast model. Audit completeness, enforce stage discipline, change how managers use pipeline data, and design fields around the decisions the data needs to support. These are not exciting interventions. They do not generate conference talks or vendor demos. But they are the prerequisite for every other commercial analytics initiative your operations team is trying to run. Clean data is boring. Wrong data is expensive. Choose accordingly.