Explore real-world item set scenarios focusing on detecting and correcting inaccurate financial data, normalizing earnings, and spotting non-recurring items to build more reliable equity valuations.
Enhance Your Learning:
“Data scrubbing”—sometimes also called “data cleaning”—is one of the less glamorous but most important tasks in equity valuation. You know how sometimes you stumble on a dataset that looks great at first glance, but then you realize some entries just don’t make sense, or maybe there’s a suspiciously large revenue jump in Q4? It’s in these moments you discover that data integrity is everything. And for CFA Level II, success often comes down to how quickly you can spot such anomalies, correct or adjust for them, and move on to the next question.
This section focuses on “Vignette Drills,” which imitate exam-style item sets. You’ll see the messy data, footnotes, uneven trends, non-recurring charges, and potential accounting manipulations, then decide how to scrub and normalize those figures so you can arrive at a decent valuation. We’ll explore an example scenario based on a mid-cap manufacturer, highlight typical pitfalls, and see how these adjustments can feed into multi-stage forecasting. Let’s jump right in.
Vignettes are deliberately designed to simulate the real-life chaos you might face as an analyst. On exam day, you’ll have a series of item sets with a limited time to read through all details—footnotes, MD&A remarks, financial statements—and make quick but accurate judgments about what is or isn’t important.
• In the real world, unscrubbed data can mask the true performance of a firm. A one-time benefit might inflate net income, or a stealthy change in depreciation policy can artificially bolster earnings.
• In the exam world, you’ll face the same sort of puzzle but must respond under tight time constraints.
The end goal is to demonstrate you can spot red flags, reclassify suspicious items, and produce an adjusted data set for more reliable equity valuation. That’s what these exercises are all about—practice doing it fast and doing it right.
Picture this: You’re analyzing a 5-year financial history for a company we’ll call Mopart Inc., a mid-cap manufacturer specializing in machinery parts. The CFO changed in Year 3, followed by a sudden shift in how certain expenses were classified. Then in Year 4, Mopart sold a piece of its business for a one-time gain. Meanwhile, revenue growth looks great, but you notice a significant jump in new distribution deals—even though shipping costs have oddly remained flat.
If that’s not enough to raise your eyebrows, Mopart’s R&D costs dropped significantly in Year 5, just as they announced a brand-new product line. “Huh,” you might say. “How do you develop a new product line with less R&D spending?” This scenario is ripe for a data-scrubbing test.
Let’s break down a process you might follow to keep your analysis squeaky clean:
• Look for “weird spikes”: A year or quarter where revenue or expenses deviate wildly from the trend.
• Check footnotes or MD&A (Management Discussion & Analysis) sections for major events, such as asset sales, restructurings, or changes in accounting policy.
• Verify any huge intangible asset write-downs, big legal settlements, or insurance payouts.
Mathematically, many analysts do a z-score test to find outliers in historical data:
If |z| > 3 (for instance), that data point might be an outlier worth investigating.
A new CFO might decide to:
• Change depreciation schedules (e.g., from straight-line to accelerated).
• Reclassify certain operating expenses as capital expenditures.
• Tweak inventory accounting from LIFO to FIFO (where permissible).
Each of these changes can make older periods non-comparable with the most recent data. So, watch for subtle clues in footnotes or differences in line item headings from year to year.
If you think a one-time gain artificially boosts net income, consider removing it from your normalized figures. For example:
• A big disposal gain from selling property might inflate net income this year, but it doesn’t reflect ongoing operating performance.
• Conversely, a large one-time lawsuit settlement might deflate net income artificially.
After removing these effects, you can recast the financial statements. This might mean you produce an “adjusted net income” figure or restate historical depreciation charges to create consistent data across all 5 years.
Management Discussion & Analysis is often where companies provide explanations for big changes in results. They might talk about supply chain disruptions, reclassification of intangible assets, or new partnerships. The MD&A can be your best friend: sometimes they come right out and say, “We sold a major division for a $40 million gain.” That’s a big help in explaining suspicious figures.
Sometimes a short-term anomaly is genuinely short-term. Other times it’s the start of a new “normal.” Distinguishing these requires both quantitative analysis and a qualitative assessment of the company’s fundamentals. A single quarter’s revenue spike might not be part of an upward trend if management is pulling forward future sales or offering deep discounts (aggressive revenue recognition, anyone?).
Below is a simplified flowchart of how data scrubbing often flows from raw data to a final, normalized dataset:
flowchart LR A["Raw Data <br/>(Financials)"] --> B["Identify Outliers <br/>(Revenue, Expense)"] B --> C["Examine Footnotes <br/>and MD&A"] C --> D["Adjust or Recast <br/>Statements"] D --> E["Normalized Data <br/>For Valuation"]
For exam prep, you might get a single item set with Mopart Inc.’s 5-year statements—revenue, cost of goods sold, operating expenses, net income, plus a snippet from the MD&A. The questions will test you on:
• Which line items to adjust or remove for normalizing earnings.
• Whether certain expenses are recurring or one-offs.
• Whether the changes in accounting policy require you to reclassify older data.
• How you’d handle newly discovered off-balance-sheet items (like operating leases that are not capitalized).
In a typical exam scenario, each question might be multiple-choice or short-answer style. Some queries focus on the rationale: “Why would you adjust for this gain?” Others test your calculations: “What’s the adjusted EPS for Year 4 if we remove the entire after-tax effect of that $20 million disposal?”
Data scrubbing isn’t just about mechanical adjustments; it’s also about integrity and analysis ethics. Over-adjusting can be as misleading as ignoring the outliers:
• It’s tempting to keep removing “bad stuff” to make the numbers look nice. But that might yield an overly rosy forecast.
• Or, if you have a hunch the stock is overpriced, you might be too quick to label everything a “recurring expense.”
The CFA Institute Code of Ethics demands objectivity. You want to reflect the underlying economic reality, not push an agenda. Keeping robust documentation of each adjustment, referencing official disclosures, and verifying your assumptions are essential.
Ultimately, the reason you’re doing all this scrubbing is to feed more accurate forecasts into your valuation model—like a multi-stage DCF or a multi-year FCFE approach. For instance:
• If you correctly identify the one-time disposal gain, you won’t carry that forward as part of free cash flow.
• If you uncover that R&D was artificially low, you might adjust your forecasted R&D expense upward to reflect the normal operational level.
• If you see an accounting policy shift that reduces depreciation temporarily, you’ll want to reflect the correct depreciation expense in your forward estimates.
A well-scrubbed data set helps forecast stable growth patterns or cyclical fluctuations that align with the firm’s actual operations, not illusions caused by misclassification or unusual transactions.
It helps to look at actual companies to see how complicated real data can get. For example, in the SEC’s EDGAR system, you’ll find annual reports containing footnotes describing intangible write-downs, goodwill impairments, or acquisitions that change the entire corporate structure. If you have time, skimming a real 10-K can give you a sense of the layers of detail.
Case studies such as Enron (though quite dated now) highlight the extremes of data manipulation, while more recent controversies involving revenue recognition or off-balance-sheet financing remind us that the lines can still be blurred in modern times. In the CFA exam, you won’t see actual brand-name companies, but you’ll see plenty of fictional ones that display the same issues.
When you’re in the actual exam environment, the key to rocking a data-scrubbing vignette is systematic thinking. Check each suspicious number, dig into the footnotes, adjust or recast as needed, and remember to apply consistent logic across all periods in your analysis. Resist the urge to get emotional or make knee-jerk decisions about what “should” be removed. Instead, be methodical, unbiased, and thorough.
That’s data scrubbing in a nutshell—maybe not the most glamorous step, but arguably one of the most critical in equity valuation. The payoff is that your final forecasts and intrinsic value estimates will be a whole lot more believable.
• CFA Institute. Practice Problems on Advanced Data Issues and Normalization.
• Sample real-world annual reports on SEC EDGAR (SEC.gov) with known anomalies.
• Gordon, E. A., Henry, E., & Palia, D. (2004). “Related Party Transactions and Corporate Governance.”
• Official MD&A sections in company filings: an indispensable source for context on suspicious items.
Additional Notes & Exam Tips
• When you see an unusually high or low figure, always ask “Is it sustainable?”.
• Use footnotes, MD&A, and changes in key policy disclosures to figure out if you need to recast statements.
• Remember that exam item sets might isolate a single line item (like an inventory write-down) or combine multiple issues (like intangible write-downs plus a shift in revenue recognition). Stay organized!
• If you must guess on an adjustment, justify it logically. The exam is about your process as much as your final number.
Good luck with your practice, and remember: scrubbing data is one of those unflashy tasks that can set you apart as a thorough, ethical analyst. It’s not just seeing the numbers; it’s seeing the story behind the numbers—and a story that’s free of noise or illusions is a story well told.
Important Notice: FinancialAnalystGuide.com provides supplemental CFA study materials, including mock exams, sample exam questions, and other practice resources to aid your exam preparation. These resources are not affiliated with or endorsed by the CFA Institute. CFA® and Chartered Financial Analyst® are registered trademarks owned exclusively by CFA Institute. Our content is independent, and we do not guarantee exam success. CFA Institute does not endorse, promote, or warrant the accuracy or quality of our products.