Every carrier with a Direct Repair Program grades you. State Farm Select Service, Progressive Pro, Geico ARX, Allstate Good Hands Repair Network — all of them score your shops monthly, some weekly, on a small set of KPIs that decide whether the assignments keep coming.
The problem isn't that the metrics are secret. The metrics are well-known. The problem is that most MSOs don't find out they slipped until the carrier scorecard lands a month later — at which point the damage is done and the assignments have already rerouted.
What Carriers Actually Score
The exact weights differ by carrier and by region, but the metric set is remarkably consistent:
- Cycle time — keys-to-keys in calendar days, usually with a carrier-specific target (typically 5–10 days depending on severity).
- Severity — average repair cost per claim, compared to regional benchmarks.
- CSI — customer satisfaction from the carrier's own post-repair survey.
- Rework / comeback rate — customers returning for the same repair.
- Supplement rate and timing — how often and how late supplements hit.
- Rental days — indirect cost to the carrier, usually tied closely to cycle.
- Photo and documentation compliance — did the shop submit what the carrier asked for.
Every one of those is computable from data you already have in CCC ONE. None of them require the carrier's scorecard to calculate yourself.
Why Waiting for the Carrier Scorecard Is Too Late
The carrier's scorecard is a trailing indicator. It tells you what happened last month. By the time you see you slipped from Tier 1 to Tier 2 on Progressive Pro, the shop has already had four weeks of bad throughput — which means the next scorecard is probably going to be worse, because the underlying process didn't change.
This is the same reason nobody drives a car by looking only in the rear-view mirror.
The Internal Scorecard: What to Build
The thing to build is your own shadow scorecard per carrier, per shop, updated daily, using the same logic the carrier uses as closely as you can reconstruct it. Three layers:
Layer 1: Current month, running
For each (shop × carrier) pair, show month-to-date: cycle time, severity, supplement rate, rental days, CSI if you have it. Compare against the carrier's tier thresholds (you know these — they are shared in the DRP agreement and the program guide).
Layer 2: Trailing 90 days
Most carriers weight a rolling window, not a single month. Show trailing 90-day averages with a trend arrow. This is the metric you can actually move with operational changes.
Layer 3: Alerting
Set a threshold tighter than the carrier's threshold. If State Farm wants cycle under 8.0 days and you are at 7.5 this week, you get a proactive flag — not because you're failing, but because you're one bad week from failing.
What the Scorecard Surfaces That Month-End Reports Miss
- Carrier-specific patterns. Shop 7 might be fine on Progressive and terrible on State Farm. That's a workflow problem, not a shop problem, and you won't see it in a shop-level P&L.
- Severity drift. If your severity for a specific carrier starts trending toward a regional average you haven't been tracking, that's a leading indicator of a DRP review.
- Supplement timing. Carriers penalize supplements that hit after the vehicle is in the shop for more than 48 hours. This is easy to measure and almost nobody does.
- Mix shifts. A shop taking on more heavy work will naturally see cycle go up. A shadow scorecard adjusts for mix; a carrier scorecard often doesn't.
What This Looks Like Built Out
The version we run for MSO customers sits on top of the CCC ONE data already flowing into the warehouse. It's:
- One page per carrier, one row per shop.
- Current-month and trailing-90-day values for each scored metric.
- Red/yellow/green vs the carrier's tier thresholds (and vs a tighter internal threshold).
- A trend arrow and a sparkline.
- An automated alert when any (shop × carrier × metric) crosses a threshold.
- A monthly PDF that leadership can put in front of a carrier rep at a performance review.
Total build time once CCC data is already in the warehouse: roughly two weeks. Maintenance: near zero, because the metrics and thresholds don't change often and the warehouse data is already reliable.
The Political Upside
The other reason to run an internal scorecard: when you walk into the carrier performance review with your own dashboard, the conversation changes. Instead of reacting to the carrier's data, you're bringing your own — broken down by adjuster, by claim type, by severity band. Carrier reps notice. It frames you as an operator who manages the relationship, not one who waits to be graded.
That's a soft benefit, but every MSO owner who has done it tells us the same thing: the carrier relationship gets easier the moment you show up with your own numbers.