Performance calibration is where the real decisions get made. Not in your 1:1s, not in the written reviews, but in the room where managers debate and negotiate final ratings.
Most calibration sessions are political theater. The loudest voices win. The best-prepared manager gets their people rated higher. And the quiet contributor with the introverted manager? They get averaged down.
This guide covers how calibration actually works, how to prepare for it, and how to make the process fairer for everyone.
What Performance Calibration Is (And Why It Matters)
Calibration is the process where managers collectively review and adjust individual performance ratings to ensure consistency across teams. The goal is fairness: someone rated "exceeds expectations" in Team A should be comparable to the same rating in Team B.
In theory, calibration removes bias and creates consistency. In practice, it often introduces new biases while managers trade ratings like currency.
The Three Biases That Destroy Calibration Sessions
Recency Bias
What happened in the last two months matters more than the previous ten. That big win in Q4? It overshadows the slow start. The recent mistake? It tanks the whole year.
The fix: Document evidence throughout the year, not just before review time.
Halo Effect
One impressive trait colors everything else. The engineer who speaks well in meetings gets rated higher on technical skills. The designer with beautiful slides gets credited for strategic thinking.
The fix: Rate specific competencies separately before combining into an overall rating.
Similarity Bias
We overrate people who remind us of ourselves. Same background, same communication style, same career path. We "get" them, so we rate them higher.
The fix: Use structured criteria and challenge yourself on ratings for people unlike you.
How to Prepare for Calibration
Step 1: Gather Evidence for Each Person
Before calibration, compile specific examples:
- Key accomplishments with business impact
- Challenges they overcame
- Development areas with specific instances
- Peer feedback themes
Step 2: Anticipate Challenges
For each person, ask: What would someone argue against this rating? Prepare your response.
Step 3: Map Network Value
Who does this person enable? Who relies on them? Traditional reviews miss network contribution. If you have ONA data, use it.
Step 4: Prepare Your Narrative
You have 2-3 minutes per person. Make it count with a clear story: role, key contributions, rating rationale, development focus.
Making Calibration Data-Driven
The best calibration sessions use data beyond manager opinion. Organizational Network Analysis shows who people actually turn to—revealing hidden high performers and over-promoted under-contributors.
When someone challenges your rating, "the data shows they're in the top 10% for cross-team collaboration" is stronger than "I think they're really collaborative."
Common Calibration Scenarios
Your High Performer Gets Challenged
Response: Lead with impact, not effort. Quantify results. Show network value if available.
Distribution Forces Changes
Response: Know your hill to die on. Protect your true top performers; be willing to adjust the middle.
Recency Bias Surfaces
Response: "That was one quarter. Here's the full-year picture." Evidence over anecdote.
After Calibration
- Review final ratings — Understand what changed and why
- Prepare conversations — Your people deserve honest explanations
- Document learnings — What will you do differently next cycle?
The Performance Calibration Checklist
- Evidence gathered for each person (2+ specific examples)
- Ratings defensible with business impact
- Network contribution documented
- Challenges anticipated and responses prepared
- Narrative clear and concise (2-3 minutes each)
Calibration doesn't have to be political. With preparation and data, you can advocate effectively for your team while keeping the process fair.
Read the full manager effectiveness guide | Learn when to coach vs direct
