Blog post

What Actually Happens During a Calibration Session: A Fly-on-the-Wall Account

Watch how real calibration sessions differ with and without data. A narrative look at bias, groupthink, and how structured data corrects manager judgment.

What Actually Happens During a Calibration Session: A Fly-on-the-Wall Account
Last updated: April 2026

You're invited to sit in on two calibration meetings. Same company, same 12 managers, same 200 engineers. One happens the way it always has. The other happens with structure. Pay attention to what changes.

The First Room: Without Data

The conference room is already tense when you arrive. It's 2 PM on a Thursday, and everyone knows this takes hours.

"Alright, let's start with engineering," says Claire, the VP of Engineering. She's got a spreadsheet open, but nobody's really looking at it. "Marcus. Everyone have thoughts on Marcus?"

Sarah leans back in her chair. "Marcus doesn't collaborate well. He's the type to just do his own thing. I've heard feedback from multiple people that he's not a team player."

Sarah is the Engineering Manager for backend systems. She says this with total confidence. Several people nod.

"Yeah, I've noticed that too," adds Tom from infrastructure. "He's solid technically, but he doesn't play well with others."

Marcus's skip-level manager, James, says nothing. He works for someone else's team.

"So Marcus is what, a 3?" Claire asks. On a 5-point scale, a 3 is below expectations. It's the beginning of a paper trail toward exit.

"Yeah, 3 feels right," Sarah says.

Nobody challenges this. The data is just vibes. The loudest voice in the room has set the tone, and groupthink is doing the heavy lifting.

This exact moment happens to high performers every day in companies that aren't measuring.

Move to the next person. Jessica is a designer who has won company design awards twice. Her manager, Paul, describes her as "solid, good designer, keeps her head down."

"Is she high-performer territory?" Claire asks.

"I think she's borderline," Paul says. "Good work, but she's not super visible."

"Okay, borderline," Claire repeats, writing this down. Nobody pushes back. Nobody asks Paul what "visible" means or how he's measuring it. The meeting moves on.

Three hours later, the room has made calibration decisions about 200 people. They've done it with no shared data, no common framework, and no way to audit any of it. If Sarah's bias against Marcus shows up in his actual work output, they'll never catch it because they never looked at his actual work output. If Jessica's quietness was mistaken for lack of impact, nobody will know because nobody quantified Jessica's impact.

The whole thing takes the form of objectivity. There's a spreadsheet, rankings, notes. But the substance is pure human judgment, uncalibrated and untested.

The Same Room, Different Approach

Now watch what happens when the same company brings structure to the same meeting.

Claire starts with a brief intro: "We're going to work through our engineering teams using five data points. We've prepped these so we're not debating opinions."

She clicks to the first slide. It shows five engineers, all from backend systems (Sarah's team). Before they calibrate anyone's performance rating, they look at work data.

Marcus appears first. Alongside his name:

  • Lines of code reviewed this quarter: 287 reviews completed
  • Cross-team projects: 12 active collaborations (Infrastructure, ML, Frontend)
  • On-call incident resolution: 18 resolved, avg. response time 23 minutes
  • Code quality metrics: 8.3/10 (peer review score)
  • Peer feedback on collaboration: "Clear communication in async channels," "Always explains decisions," "Helped me debug authentication issue at 11 PM"

The room goes quiet.

"Wait," says Sarah. "Twelve cross-team projects? I wasn't aware of all of those."

"These are from our collaboration tracking tool," Claire says calmly. "They logged hours on external projects. If you look at the peer feedback (from anonymous surveys we ran), the collaboration story is pretty different from what we discussed in previous sessions."

Sarah shifts in her seat. This is awkward, but productive. Nobody is attacking her judgment. The data is doing the talking.

"So Marcus isn't a 3," James says, speaking up for the first time. "Based on this?"

"Let's see what calibrated looks like," Claire says. "Compare Marcus to the others on his level."

They look at four other engineers with similar technical rankings. Marcus's cross-team work is higher. His code quality is comparable. His incident response is faster. His peer feedback is more consistently positive.

"I'd put him at a 4," says Tom, the infrastructure guy. "Maybe even a solid 4."

"Fair," Sarah says. She doesn't like it, but the data isn't debatable.

The meeting moves to Jessica. Same treatment.

Her data:

  • Projects completed: 8 major designs shipped
  • Stakeholder feedback: "Thoughtful, asks great questions," "Simplifies complex problems," "Always available for feedback loops"
  • Design system contribution: 23 components documented
  • Cross-functional collaboration: 7 departments listed her as key collaborator
  • Time to production: 32 days average (company average: 47 days)

"This doesn't look like a borderline performer," Claire says. "What does this look like?"

"This looks like someone who works quietly but moves faster than most people in the building," James says. "4, easily."

Paul, Jessica's manager, is quiet. He was wrong about her impact. But the mechanism that proved him wrong wasn't public humiliation. It was data. The data made it impossible for anyone to stay attached to the wrong conclusion.

What Changed

In the first room, they made 200 calibration decisions in three hours with no common framework. Decisions landed on who was most memorable in the last 30 days, who talked loudest in meetings, whose manager fought hardest to defend them, and gut feeling dressed up as intuition.

In the second room, 200 calibration decisions landed on what the person actually accomplished, what their peers actually experienced, measurable collaboration patterns, how they compared to people at similar levels, and a consistent method everyone could see.

The first room had bias and called it judgment. The second room had judgment informed by data that corrects for bias. It's the difference between "I think" and "Here's what happened."

It's also auditable. Six months later, if anyone asks "Why was Marcus a 3 when he was promoted to a 4?", the answer is right there: "We looked at his collaboration data and realized we had incomplete information. His actual cross-team project count was 12, not the 2 we were aware of."

In the first room, the answer is: "Sarah said he didn't collaborate well, and we believed her."

Who This Matters For

If your calibration sessions look like the first scene (full of debate, consensus-seeking, and confidence in people who don't actually know), then you're making career decisions on incomplete information. High performers are getting rated as mediocre because they're quiet. Political operators are getting rated as stars because they're loud. Bias is unopposed because nobody's measuring anything.

The second room isn't perfect. Data isn't perfect either. But it's a lot harder to be systematically wrong when you're looking at what people actually did instead of how you remember them.

If your calibration sessions look more like the first scene, we should talk. Book a demo and see how Confirm's calibration software lets you ground your sessions in data.

See Confirm in action

See why forward-thinking enterprises use Confirm to make fairer, faster talent decisions and build high-performing teams.

G2 High Performer Enterprise G2 High Performer G2 Easiest To Do Business With G2 Highest User Adoption Fast Company World Changing Ideas 2023 SHRM partnership badge — Confirm backed by Society for Human Resource Management Brandon Hall Group Excellence in Technology Award 2023 HR Executive Top HR Products 2023 Tech Trailblazers Award Winner 2023