Tech Company Calibration Playbook
How engineering-led companies run fast, bias-resistant calibration sessions — from 50-person startups to multi-team engineering orgs with dual IC and management tracks.
Why Tech Companies Calibrate Differently
Tech companies have structural characteristics that make generic calibration advice a poor fit. Dual-track career ladders, remote-first teams, project-based impact attribution, and fast review cycles all create calibration challenges that HR templates from traditional industries don't address.
The three most common calibration failures in tech: (1) engineers on the IC track get compared to managers and lose because management work is more visible; (2) remote engineers are rated lower despite equivalent or superior contribution; (3) calibration sessions become de facto promotion discussions rather than rating alignment exercises.
Key PrincipleTech calibration must separate two distinct activities: rating employees against the bar for their current level, and deciding who is ready for the next level. Mixing promotion decisions into calibration inflates advocacy noise and distorts the whole session.
Calibration Cadence for Tech Companies
| Company Stage | Recommended Cadence | Typical Format |
|---|---|---|
| Startup (50–150) | Annual | Single session, all managers, 2–3 hours |
| Growth-stage (150–500) | Twice yearly (Q2 + Q4) | Department-level sessions + cross-dept sync |
| Mid-market (500–2,000) | Annual with mid-year check | BU-level calibration + executive summary session |
| Enterprise (2,000+) | Annual (phased by org) | Multi-phase: team → dept → BU → exec |
The Quarterly Calibration Option
Some high-growth tech companies run calibration every quarter, aligning it with quarterly performance check-ins. This works well when: review cycles are short (6 months), headcount is growing faster than 25% annually, or when promotion velocity needs to accelerate. The tradeoff is calibration fatigue — quarterly calibration only works when preparation is lightweight and sessions are under 90 minutes.
IC Track vs. Manager Track Calibration
The most important structural decision for tech company calibration is whether to run IC and management tracks separately. For companies below 100 people, combining them is workable. Above 100, separate sessions are almost always better.
Why separate sessions matter
- Comparison errors: When ICs and managers are calibrated in the same session, calibrators unconsciously weigh leadership behaviors more heavily — even on teams where senior IC work is the primary value driver.
- Advocacy asymmetry: Engineering managers calibrate their IC reports and attend calibration. Senior ICs who don't manage rarely have an advocate in the room who can speak to their work at the same level of detail.
- Level rubric mismatch: An L5 IC and an EM both at "Senior" have completely different expectations for scope and impact. Calibrating them against the same mental model produces grade inflation for one and deflation for the other.
How to run separate track sessions
Pre-session: Separate the pools
Before calibration, segment employees into IC and manager tracks. Assign each employee a proposed rating and a supporting justification using the track-specific rubric.
IC calibration session
Calibrate IC employees with the engineering leaders (VPE, principal engineers, senior ICs who can assess technical scope) as the primary calibrators. HR or HRBP facilitates.
Manager calibration session
Calibrate managers separately, with VP+ and CHRO/CPO as calibrators. Assess on team outcomes, people development, and organizational influence — not technical output.
Distribution alignment
After both sessions, check that rating distributions across IC and manager tracks are consistent. Systematic rating inflation on either track is a calibration signal, not a talent reality.
Remote-First Calibration: Addressing Visibility Bias
Remote engineers are rated lower than in-office counterparts across every industry, but the effect is especially pronounced in tech because so much performance signal is informal — who joins the meeting, who speaks in all-hands, who gets looped into the architecture discussion. Remote engineers miss most of those signals.
Data-based mitigation strategies
- Require contribution documentation: Before calibration, ask all managers to document specific project contributions with scope context. Eliminates "I don't have visibility into their work" as a calibration driver.
- Use ONA data: Organizational network analysis surfaces cross-team collaboration and peer trust signals that don't depend on physical presence. Remote engineers often have stronger collaboration graphs than their in-office counterparts.
- Flag remote status statistically: Before the session, show calibrators whether remote employees in their team are rated systematically lower. If the average remote employee in a team is 0.5 points lower than in-office peers, that's a bias signal, not a performance signal.
- Blind first pass: Have calibrators submit initial ratings before the session opens, without seeing other managers' ratings. This locks in assessments before social pressure can create convergence toward visibility-based ratings.
Watch ForThe "I don't have visibility" excuse is the most common way calibrators avoid accountability for remote team members. If a manager doesn't have enough visibility to rate an employee, that's a management gap — not an employee gap. Flag it, don't compound it in calibration.
Leveling Alignment in Calibration
Tech companies with formal leveling frameworks (L1–L7 or equivalent) should anchor every calibration to the level rubric, not to peer comparison. "Better than their peers" is not the same as "performing at the expected bar for their level."
Rating scale recommendations for tech
| Company Size | Recommended Scale | Distribution Target |
|---|---|---|
| Under 150 employees | 3-tier: Below/Meets/Exceeds | ~70% Meets, 20% Exceeds, 10% Below |
| 150–500 employees | 4-tier: Below/Meets/Strong/Exceptional | 60% Meets, 25% Strong, 10% Exceptional, 5% Below |
| 500+ employees | 5-tier with forced distribution guidance | Calibrate to a target distribution by org/level |
Bias mitigation at promotion decisions
In calibration, keep promotion discussions separate. Rate every employee against the bar for their current level. Then, in a separate session (or a clearly separated segment of the same session), review who is ready to be considered for promotion. The promotion bar is higher than "performing at level" — it requires demonstrated performance at the next level, not just strong performance at the current one.
Pre-Calibration Checklist for Tech Companies
- Ratings locked in the system before the session starts — no live input allowed
- IC and manager tracks separated into distinct employee pools
- Each manager has submitted a written justification for each rating, tied to the level rubric
- ONA or peer signal data pulled and distributed as pre-read material
- Remote employee list flagged for visibility-bias review
- Distribution report generated: show each manager's rating curve before the session
- New hires (under 6 months) and on-leave employees flagged — different standard applies
- Session agenda sent 48 hours in advance, including time allocations
Time EstimateA well-prepared calibration for 40–60 IC engineers takes 2.5–3.5 hours. A manager calibration for 10–15 managers takes 1.5–2.5 hours. If sessions are running longer, the pre-work is insufficient — managers are arriving without documentation and calibration is becoming the data-gathering exercise.
Tech Calibration FAQ
See Confirm in action
Confirm helps tech companies calibrate faster and more fairly — with ONA data, automatic bias detection, and level-based rubrics built in. See it in action.
