How to Measure Any Leadership Responsibility (Even the Ones That Feel Unmeasurable)
What gets measured gets managed. Here's how to find a real metric for any responsibility on a role scorecard.
The CEO had pulled up the Role Scorecard for his Head of People on the screen between us. We were three weeks into building scorecards for his leadership team. The CRO had a number. The CFO had a number. The Head of Engineering had several. The team was getting clearer about what each role was actually for.
This one was different. The responsibilities were listed. Develops talent. Builds culture. Drives engagement. Supports the leadership team. Runs people operations. Underneath each one, the metric column was either empty or filled in with something soft. Engagement: "team feedback." Culture: "tone of the office." Talent development: "leadership team strength."
He looked at me and said the line I hear constantly when this part of the scorecard work hits a wall. "I don't think you can really measure most of this." The implication being that the role just is what it is. You hire someone you trust and hope it works.
I told him no. The pattern on his screen is The Toothless Standard. The scorecard names the responsibility but has no metric with enough bite to drive accountability. The standard exists. It just doesn't bite. And every leadership responsibility, including the ones that feel resistant to measurement, has a measurable expression. The work is finding it.
What The Toothless Standard Sounds Like
The pattern shows up in language before it shows up on the page. Phrases like "more art than science," "you'll know it when you see it," "it's about judgment," or "we just need someone good in this role." Each of these is a flag that the leader has stopped trying to define what good looks like.
There's a legitimate version of these statements. Some leadership responsibilities genuinely require judgment that no single number captures. There's also a cop-out version, where the language gets used to avoid the harder work of defining the outcome. The first is honest. The second leaves the role and the company exposed.
The Toothless Standard concentrates at senior levels. The further up the org chart, the more the responsibility is to deliver an outcome rather than perform specific activities. The further down, the more activity metrics make sense. A senior leader's job is the result. An individual contributor's job is often the activities that produce the result.
A Head of Sales is measured on total new revenue closed. That's an outcome. An appointment setter is measured on meetings booked. That's an activity. Both metrics are appropriate for the level. If you measured the appointment setter on monthly closed revenue, you'd be holding them accountable for an outcome they can't directly drive. If you measured the Head of Sales on calls dialed, you'd be under-scoping a role whose accountability is the number, not the activity that produces it.
Toothless metrics tend to appear when senior leaders have responsibilities that require outcome-level measurement, and the leader doing the scorecard hasn't done the conversion work yet. The activity metric or the soft phrase fills the gap. The standard looks complete on paper. It has no bite.
From Responsibility To Metric: A Four-Step Conversion
The conversion from a responsibility to a working metric is four steps. None of them requires special tooling. They require thinking about the role differently.
Step 1: State the responsibility as an outcome, not an activity. Most scorecard responsibilities get written as activities because activities are easier to describe. "Develops people." "Manages the team." "Drives strategy." None of those describes an outcome. They describe a category of work. Convert: "Builds and retains a leadership bench that can run the business as it scales." "Delivers operational performance against the plan." "Translates strategy into a roadmap the team can execute." The outcome statement makes the metric possible by naming what should change in the world if the role is done well.
Step 2: Pick a measurement pattern, or invent one. The five patterns in the next section are starting points, not a closed list. The goal is to make the measurement as objective and independently verifiable as you can. If one of the five fits, use it. If you need to invent something else that does the job, go ahead and invent it. The constraint is the property of the measure, not the pattern.
Step 3: Set the cadence. Most leadership-level metrics are reviewed monthly or quarterly. Some leading indicators make sense on a weekly basis, but the headline metric for a senior role usually doesn't move on a weekly cycle. Cadence has to match the natural rhythm of the outcome. A retention metric is meaningless weekly. A pipeline metric is meaningful weekly. Set the cadence to match the signal.
Step 4: Accept the trade-off. The most precise possible measurement is often too costly to capture. The cost shows up in time, in tools, or in the friction of asking the team for data they don't already produce. A directionally correct proxy beats a precise metric you can't actually run. If the cleanest metric for "drives leadership team execution" requires an external project audit, the proxy of "percentage of leadership team commitments closed on time, tracked in the weekly meeting" gets you most of the way there for almost no cost. Take the proxy. Sharpen later if you need to.
Five Measurement Patterns To Choose From
The five patterns below cover most leadership responsibilities. They are framed for growth companies, which means they assume you have the data the business already produces, calendar and meeting cadences, and direct relationships with the people doing the work. None of them requires formal HR programs.
Direct performance metric. A number from the business itself. Revenue, retention, gross margin, on-time delivery, churn rate, defect rate, pipeline coverage. Use this when the responsibility is directly tied to a business result the role owns. The cleanest case.
Process or event count. Something countable from how the work happens. Number of one-on-ones held against the schedule. Number of cross-functional decisions documented. Number of escalations resolved within the agreed window. Number of leadership team members with completed development plans. Use this when the responsibility is about doing certain things reliably and consistently, even when the downstream outcome is influenced by many factors. Often a leading indicator paired with a lagging direct metric.
Binary or threshold. Yes or no. Did the thing happen at the agreed quality bar? The talent calibration was completed this quarter. The annual operating plan was published by the deadline. The succession plan exists for the top ten roles. Use this when the responsibility is inherently discrete. Some critical leadership work is binary. Either the strategy got communicated to all teams by the deadline or it didn't.
Artifact quality check. A deliverable exists and meets a quality bar. The strategic plan is current and reviewable. The operating model document reflects the actual org. The customer feedback report is published monthly. Score against a simple rubric, with the CEO or a peer doing the scoring. Use this when the responsibility produces a document or framework that drives the rest of the work. Quality is checkable. The check creates accountability.
Stakeholder signal. Someone with line of sight scores the work on a structured scale. The CEO rates the Head of Strategy's quarterly framework on a 1-5 rubric. Three peer leaders rate the Chief of Staff's effectiveness as a coordination function. The board scores the CFO's reporting clarity. Use this when the value of the role is felt by specific stakeholders, and there's no clean number that captures it. Keep it simple. Three raters, a 1-5 scale, and a structured rubric beat anonymous surveys with thirty questions.
Five Roles, Five Scorecards, Twenty-Five Metrics
Worked examples for five roles where The Toothless Standard tends to show up. Each role has a brief overview and five Key Responsibilities paired with a KPI and Red/Green/Wow thresholds. The format is intentionally tight. Real scorecards are short, scannable documents that drive weekly conversations. They are not essays.
The thresholds below are illustrative. Real numbers depend on the company's stage, plan, and history. Use the format. Set your own targets.
Head of People
The role exists to build and protect the talent infrastructure that scales with the company. At $10M to $100M, this often combines what would be three separate roles at larger companies.
Builds the leadership talent bench. KPI: Succession coverage % for top ten roles, defined as roles with at least one ready-now or ready-soon successor (R = <60%, G = 80%, W = 100%)
Drives retention of A-players. KPI: Voluntary regrettable turnover % on leadership team and one level below (R = >12%, G = 8%, W = <5%)
Hires A-players into the company. KPI: 90-day new hire performance rating from hiring managers, 1-5 scale, quarterly cohort average (R = <3.5, G = 4.0, W = 4.5+)
Develops leadership team capability. KPI: Completion % of agreed development commitments across the leadership team (R = <70%, G = 85%, W = 95%+)
Builds and protects company culture. KPI: Quarterly all-team engagement score on "I am proud to work here" (R = <70%, G = 80%, W = 90%+)
Chief of Staff
The role exists to extend the CEO's leverage. Responsibilities span operating rhythm, leadership team coordination, CEO time, special projects, and decision support. The role rarely has a measurement template, which is why it tends to drift.
Drives the leadership team's operating rhythm. KPI: % of weekly leadership team commitments closed on time (R = <70%, G = 85%, W = 95%+)
Coordinates execution on company priorities. KPI: % of quarterly company priorities at green status by quarter end (R = <60%, G = 80%, W = 95%+)
Manages CEO time and priorities. KPI: % of CEO calendar time on Defensible Calendar priorities (R = <50%, G = 70%, W = 85%+)
Delivers special projects on assignment. KPI: % of assigned projects delivered on time at agreed quality (R = <70%, G = 85%, W = 95%+)
Supports leadership team decision quality. KPI: Stakeholder rating from CEO and two peer leaders on coordination effectiveness, 1-5 scale (R = <3.5, G = 4.0, W = 4.5+)
Chief Operating Officer
The role exists to run the operating engine of the business. The breadth of the role is what makes the metrics drift. The fix is splitting Run from Build and measuring both.
Delivers core operational performance. KPI: Performance against operating plan on revenue, gross margin, and budget (R = <90% of plan, G = 100% of plan, W = 110%+ of plan)
Drives execution on strategic initiatives. KPI: % of quarterly strategic initiatives delivered on time and on adoption targets (R = <60%, G = 80%, W = 95%+)
Maintains the operating model and decision rights. KPI: Operating model and Decision-Making Rights map reviewed and updated on agreed cadence (R = missed reviews, G = all reviews on time, W = on time with documented improvements adopted)
Runs the company's execution rhythm. KPI: % of leadership team commitments closed on time across the operating leadership team (R = <70%, G = 85%, W = 95%+)
Develops the operating leadership team. KPI: Stakeholder rating from peers on coordination effectiveness, 1-5 scale (R = <3.5, G = 4.0, W = 4.5+)
Chief Marketing Officer
The role exists to build the company's market presence and demand. Demand gen has clear numbers. Brand, positioning, and customer insight do not, which is where the toothless standards usually live.
Generates a qualified pipeline. KPI: Marketing-sourced pipeline dollars vs plan (R = <85% of plan, G = 100% of plan, W = 115%+ of plan)
Builds and maintains brand and positioning. KPI: Annual brand health study composite score (R = <3.5, G = 4.0, W = 4.5+)
Enables sales with effective collateral and tools. KPI: Sales team rating on marketing collateral usefulness, 1-5 scale (R = <3.5, G = 4.0, W = 4.5+)
Develops customer insight that informs strategy. KPI: Quarterly customer insight brief rating from CEO and Head of Product, 1-5 scale (R = <3.5 or missed quarter, G = 4.0, W = 4.5+)
Drives go-to-market for product launches. KPI: % of major product launches meeting agreed readiness criteria on time (R = <70%, G = 90%, W = 100%)
Head of Customer Success
The role exists to keep customers, expand them, and turn customer experience into a competitive advantage. Retention has a number. The relationship work that produces retention is harder to measure directly.
Drives customer retention. KPI: Percentage variation of Net Retention Rate vs operating plan (R = <-10%, G = 0%, W = +10%)
Expands existing accounts. KPI: Expansion ARR vs plan (R = <85% of plan, G = 100% of plan, W = 115%+ of plan)
Delivers customer onboarding at quality. KPI: % of new customers reaching agreed activation milestone within thirty days (R = <70%, G = 85%, W = 95%+)
Builds customer advocacy and references. KPI: Number of new active customer references and case studies per quarter (R = <3, G = 5, W = 8+)
Develops the customer success team. KPI: Voluntary turnover % on the customer success team (R = >12%, G = <8%, W = <5%)
Diagnostic: How To Audit Your Scorecards For Toothless Standards
Run these four questions across your current leadership scorecards. The first two test the metric itself. The second two test whether the metric is scoped to the role.
For the softest-feeling responsibility on each scorecard, can you state the metric, the data source, and the cadence in one sentence? If you cannot, the metric is not real yet. The vagueness is doing the work of avoiding the conversation.
When you read each metric out loud, does it describe an activity or an outcome? Activity metrics like "calls made," "campaigns shipped," or "interviews conducted" are the most common form of toothless standard. They confirm work is happening but do not tell you whether the work is producing the outcome the role exists to drive.
Is the metric scoped to what this person actually controls, and does it match their level? A Head of Sales measured on calls made is under-scoped. An appointment setter measured on closed monthly revenue is over-scoped. Senior leaders need outcome metrics tied to what they own. Individual contributors need activity metrics tied to what they directly do. Mismatches in either direction are a flag.
Are the metrics as objective and independently verifiable as you can make them? Some softness is unavoidable. Some responsibilities legitimately require rater input or judgment calls. But if every metric on a scorecard is soft, the leader has not done the work. Aim for objectivity wherever it is available. Accept softness only where you have to.
Where To Start
Pick the role on your leadership team where you are most likely to dodge the conversation. The one you have been telling yourself that the metrics will come into focus eventually. Start there.
Sit down with that scorecard and walk through the four-step conversion for one responsibility at a time. State the outcome. Pick a measurement pattern, or invent one that meets the bar of objective and verifiable. Set the cadence. Accept the trade-off when the most precise metric is not worth the cost.
Do not try to fix every scorecard at once. Do not try to fix all five responsibilities on a single scorecard at once, either. The work is iterative. A scorecard with one sharp metric and four soft ones is better than the scorecard you started with. Done is better than perfect.
The standard to hold yourself to is that every responsibility on every leadership scorecard has a metric you can describe in one sentence. What is being measured? What data source produces it? Who reviews it? How often? If you cannot yet do that for every responsibility, you have a Toothless Standard somewhere in your organization. The good news is the work is finite. Pick the softest responsibility. Sharpen it. Move to the next.
What gets measured gets managed. Most of what is not getting managed on a leadership team is sitting in plain sight on a scorecard, with nothing actually measuring it.
Questions for You and Your Team
The fastest way to find The Toothless Standard in your own organization is to look at where you have been letting metrics slide. These three questions will surface where the conversion work is still ahead of you.
Which role on your leadership team would you struggle to write a real metric for if I asked you right now? The role you cannot write a clean metric for is the one with The Toothless Standard hiding inside it. The discomfort is the signal.
When you read your senior leaders' scorecards out loud, are the metrics scoped to outcomes they actually own, or to activities that look more like the work of the level below them? Senior roles with activity metrics are under-scoped. The conversion work has not been done yet.
If you removed every soft metric from your leadership scorecards, would there be enough left to actually run a performance conversation? If the answer is no, you are managing leaders on impressions rather than measurements. That is where drift accumulates.
The goal is not to have perfect answers. It is to surface whether The Toothless Standard might be affecting how your leadership team is actually being held to account.
Take the Next Step
If you want to test how clearly your leadership team is operating against measurable standards, take the Leadership Team Assessment. It will surface where roles, expectations, and metrics are clear and where they are soft.
To talk through your scorecard situation directly, book a call.