The tyranny of Metrics

The tyranny of Metrics

In the world of statistical measurement, there is an old adage: “The problem with statistics is that too many people use them like a drunk uses a lamppost; for support instead of illumination”

In the past quarter as companies were finalized 2021 business plans, many business conversations included discussions of OKR’s, KPI’s and other metrics, as companies cascaded their operational plans down through their organizations. As companies drive towards the use of better metrics driven planning and feedback tools, too often I see a tendency for the metric to become the goal instead of a path to the goal.

Do you understand the process before you measure it?

My father started his career in the 1950’s as an industrial engineer at the Steel Company of Canada. One day, he was escorting a senior cost accountant on a plant tour. As they watched steel ingots being produced, the accountant noticed a worker standing a few feet away from a cooling but still pink hot ingot with a tape measure. The worker was estimating the length, width and thickness of the ingot which would weigh a few tons, and then would look to a chart on the wall. Based on the dimensions, the chart specified the weight in pounds to several decimals of accuracy. The worker would then dutifully write down the weight (including several decimals of accuracy) and send the production sheet off to where it would eventually end up in the hands of cost accounting to assign a dollar value (to several decimals of accuracy) to the ingot. As the ingot continued to be processed, it was made into plate, rod, coils, or other products, cut or sliced to smaller dimensions, and possibly treated or coated. At each stage, the cost accountants manually re-calculated its value (to several decimals of accuracy).

Now, as the accountant understood the process, he was horrified thinking about the hours and hours of calculation done by his department (this was in the early days of computers and electro-mechanical calculators) and nearly apoplectic with rage. He demanded to know from my father why the worker was not weighing the ingot on the scale. My father calmly replied: “The scale is broken but even if it wasn’t, it is only accurate to the closest 100 pounds.” Afterall, why would you spend money on better scale accuracy for a product weighing thousands of pounds with a value of less than pennies per pound.

Are your measurements accurate or biased?

An area where I see companies struggle is to determine a success measurement system for the Customer Success organization. While many use Net Promoter Score (NPS) as a measure of customer satisfaction, I remain skeptical of its use as a predictor of future financial results. While the measurement itself may be accurate, does the process to collect it really reflect the client’s true intentions. Throughout the 1970’s and 80’s, IBM was viewed as an arrogant company that bullied customers and didn’t listen. As part of its transformation in the 1990’s, employees’ bonus and managers’ career trajectories were tied to customer satisfaction measurements. At one point, customer satisfaction was measured by a survey of a statistically valid sampling of customers, administered by an independent third party company. By doing so, IBM management believed that the results would be an accurate representation that could be used for analysis and decision-making. By tying the results to employee bonus, it was felt this would inspire correct behaviours.

Unfortunately, the process had flaws. The first flaw was someone had to produce a list of who to phone at each customer. Obviously, those closest to the customer (the account managers) would be the best source of whose opinion at each customer best represents the customer’s company. Since the account managers’ (and their boss’) bonus was tied to the results, the list was subject to biases towards those at the customer who had the most favourable impression of IBM.

In addition to prevent calls to customers who were in the middle of sensitive interactions with the account team such as negotiating a new contract or other “deemed sensitive” events, customer could be flagged by the account team for exclusion from the survey in any given quarter.

Although the list was randomly called each quarter, another flaw was to avoid survey fatigue, customers were never called more than once a year. As a result, managers could quickly isolate the list to subsets of customers each quarter that were likely to be called. This spawned quarterly campaigns to check in with every customer who might be called, to ensure they knew about the process, knew that bonuses were dependent on their answer and to inquire if they had any outstanding issues that could be taken care of for them.

No question, IBM’s customer satisfaction ratings rose but was it because customers were happier or was it because managers were cooking the system.

Does the methodology-used create metrics that predict outcomes?

One of my university professors was Polish. As a young scientist during WWII, he was conscripted by the Nazi to build the V1 and V2 rocket systems. As a conscripted worker, working conditions were hard. His captors were rigorous in checking calculations and any attempt at sabotage was dealt with by summary execution. To thwart them, my professor and his colleagues adopted a different strategy. They applied correct engineering formulae and calculations to the wrong problem. By doing so, failures occurred and progress of developing the rockets was slowed yet the failures could never be tied back to them.

Using a valid methodology to measure something does not ensure the metrics are valid if the measurement methodology is misapplied.

Is the sampling valid?

Many employee performance management systems use a “bell curve” to manage the distribution of performance ratings. In the world of probability and statistics, in certain situations, the Central-Mean Theorem underlies the principle that in a large sample set, a normal distribution (Bell Curve) will result. As a result, managers are told that 5% of your employees are in the bottom tier, 15% are in the next tier, 60% are in the middle, 15% are almost superstars and only 5% can be in the superstar category. While it is likely valid, in the large company with many employees, that the performance ratings of all employees will follow a normal distribution, it is also likely invalid that the performance ratings of a small team of employees will be normally distributed. Yet annually, managers of small teams are told to fit their performance ratings into the Bell Curve.

Are you measuring the “Cause” or the Effect”?

Understanding the process also helps understand and explain the relationships between “cause and effect”. Have you ever heard the argument that the price of our product is too high and therefore the sales team cannot sell their quota so they need to discount more? Is the price the cause of poor sales results or is poorly skilled salespeople, a lousy sales process, or targeting prospects who won’t receive high value from the product really the cause of why the company can’t earn the price they think their product deserves. Don’t confuse “cause” versus “effect”.

Have you validated the business impact of the cause-and-effect relationship?

A few years ago, a client was very hawkish on only hiring candidates who scored high on a particular aptitude test that their CEO liked. One of their senior leaders confessed to me that she wasn't sure that she would have been hired if the test had been in place when she had applied.

I asked a friend, who is a senior employment lawyer, whether a company could ask a prospective employee to take an aptitude test as part of the interview process. His opinion was they probably could not, unless they could demonstrate a validate correlation between the test results and performance of actual employees in their company. If they could not, he would argue that the test may have systemic biases imbedded in it, thus making it discriminatory.

While I often see companies create hypotheses about metrics like

  • This employment test is a good indicator of likely hiring success
  • This customer success process will result in higher customer retention
  • This sales methodology will increase win rates

rarely do I see follow-thru to rigorously confirm the hypothesis is correct or its applicability to their specific situation.

For example, when was the last time you saw metrics like:

  • Developers who score above a certain level in a coding employment test are x% more productive than those that fall below that level
  • Each additional point in NPS translates to $x of additional revenue in the lifetime value (LTV) of a customer
  • Implementing this new sales methodology has increased our win rate by x% while reducing the cost of customer acquisition (CAC) by x% and increased the LTV by x%

Metrics are a great business tool that support the management philosophy of “you can only manage what you measure”, but are your metrics valid measure of the objective or have they become "the objective"?

One final thought. Many software companies measure and celebrate the high NPS scores they receive from clients when they call their Support Line. Is this success? Or, is every time a customer phones a failure… a product quality issue… software that is not intuitive to use… product documentation or help functions that don’t solve the problem… a failure to set the proper expectation during the sales process… or software that was not properly configured during implementation…

If so, why are we celebrating a good NPS score?

Jenn Krieger ??in positivity and kindness ??

Customer Engagement Marketing Manager

3 年

Excellent info.

回复

要查看或添加评论,请登录

Peter Smith的更多文章

  • Change

    Change

    45-years-ago, on April 10th, a day later than expected, I started my full-time post-university career (see yesterday’s…

  • Indulge my reminiscences.

    Indulge my reminiscences.

    45-years-ago today was supposed to be my first day of work at IBM. I had written my last university exam the previous…

    15 条评论
  • Not knowing what you don't know

    Not knowing what you don't know

    For those of you who are working on your career plan, hopefully here is some food for thought. In 1977, I invented the…

    12 条评论
  • A new rule for return-to-office

    A new rule for return-to-office

    From “pandemic” to “endemic” … from “Work from Home” to “Return to Office” … what will 2022 bring as we “return to…

    10 条评论
  • The Power of “almost” Quitting

    The Power of “almost” Quitting

    I was chatting with a friend. She is an individual contributor and very frustrated with her job.

    9 条评论
  • The Peter Principle

    The Peter Principle

    Despite the title, this post is not about me. Have you ever worked with someone whose career is on an upward trajectory…

    16 条评论
  • Burnt out or just grumpy

    Burnt out or just grumpy

    I was talking with a friend who leads a recruiting company, and we were discussing the reality and the hype of the…

    12 条评论
  • How do you put a price on “trust”

    How do you put a price on “trust”

    In the consulting world, there is an old joke. Two people buy a ship.

    17 条评论
  • What hiring biases do you have?

    What hiring biases do you have?

    As is often the case, one of my conversations in my role as a mentor, sparked this post. We were discussing the…

    3 条评论
  • Is it a bad thing to be described as "super-smart"?

    Is it a bad thing to be described as "super-smart"?

    Are you described by others as being “super-smart”? Is that a bad thing? It may be if that is the first words used to…

    9 条评论

社区洞察

其他会员也浏览了