KPIs, the facile way to manage

by Ross Gittins.  This article was first published in The Age on 18th December 2013.  It adds weight to our contention that Managerialism may be seriously flawed.

You have heard of painting by numbers, but these days the great fad is management by numbers. I call it the metrification of business – although it’s just as prevalent in the public sector. If you know what the initials KPI stand for, you’ll know what I’m talking about.

I’ve been a bean-counter all my working life, first as an accountant, then as an economic journalist. So I’ve long believed in the importance of measurement, of getting the measurement as accurate as possible and of understanding the strengths and weaknesses of particular measures.

Much of my apprenticeship as an economic journalist was spent making sure I understood how all the economic statistics were put together. But as I’ve watched the enthusiasm for ”key performance indicators” and other ”metrics” grow, I’ve become increasingly sceptical about their usefulness.

The rise of metrification has been built on repetition of the seeming truism that what gets measured gets managed. It follows that what isn’t measured isn’t managed.

Advertisement

It’s certainly true that what we measure affects what we do and the way we think about the phenomenon being measured. So in the drive to hold workers accountable – that is, to increase control over them – and improve their performance, it’s become fashionable among managers – private and public sector – to measure key aspects of an organisation’s performance.

It’s only a small step to setting targets for the performance measures and, often, raising those targets from year to year. You might have targets for the performance of organisations, but also for the performance of individuals. And it’s only another small step to link performance to remuneration. We pay for results – what could be fairer?

The rapid escalation of executive remuneration over the past 20 years has occurred not so much because of the rise in basic salaries as the proliferation of ”performance pay”. A study by Mihir Desai, of the Harvard Business School, found that, in America, the proportion of the total pay of senior managers that is based on share prices rose from 20 per cent in 1990 to 70 per cent in 2007.

But, as The Economist reported last year, Desai found this had warped incentives and fostered malfeasance. Managers had won huge payouts simply because the share market had gone up, regardless of whether they had personally added value.

”They have also gamed the system, sometimes illegally, to hit targets that put fat sums in their pockets,” it said.

The trouble with this measurement approach to accountability and reward is that, as the American psychologist Martin Seligman has said, ”If you don’t measure the right thing, you don’t get the right thing.”

When the push for micro-economic reform was at its height, someone got the bright idea that if you calculated and made public the equivalent of key performance indicators for all the many responsibilities of the state governments, you’d encourage them to compete among themselves to improve their standing in the league tables.

The Productivity Commission has now compiled and published these indicators for many years. But someone who should know warned me they’d become completely unreliable. Why? Because managers in each state had manipulated their results to make them look good against the competition.

In 2005 the website Crikey.com.au ran a letter from an anonymous public servant reporting their experience with management by KPI.

”Early in June our manager discovered we were a few percentage points away from meeting operational requirements for the financial year. Rather than explain to his boss that staff cannot perform well when there are continual computer problems and weekly changes in procedures and priorities, he instituted a series of ludicrous schemes to improve the statistics,” the person wrote.

”Any work that was already out of time was placed on the backburner, not to be touched until after July 1, when it would be counted in the next year’s statistics. In other words, work that was overdue would not even be looked at for another fortnight.

”For two days staff did nothing but go through their files searching for cases that could be closed without further action or referred to another area. We achieved absolutely nothing in terms of genuine output for those two days, but our percentage of resolved cases sky-rocketed. We then started on the new work, but only worked on simple cases that could be closed well within the acceptable operational time frame…

”On June 30 our manager proudly announced that we had achieved operational requirements.”

I’ve been around long enough to know that measurement can be a trap. People can be mesmerised by numbers. Because they’re objective, people take them to be infallible. They forget (if they ever knew) the assumptions and other limitations on which they’re based, they take them to be measuring something they’re not and they forget how easily others can manipulate them for their own purposes.

It’s easy to measure quantity, but much harder to measure quality. Most jobs are multi-dimensional, but you can’t have a KPI covering every dimension. In which case, I can always meet my KPIs by cannibalising some dimension – some aspect of quality – not covered by a KPI.

Einstein said ”not everything that counts can be counted”. The modern preoccupation with metrics is an attempt to over-simplify the managerial task by confusing quantity with quality.

Sorry, life wasn’t meant to be that easy.

Read more: http://www.theage.com.au/comment/not-everything-that-counts-can-be-counted-20131217-2zj6e.html#ixzz2nlp2cDOV