VI — Contemporary (2000 – today) · Chapter 26
Knowledge Work and Its Discontents
Why we still can't measure it

In 1999 Peter Drucker published an essay titled 'Knowledge-Worker Productivity: The Biggest Challenge.' The essay argued that the productivity of the manual worker had been the great accomplishment of twentieth-century management — a roughly fifty-fold increase over the century, by Drucker's reckoning — and that the productivity of the knowledge worker would be the great challenge of the twenty-first. He noted that the field was, in 1999, roughly where the field of manual-worker productivity had been in 1900: aware that the question existed, possessed of a few preliminary ideas, but mostly unable to measure or systematically improve the thing it was supposed to be managing. Twenty-five years later, and against substantial competition from other candidates, the knowledge-worker productivity problem remains roughly where Drucker left it. The honest attempts at solutions are interesting. The dishonest attempts are doing real damage.
What Makes Knowledge Work Different
Manual-worker productivity is conceptually simple even when it is operationally difficult. The output is a physical thing — pins, cars, bushels of wheat — that can be counted, weighed, and measured against an input of hours, materials, and capital. Improvement is a matter of getting more output per unit of input, which means time-and-motion study, mechanization, process redesign, and the long Lean lineage of Chapter 23.
Knowledge work breaks this model in three ways. First, the output is often invisible: a decision not made, a customer not lost, an architectural choice that won't pay off for two years. Second, the output is highly variable in quality: a software engineer's commit can range from negative-value (introducing a bug that costs a customer's trust) to civilization-changing (the Linux kernel, the BLAS library), and the difference is rarely visible at the moment of production. Third, the optimal pace of knowledge work is non-monotonic: working harder beyond a moderate point usually produces lower-quality output, not higher, because the work depends on judgment that fatigue degrades faster than it improves throughput. None of this is a reason to abandon the project of measurement. It is a reason to be honest about what is being measured and what is not.
Tom Allen and the Geography of Communication
In the 1970s, Thomas J. Allen, an MIT professor, conducted a multi-year empirical study of engineering teams in defense and R&D laboratories. His finding, now known as the Allen Curve, was that the probability of two engineers communicating about a problem fell off sharply with the physical distance between their workstations — by roughly 10 meters of horizontal separation, the probability had collapsed to near baseline. Engineers on different floors of the same building communicated about as often with each other as with engineers in entirely different cities.
The finding implied that the spatial design of an engineering organization is, materially, part of its productivity. Allen's work informed several decades of office design, including some of the famous research-campus layouts of the 1970s and 1980s (Bell Labs, Xerox PARC, IBM Yorktown Heights). It also informed the move to open-plan offices that became fashionable in the 2000s — though, as later research showed, open-plan offices typically reduce face-to-face communication rather than increase it, because workers cope with the sensory overload by retreating into headphones and asynchronous channels. The Allen Curve was about voluntary cognitive engagement, not raw exposure. Modern remote and hybrid work has reshaped the question without resolving it; the geography of communication still matters, the geography is just more complicated.
DORA and the Software Productivity Question
The most rigorous modern attempt to measure productivity in any knowledge-work domain is the DORA research, conducted by Nicole Forsgren, Jez Humble, and Gene Kim and published in the 2018 book Accelerate. DORA identified four metrics — deployment frequency, lead time for changes, change failure rate, and time to restore service — that consistently distinguished high-performing software organizations from low-performing ones across thousands of survey respondents. The metrics were chosen carefully to focus on the throughput and reliability of the system, not on individual engineers' output, and the research repeatedly emphasizes that the metrics should not be used for individual performance evaluation.
The DORA framework has been validated and refined over multiple subsequent years of research. It has also been widely misapplied. Organizations that adopt the metrics and use them as individual targets typically destroy the productivity they were trying to measure, because engineers respond to the targets by gaming them — splitting deployments to inflate frequency, declaring fewer issues as 'changes' to lower failure rate, narrowing the definition of 'restore' to game time-to-restore. Goodhart's law applies. The metrics are diagnostic when used at the system level and counterproductive when used at the individual level, and the line between is thinner than most organizations want to admit.
SPACE and the Modern Frameworks
The 2021 SPACE framework, also from Forsgren and colleagues, attempts to broaden the measurement of software productivity beyond DORA's four metrics into five dimensions: satisfaction and well-being, performance, activity, communication and collaboration, and efficiency and flow. The argument is that productivity is multi-dimensional and that any single-metric framework will fail to capture it. SPACE recommends that any attempt to measure productivity should sample across at least three of the five dimensions to avoid the gaming-the-metric problem that single-metric frameworks suffer.
The DevEx (Developer Experience) framework, popularized in 2023 by similar researchers, focuses more narrowly on what the developer's work day actually feels like — flow state preservation, feedback-loop speed, cognitive load, friction in the development environment. The thesis is that productive engineers are usually engineers whose tools, processes, and incentive structures make productivity natural. Inverting the question — what makes engineers unproductive? — is often more useful than asking how to extract more productivity from them. The DevEx framing has been picked up across the industry and represents the current frontier of honest knowledge-work productivity research.
The Surveillance Failure Mode
On the dishonest side of the question sits the modern industry of employee-surveillance software — keystroke monitors, screen recorders, mouse-movement trackers, productivity-score dashboards. The category exploded during the 2020-2022 shift to remote work, when many managers, deprived of the physical visibility that had been their main proxy for productivity, reached for software substitutes. The category has continued to grow.
The operational failure mode is consistent. Workers respond to surveillance by performing performative work — moving the mouse, opening windows, generating activity that satisfies the metric without producing value. Real productive work — thinking, reading, talking with colleagues, prototyping in unexpected directions — registers as low activity on the dashboard and gets penalized. Within a year, the workforce has been Goodharted into producing the metric rather than the value, and senior management cannot easily distinguish between teams that have been surveilled into compliance and teams that have been surveilled into theater. The available research, both from the surveillance vendors themselves and from independent researchers, suggests that the surveillance technology rarely produces the productivity gains it promises and routinely produces measurable degradation in employee retention and engagement. It is, in Deming's terms, a Taylor-style intervention applied to work that is structurally unsuited to Taylor-style intervention. It is unlikely to age well.
Drucker was right in 1999 and remains right today. Knowledge-worker productivity is the open frontier of management research, and the field is still in something like the position manual-worker productivity was in around 1910 — possessed of a few useful frameworks, several promising research programs, and a substantial industry of bad ideas dressed up in good vocabulary. The honest practitioner's job is to read DORA, SPACE, and DevEx seriously, to use them at the system level rather than the individual level, to refuse to confuse activity with output, and to keep the worker's actual work environment central to the analysis. The dishonest path — surveillance, score-carding, individual productivity metrics — is faster, easier, and reliably destructive of the very productivity it pretends to manage.
Sources
- 1.Knowledge worker · Wikipedia
- 2.Allen Curve — Tom Allen · Wikipedia
- 3.Accelerate (DORA research) · Wikipedia
- 4.DevOps Research and Assessment · Wikipedia
- 5.Goodhart's law · Wikipedia
- 6.Employee monitoring software · Wikipedia