Academia.eduAcademia.edu

Making Measuring Instruments

T h e A g e o f E c o n o m i c M e a s u r e m e n t A n n u a l S u p p l e m e n t to Volume 33 History of Political Economy Edited b y Judy L. and M a r y S. K l e i n M o r g a n D u k e University Press D u r h a m a n d L o n d o n 200,1 Perspective Making Measuring Instruments M a r y S. M o r g a n In the mid-nineteenth century, economists had many numbers but relatively few measurements; by the mid-twentieth century, they began to take for granted that they had measurements for most of the phenomena about which they theorized. This is the change that marks the impact of the age of measurement, and my interest in this prespective essay is to provide an understanding of what this change entailed. Out of many different economists' individual practical measuring projects around the late nineteenth and early twentieth centuries, there emerged a number of different kinds of measuring instruments, each carefully fashioned for economics. By looking carefully at the materials collected in this volume, we can discern that these instruments can be grouped according to the strategies or recipes that were followed in making such measuring instruments. As a historical development, first the instruments, then the strategies, served to connect the ambitions of economists seeking accurate scientific measures of their world with the practically necessary, sometimes mundane, but never trivial tasks of data collection. Correspondence may be addressed to Mary S. Morgan, Department of Economic History, London School of Economics, Houghton Street, London WC2 2AE, U.K.; e-mail: [email protected]. My thanks go to all the members of and contributors to the measurement workshops at the University of Amsterdam and the London School of Economics over the last few years and to the hard-working contributors to the HOPE workshop and this volume. I thank particularly Marcel Boumans, Harro Maas, and Hasok Chang for guiding my thinking about measurement issues. I acknowledge support from the British Academy for funding this work. 236 Mary S. Morgan Both measuring instruments and measurement strategies are a consequence of the age of measurement, not a prerequisite for it. Making measurements of economic entities requires instruments of a certain kind. The second broad historical thesis suggested by the essays in this volume is that the measuring instruments that economists built in this period were also analytical devices. These instruments embodied frameworks and techniques to turn observations into measurements and to organize the empirical data into particular types of arrangements so that economic phenomena might be arrayed before us. On one view, we can see this as a historical process of establishing facts about phenomena, paralleling the epistemological interests of Bogen and Woodward (1988).' On another view, this process can be seen, ultimately, as one of creating new categories of phenomena, paralleling the historical interests of Theodore Porter ([1994] 1996). I begin in the former tradition and will return to the latter interpretation. Measuring Instruments It is a conventional, and surely uncontested, claim that taking measurements requires instruments of measurement. From our childhood use of fingers and rulers to the technicalities of clocks and Geiger counters, we rely on tools to help us measure our world. Economics is no different, except perhaps in being relatively slow to develop such measuring instruments. The discipline did not develop such devices until the late nineteenth and early twentieth centuries, whereas there had been a dramatic increase in the number and range of measuring devices in the natural sciences in the seventeenth and eighteenth centuries (Heilbron 1990). Economists of the late nineteenth century wanted to measure all sorts of things, from the close-up single choice of an individual to the aggregate price level changes in the economy as a whole. But they had few instruments that would allow them to do these things. Whereas some might suppose that economic measurement is merely counting "what is there," a similarly naive view would have us think that x-ray machines merely look through our flesh to reveal our bones. We don't see a macroeconomy, nor a consumer price index, nor an individual choice decision, hence fashioning measuring instruments in economics has been, in part, 1. Within this framework, the problem of differentiating facts from artifacts, and the role of measuring instruments in this process, are particularlyfinelytreated in Franklin 1986. Making Measuring Instruments 237 a matter of developing ways of observing the economy. It is consistent with this to note, from the history of science, that the design and use of measuring instruments are often aligned to experimental investigations.2 Our analogy continues with the advanced medical scanners, instruments of investigation that are designed and programmed to enable us to "see" certain things inside us by picking them out in a particular way. The ways in which the economic body is investigated and data are collected, categorized, analyzed, reduced, and reassembled amount to a set of experimental interventions—not in the economic process itself, but rather in the information collected from that process. Economic observations must not only be registered but also converted into measurements, and converted in ways which serve particular theoretical or empirical or bureaucratic purposes. All this is to say that modern economics has rulers that are as complicated and as technical in design as those of modern medical science. Economic measurement requires counting and arithmetic on a grand scale according to certain procedures. Of course, those procedures involve bureaucratic regimes, clerical workers, automated information systems, and so forth. But they also require sets of instructions, provided, as Marcel Boumans (1999) has recently suggested, by certain mathematical models, or parts of models, that we can think of as constituting measuring instruments in economics. Using this important insight here, we can interpret the mathematical formulae, statistical formulae, and accounting rules, as well as the rules of data elicitation and manipulation, described in the essays in this volume in terms of measuring instruments. The appropriateness of such a labeling lies in the fitness and fruitfulness of the historical reinterpretations it prompts. The terminology and concept of measuring instrument fit happily onto the work of several late-nineteenth-century masters of measurement. Marcel Boumans (this volume) uses his earlier insight to understand Fisher's index number formulae; Harro Maas (this volume) shows how William Stanley Jevons can be understood as using a "virtual balance"; and Franck Jovanovic and Philippe Le Gall (this volume) recount how 2. See for example, Chang 2001. Heilbron (1990, 6) suggests that the early scientific instruments can be labeled either as "measurers" (for example, the barometer) or as "explorers" (such as the air pump) which "produced artificial phenomena for demonstration, investigation, and measurement." This distinction might be difficult to maintain for instruments made by economists in the age of economic measurement, where the relation between these aspects of measuring instruments is very close. 238 Mary S. Morgan instrument terminology was explicitly used by Lucien March to describe his "business barometers." When such economists wanted to solve a new measurement problem, they adapted, or fashioned anew, an instrument to make those measurements. These measuring instruments can now be grouped and typed according to certain general features. For example, this is a period when many different index number formulae are developed, each one offering a differentiated measuring instrument for a particular purpose. But we can also see that all index numbers follow a generic recipe or strategy of "weighted averages." For another example: we can understand how the umbrella strategy of "using causes to make measurements" can describe the various ways Jevons used assumptions about causes to make measurement possible, although the different ways might be thought of as different measuring instruments. Once such instruments are forged, they tend to be taken for granted, and once strategies are recognized, they provide recipes for new instruments. Each individual instrument has its own history, but the sum total of the age of measurement is the emergence of strategies for making such instruments. Strategies for Making Measurements Using Causes to Bootstrap Measurements One strategy of measurement is to pay attention to causes. Since the 1920s or 1930s, econometrics has focused its energy on building models to represent relations in the economy. It then uses these models to establish which are the relevant causal factors and to measure the relevant strength of those factors in explaining certain economic phenomena represented in the models. This approach relies both on techniques, developed only in the twentieth century, to measure causal relations in statistical data and on the already established measurements of the variables in the relationships (see Morgan 1997). Econometricians practice a kind of indirect measurement of the causal connections in relationships based on their theoretical preferences and what are perceived, at best, as direct measures (or at worst proxies for those) of the elements to be related. But we find the role of causes was rather different in the measuring instruments of earlier times. Whereas twentieth-century econometricians assume measurements to access causes, nineteenth-century economists interested in measuring assumed causes to make measurements. Making Measuring Instruments 239 The problem as understood from Mill's time had been how to deal with multiple causes, some permanent and some accidental, and to pick out the effects of single causes from the multitude. Though it appeared possible in thought experiments to consider man as driven by a single motive of wealth acquisition, attempts to isolate and make measurements of single, even constant, causes in such circumstances were fraught with difficulty. Simple averages, the main statistical technique of the day, were thought to work on the assumption of many small disturbing causes which cancel out in the average, as in measurement errors. Sometimes these averages worked well, as in Jevons's measurement of the wear on coins discussed by Sandra Peart (this volume). Other times, as in Jevons's investigation of poverty, they were inappropriate—in such situations, often found in economics, each item involved in the aggregate is subject to many causes, all of them valid, and none necessarily small or assumed to cancel. We can also see these difficulties of multiple causes in Adam Smith's thinking about the value of money (in Kevin Hoover and Michael Dowell's discussion in this volume) and in Jevons's work on the same topic (as discussed by Hoover and Dowell, Peart, and Maas all in this volume). The solution used by Jevons in his study of gold can be interpreted as assuming a cause (namely, variation in the supply of gold, which alters all prices equally) in order to bootstrap a measure of the change in the value of gold money, whereas later economists might start from measurements of money stock and prices and use these to measure the parameter relating the two as a test of the quantity theory. Jevons assumed that there was a reciprocal relationship between the value of gold money and the general level of prices, that the supply of gold would determine the general level of prices, and that other causes of individual price changes reflected in the data on individual prices were independent and would cancel out. He could then interpret the average change as attributable to variations in gold supply and thus provide a measure of the change in the value of gold: the assumptions about causes provided the basis for the measurement of the element of interest, namely, the change in the value of gold. The kind of circularity inherent here, that of inferring a quantity on the basis of assuming a lawlike relation, is found in many fields, and need not necessarily be dismissed on the basis of its circularity, for as Chang (1995) argues, such circularity can be fruitful in concept building and inferences may mesh with other measurements obtained elsewhere. 240 Mary S. Morgan The circularity would only become dangerous if Jevons's measure of the change in the value of gold was then followed by the kind of work undertaken in modern econometric practice in which such a measurement is used in a test of the quantity theory. It is to Fisher's credit that his work on this problem in the early twentieth century steered away from this measurement circularity in two connected steps: first, he represented the relation between money and prices in an equation of exchange which incorporated two other elements, velocity and trade; and second, he constructed independent measurements of all four elements in the equation. Nevertheless, using causal assumptions to measure money changes remained fraught during the period from Jevons through the twentieth century. Thomas Humphrey's analysis (this volume) of the 1920s debate between Federal Reserve Board economists, who believed in the real bills doctrine of money, and the quantity theorists shows how intimately causes, concept definitions, and measurement structures (including the rules of the institution) are all bound together. The inextricable mixture of definitional questions and the use of causal claims with measurements represented in this episode might be considered common for the period between Jevons's usage of causes in constructing measurements and the methods generally accepted after the 1940s for establishing causal relations (assuming existing measurements of variables) which came with econometrics. Adding with Weights to Make Index Numbers Index number formulae conceived as measuring instruments are based on the strategy of aggregating in a way that allows each individual element to be assigned its due weight in the whole. Such a "weighted average" strategy provides a solution to a general problem in economics, namely that many concepts refer to aggregates of things which may be considered homogeneous in the dimension of prices or money value, but are nonhomogeneous in another dimension, namely amounts consumed or produced. The solution is to use weights to overcome the problem of how to average in a manner that takes account of both amounts and values. The weighted averaging strategy of measurement that forms the basis of index numbers was developed in the late nineteenth century and became well accepted by the early twentieth century, although the problem Making Measuring Instruments 241 of designing good index numbers is far from trivial and vibrant arguments remain over which particular design should be used in which circumstance. For example, the arguments recounted by Spencer Banzhaf (this volume) are concerned with the problem of taking into account a third dimension, namely quality, alongside quantity in the price index. In his account we see arguments from microeconomic theorists that index numbers should measure constant utility equivalents and complaints from aggrieved consumers (during times of price rises) that official indices did not measure price changes accurately because of hidden quality changes. The hopes of both groups were defeated by the practical problems about how to weight items to take account of quality changes— despite the advice of academics on committees of inquiry and the willingness of bureaucrats, who collect the data and construct the measurements, to take these problems seriously. Measurement strategies, and the particular measuring instruments they spawn, often have an inherent design tension between satisfying consistency requirements and providing a measure of a concept that coheres with features of the empirical world. Instruments often embody both theoretical and empirical requirements in their design, creating two forms of internal consistency which ought to be met. Both these may place strain on the design and use of the instrument, a strain exacerbated by the desire for coherence. For example, it is well known that there is no perfect index number, just a very large set of possible ones. Some satisfy particular criteria from economic theory, others satisfy criteria which come from data rules. But the instrument design also needs to mediate between these two levels if it is to provide coherent measurements that link the empirical to the theoretical levels. Boumans's detailed examination of Fisher's work (this volume) shows how, during his work on index number design, Fisher moved from a design focus on theoretical consistency, regarded as mathematical features of the index formulae, to a realization that since index numbers could not meet all such criteria at once, his tests would be better interpreted as ones of coherence—how well a particular index number captured the relevant theoretical concept in a way that minimized distortion of the aspect of the empirical world being measured. As Roy Weintraub (this volume) reminds us, mathematics connects to the other sciences in different ways as its image changes. The mathematical virtues of truth and exactness might sometimes be connected with measurement virtues of accuracy and precision, while at other times they might cut loose from 242 Mary S. Morgan them. Boumans (this volume) uses the history of mathematics to analyze Fisher's changing understanding of measurement: theory dictates the nature of possible measurement assuming the theoretical picture of the world is correct, but that world itself may be differently arranged. A parallel example comes from the early work on the "identification problem" in econometrics. Econometricians of the 1920s and 1930s were adamant that in order to measure the parameters in relationships, not only did the mathematical structure representing those relations need to have certain features, but the statistical data needed to have variability over the same features (see Morgan 1990). The issue of coherence between these two sets of conditions lay at the heart of the 1930s "pitfalls" debate.3 The problem of coherence between theory and the world raises dangers for any measurement strategy. Using Properties of the Balance The strategy of using a balance for measurement purposes should come as no surprise: that is what a balance is for, and as Peart (this volume) shows, Jevons used a physical balance to measure the wear on coins. What is unusual in economics is that the balance is used not as a physical measuring instrument, but as a virtual instrument. We can see this in Maas's analysis (this volume) of Jevons's work both on the value of gold and on the measurement of individual feelings. In the former case, the balance analogy is used in a specific argument to structure the measurement problem and so to provide a solution to the problem of measuring the value of gold: here the measurement involves a weighing up procedure, and an averaging procedure, both of which, in their detail, depend on the properties of the balance. The measurement of feelings applies the idea of the balance to the individual and her perceptions of utility: here the mind is portrayed as the balance, the mind registers both moments of equality and the points at which the balance between two things tips in one direction. Jevons describes a procedure of measuring sufficient for the individual to decide action; there is a judgment of equality or inequality and no numbers are needed. Just as an hourglass can measure the passage of an hour without telling the time of day, the balance 3. See Hendry and Morgan 1995 for extracts from the debate and earlier papers on identification along with a commentary on these matters. The insights of those years were temporarily lost as "identification" came to be understood as a characteristic of the mathematical structure alone. Making Measuring Instruments 243 can judge the comparative heaviness of two things without telling their weights. Accounting, Wholes and Parts Accounting principles provide another generic strategy for economic measurement. Identities, accounting coventions and arithmetic identities defined by theory, hold things together and provide an important rule constraining the measured parts to fit together in a particular way: if two sides of the identity don't equate, something is missing or overcounted. Balancing requirements within accounting, either by the imposition of an equation or by using double-entry systems, provide another constraint available where theoretical structures stress equivalences of inputs and outputs as in the Leontief system (discussed by Martin Kohli in this volume), or the overall total equivalents of expenditure, output and income as in Richard Stone's national income accounts (discussed by Flavio Comim in this volume). These accounting elements provide both a strength and a weakness. Their summing and balancing requirements seem to ensure accuracy, whereas in reality they ensure only a type of precision. Everything that has to balance or add up must do so: precision, but this might be achieved at the cost of only partial coverage of the aggregate: inaccuracy. The strategy rewards consistency at the empirical and theoretical levels but shortcomings of coherence between the two may be neglected. Another well-known advantage of the accounting strategy is that it is flexible to many different theoretical accounts of the economy, to different economic ideologies and to many different levels of application (see Morgan 2002). National income counting predated Keynesian economics and although it later became strongly associated with Keynesian theory, it has not, historically, been tied irredeemably to Keynesian concepts. Input-output accounts can be used within a Marxian economics framework and a planned economy accounting system as well as with a regulated or managed market economy. They can be constructed for fineor coarse-grained measurement and analysis, for use with national income accounts or as part of a national budgeting system. This flexibility points to the fact that whatever accounting scheme is adopted is largely a matter of choice and is historically contingent, determined by the requirements of governments in war, the local economists hired to make 244 Mary S. Morgan the accounts, and so forth. Thus, in the same period, we see the MeadeStone accounts developed to work out wartime expenditures in Britain; Leontiefs input-output accounts used to predict future manpower levels and as a check on the national accounts developed by Simon Kuznets in the United States (see Comim's and Kohli's articles in this volume); and a particular form of national budgeting developed in Norway (see Bjerkholt [1998] 2000). The choice becomes a conventional one once national income accounts have been standardized under the auspices of international economic agencies in the later twentieth century and those agencies used their power to insist on their own measuring instruments. Related to these accounting measures are strategies of measuring wholes and parts. One of the oldest and well-measured wholes is the population, with censuses becoming widely established just after 1800. The population might be considered something like a natural whole: difficult though it might be to count every person, in principle it is relatively easy to define the whole. Sybilla Nikolow's essay (this volume) shows how far early measurers in the German cameralist tradition could emphasize the importance of such population and geographical measures of the whole by using them to work out ratios of economic power and comparing different national results. With the twentieth-century development of macroeconomics, attempts to measure "the economy" taken as a unit whole (rather than an aggregate) proved much more difficult. Conventional definitions tend to cover only things which pass through the market place (including the government sector) and are thus accounted for in monetary terms. But this leaves out all sorts of economic activity ranging from neighborly exchanges to unpaid housework. Immediately, then, the strategy of what counts as part of the whole excludes some things and includes others in the count. The strategy of extending the accounts to wider wholes and breaking them down into narrower categories, or into different sets of parts, is associated with the accounting ideal along with a desire for classification and naming. But each such redefinition of wholes and reclassification of parts results in a complete revision of the measurement task. We can see this equally well in Comim's discussion (this volume) of Stone's desire to extend the accounts while standardizing them and in Kohli's discussion (this volume) of Wassily Leontief's difficulties in stretching the input-output accounts to focus beyond production. Making Measuring Instruments 245 Constructing Indicators The measurement strategy which makes indicators to inscribe the path of the economy over time has often been derided as hopelessly empiricist, subject to "measurement without theory" jibes. Yet the creator of the French economic barometer, Lucien March, considered his task to be the extraction of regularities out of a complicated pattern of information, in much the same way as others had tried to extract measurements on one common cause out of the whole set of interlocking multiple causes. March referred to his indicators as "instruments of observation and analysis" (Jovanovic and Le Gall this volume). He characterized the numbers the barometer inscribed as "conventions," in the same way, we might add, as numbers registered by thermometers are conventions (either degrees Fahrenheit or Celsius), but nevertheless valuable for comparisons, and once the convention is known, for indicating absolutes. Although called "index numbers" by March, these business barometer indicator measures had little in the way of principles dictating their construction compared to the weighted averages measures referred to above. Nor were they like the indicator diagrams produced in engineering and in psychology in which the indicators are a directly inscribed outcome of some physical process or some physiological reaction (see Brain and Wise 1999). The indicator diagrams of economics, seen in the Manchester Statistical Society's pursuit of a "weather glass" to indicate the "dark days" of the trade cycle (see Judy Klein's essay in this volume) and in March's business barometer of the early twentieth century, are both forerunners of the business cycle measurements of the 1930s. These measurements are the results of time-series data, carefully filtered and amalgamated according to different rules. Some rules are conventional (for example, detrending), some rules are driven by theories (for example, about which data series constitute the phenomena), and some are more ad hoc (for example, based on correlation measures). The indicators are then charted so as to trace out the movements of economic activity over time. This is a measurement strategy which depends heavily on quantitative techniques yet is barely bound by rules from accounting, balancing, or weighting conventions and makes little call on principles deriving from theory or knowledge of causes. More than other strategies, this strategy makes measurements without the benefit of any blueprint other than the techniques supplied by earlier makers of business barometers. This is ironic, for this measuring instrument is one of the ' 246 Mary S. Morgan few in economics named after an existing scientific measuring instrument which is theory bound. Though not shaped by strong principles, such time-series indicators have become one of the more well-used measuring instruments in twentieth-century economic life. Just as those who own a barometer read it on their way out the door, business people, investors, government ministers, and their economic advisors check the leading and lagging indicators of the economy to see where it is heading. Social Surveys Social surveys are perhaps the most easily conceived of as a strategy for measurement, for they embed the sampling principles and techniques from statistics alongside an open-ended, flexible observation and counting method which can be applied across a range of topics, depending on the historical context in which this demand for measurement is made and met. As Bradley Bateman (this volume) suggests, there were many examples of survey data being collected in economics in the late nineteenth century—on budgets, on insurance systems, on poverty and wages, and so forth. Though the survey went out of fashion in economics during the mid-twentieth century, replaced by the strategy of causal measurement in econometrics, it has since returned as an essential element of econometrics itself. Much of modern microeconometrics and microsimulation (which perhaps ought to be treated as a measurement strategy on its own) relies on data gained through surveys, and econometric techniques have had to be adapted to the measurements generated by such instruments. The Components of Strategies The development and adoption of strategies of measurement that I describe here may have been overlooked, but they have been no less effective for all that. Historians of economics have focused on individual cases of measurement history, and, taken individually, what we see is the development of specific measuring instruments, each of which might be described as a mixture of the conventional and the ad hoc. It is only when these developments are viewed together that we can discern and describe a history in terms of a few generic types of measuring instruments. The historical change that marks the age of measurement is not therefore to be understood just as a massive increase in the collection Making Measuring Instruments 247 of numbers but as a set of processes that only begin with counting, and then go on to analyze, reduce, aggregate, and reform our measures of the economy in ways unthinkable in the early nineteenth century. It is these processes, taken together, and with hindsight, that reveal themselves as a set of measurement strategies. What do I mean by the term strategy here? It may be defined as a method for making, doing, or accomplishing something. Its synonyms suggest an idea, plan, or design: a pleasing combination of practical and ideal, yet the terms are still too vague to describe what was happening in economic measurement during the period from the late nineteenth to the mid-twentieth century. Each economic measuring instrument can be understood as involving three elements, namely, of principle, of technique, and of judgment; but there is also an overall element of design which I call a strategy. A particular strategy constrains the choices and the combinations of the three elements, and these elements in turn shape the way individual measuring instruments are constructed and so the measurements that are made in economics. For example, we can characterize both accounting principles and the equations we get from economic theory as the kinds of principles or rules which are embodied in measuring instruments. Techniques also come in various forms: techniques for counting and calculating, for cleaning data of outliers and "noise," for choosing samples, for meshing observations, and so forth. Judgment plays a role in many places, but particularly where design decisions have to be made about classification, exclusion, and inclusion. These judgments or decisions are often determined by the purpose for which the measurement is made. But judgment, techniques, and principles are not freely chosen: in the case of each measuring instrument, they hang together within an overall strategy for making instruments of that particular kind. Economists with as diverse aims and epistemological beliefs as Irving Fisher and Lucien March could both, at the beginning of the twentieth century, write seriously about the design of measuring instruments, their construction, and the criteria they should meet. Both indeed constructed "index number" measuring devices, but with very different principles and techniques. Because they were following different strategies, they constructed different kinds of instruments, and made different measurements. In contrast, the individual designs and constructions developed by Stone and Leontief can be recognized in terms of the same generic strategy—although, since Stone and Leontief judged different categories ' : I , , • 248 Mary S. Morgan to be important, their measuring instruments were constructed to measure different things. Fisher, March, Stone, Leontief, and many others shared a strong commitment to the measurement ideal, and their measuring work forms the basis for now well-accepted measuring instruments.4 The Life of Measurements s I Si I have painted an account of measurement strategies as epistemological instruments—to the neglect of the social, economic, and bureaucratic realms in which their history lives (see the essays by Klein and Porter in this volume). Yet the change wrought by the age of measurement is at least as evident in the way that we now habitually accept that the measurements provided by these instruments give us reliable representations of what is happening in the economy, and provide an impetus on which we as individuals, and on which our governments, act. The importance of numbers to social knowledge and action is made most evident in Bradley Bateman's essay. But Porter (this volume) emphasizes another aspect of the importance of numbers and reminds us of one of his cherished themes when he suggests that only in the period after 1900 did economists "systematically construct and analyse new entities through measurement." The measuring instruments discussed here are critical to this, for as Porter ([1994] 1996, 36) argues, "standardized quantitative rules have been almost as fertile as standard experiments and massproduced instruments in the making of new things." Measurement, maybe even more than theory, contributes to the making of new ontic furniture for the economic world. Porter reminds us, in his 1994 essay, that the history of science is riven with examples in which theoretical disagreements over the nature of things or lack of knowledge of causes failed to impede measurement; that quantities often displace concepts; and that quantitative "laws" have often been taken as evidence for the reality of elements in the laws. Though the wealth of the nation or the economic power of the national economy was an early category for discussion, it was only after 1950 that the Gross National Product (GNP) 4. One indication of the importance the profession attaches to these measurement strategies is to note that nine master craftsmen of measurement have been awarded Nobel prizes in economics: for the causal account (econometrics) strategy, by Jan Tinbergen, Ragnar Frisch, Trygve Haavelmo and Lawrence Klein; for the accounting strategy by Simon Kuznets, Richard Stone, and Wassily Leontief; and most recently to James Heckman and Daniel McFadden, who fashioned the econometric techniques which enabled economists to use survey data effectively. Making Measuring Instruments 249 moved from being a category of economic analysis to being a number of everyday concern to government. Its measurement, tied at some point to a particular theory of macroeconomics, broke away from its theoretical roots to become a theory-neutral (though not value-neutral) measure of the economy's progress and health. The monthly change in the consumer price index (CPI), like the daily temperature, is a regularly reported number in our newspapers and influences all sorts of decisions at all sorts of levels. Again, almost regardless of changing knowledge and theories about what causes the CPI to change, it is treated as a real quantity by the participants in an economy. Other categories of measurement are more limited in their circulation, being "real" for certain economists, officials, and businesses, but otherwise remaining technical terms that hardly touch the general consciousness. As we have seen, these "real" quantities are constructed with measuring instruments that have become conventional quantitative tools of analysis yet whose starting point and reasoned development has often since been forgotten. Like thermometers, we need only to know how to use such instruments, not necessarily how and why they work. These measuring instruments report "facts" about phenomena for us, but the phenomena they describe have been constructed on instruments fashioned long ago from the experiments and choices made by scientists in the past (see Chang 2001). There is no "right" set of rules for GNP construction, one that fits naturally to our economic world. Like a new glove, awkward at the start, a new measurement gradually becomes comfortable, though it may never fit tightly. We choose our accounting conventions, and only after the passage of time does a measured concept like the GNP conform to our other knowledge and come to seem "real." If we do our own taxes, we make the same transition: tax authorities' accounting conventions rarely fit each human experience—our experiences have to be made to fit their conventions. This same process happens with economic measurement at other scales and with instruments other than accounting. Economic measuring instruments are not devices given by God for us to reckon his own preordained economy; they are the inventions of economists and we fit ourselves to their measures. As Werner, one of Goethe's fictional characters, remarked of double entry bookkeeping, "It is among the finest inventions of the human mind; every prudent master of a house should introduce it into his economy." The character of Wilhelm replied, "You begin with the form, as if it were 250 Mary S. Morgan matter: you businessmen commonly forget, in your additions and balancings, what the proper total of life is."5 References Bjerkholt, O. [1998] 2000. Interaction between Model Builders and Policy Makers in the Norwegian Tradition. In Empirical Models and Policy-Making: Interaction and Institutions, edited by F. A. G. den Butter and M. S. Morgan. London: Routledge. Bogen James, and James Woodward. 1988. Saving the Phenomena. Philosophical Review 97.3:303-52. Boumans, M. 1999. Representation and Stability in Testing and Measuring Rational Expectations. Journal of Economic Methodology 6:381^-01. Brain, Robert M., and M. Norton Wise. 1999. Muscles and Engines: Indicator Diagrams in Helmholtz's Physiology. In The Science Studies Reader, edited by M. Bagioli. New York: Routledge. Chang, H. 1995. Circularity and Reliabilty in Measurement. Perspectives on Science 3.2:153-72. . 2001. Spirit, Air, and Quicksilver: The Search for the "Real" Scale of Temperature. Historical Studies in the Physical and Biological Sciences (forthcoming). Franklin, A. 1986. The Neglect of Experiment. Cambridge: Cambridge University Press. Heilbron, J. L. 1990. Introductory Essay. In The Quantifying Spirit in the 18th Century, edited by T. Frangsmyr, J. L. Heilbron, and R. E. Rider, 1-24. Berkeley: University of California Press. Hendry, D. F, and M. S. Morgan. 1995. The Foundations of Econometric Analysis. Cambridge: Cambridge University Press. Jackson, M. W. [1994] 1996. Natural and Artificial Budgets: Accounting for Goethe's Economy Of Nature." In Accounting and Science: Natural Inquiry and Commercial Reason, edited by M. Power, 57-81. Cambridge: Cambridge University Press. Morgan, M. S. 1990. The History of Econometric Ideas. Cambridge: Cambridge University Press. . 1997. Searching for Causal Relations in Economic Statistics: Reflections from History. In Causality in Crisis: The New Debate about Causal Structures, edited by V. McKim and S. Turner. South Bend, Ind.: University of Notre Dame Press. 5.1 owe this wonderfully apt interchange to Jackson ([1994] 1996). The source is Goethe's Wilhelm Meisters Lehrjahre of 1795. Making Measuring Instruments 251 . 2002. Economics. In Modern Social Sciences, edited by T. H. Porter and D. Ross. Vol. 7 of The Cambridge History of Science. Cambridge: Cambridge University Press. Porter, T. H. [1994] 1996. Making Things Qualitative. In Accounting and Science: Natural Inquiry and Commercial Reason, edited by M. Power, 36-56. New York: Cambridge University Press.