18
Science and Public Policy April 2005 0302-3427/05/020119-18 US$08.00 Beech Tree Publishing 2005 119 Science and Public Policy, volume 32, number 2, April 2005, pages 119–136, Beech Tree Publishing, 10 Watford Close, Guildford, Surrey GU1 2EP, England Valuing S&T activities Public values and public failure in US science policy Barry Bozeman and Daniel Sarewitz Domestic science policy in the United States is linked inextricably to economic thinking. We seek to develop a practical analytical framework that confronts the manifest problems of eco- nomic valuing for science and technology activi- ties. We argue that pervasive use of market valuation, market-failure assumptions and eco- nomic metaphors shapes the structure of science policy in undesirable ways. In particular, reli- ance on economic reasoning tends to shift the discourse about science policy away from politi- cal questions of “why?” and “to what end?” to economic questions of “how much?” Borrowing from the “public values failure framework”, we examine public values criteria for science policy, illustrated with case vignettes on such topics as genetically modified crops and the market for human organs. Barry Bozeman (corresponding author) is at the School of Pub- lic Policy, Georgia Tech, Atlanta, GA 30332, USA; Tel: +1 404 894 0093; Fax: +1 404 385 0504; E-mail: barry.bozeman@ pubpolicy.gatech.edu. Daniel Sarewitz is at the Consortium for Science, Policy, and Outcomes, Arizona State University, Tempe, Arizona, USA. The authors gratefully acknowledge David Guston’s helpful comments. HE PROCESSES OF economic valuation do not easily accommodate public values (Sen, 1987). However, in part because of their con- creteness, a focus on economic values sometimes displaces consideration of public values in political discourse. In science policy, market-based argu- ments and outcomes dominate political thinking and action, despite the fact that science and technology hugely influence public values in ways that are in- dependent of the marketplace. In this paper, we apply a new analytical and rhe- torical framework — public-failure theory (Boze- man, 2002) — to science policy. Our goal is to make available an intellectual approach that can compete with market logic in thinking about, and making, science policy decisions. We start out by briefly discussing the ways in which marketplace logic has become dominant in science policy. We then discuss the concept of pub- lic values and introduce the theory of public failure as a complement to the idea of market failure. Using brief case discussions, we illustrate six criteria of public failure in science policy. We then present a simple graphical tool for conceptualizing the rela- tionship between market failure and public failure, before concluding with a brief discussion of why public failure might succeed in breaking the hegem- ony of economic rationality in science policy. Background Domestic science policy in the United States is linked inextricably to economic thinking, making it no different from many other policy domains. When Americans consider institutional arrangements for T

Public values and public failure in US science policy

  • Upload
    daniel

  • View
    218

  • Download
    6

Embed Size (px)

Citation preview

Page 1: Public values and public failure in US science policy

Science and Public Policy April 2005 0302-3427/05/020119-18 US$08.00 Beech Tree Publishing 2005 119

Science and Public Policy, volume 32, number 2, April 2005, pages 119–136, Beech Tree Publishing, 10 Watford Close, Guildford, Surrey GU1 2EP, England

Valuing S&T activities

Public values and public failure in US science policy

Barry Bozeman and Daniel Sarewitz

Domestic science policy in the United States is linked inextricably to economic thinking. We seek to develop a practical analytical framework that confronts the manifest problems of eco-nomic valuing for science and technology activi-ties. We argue that pervasive use of market valuation, market-failure assumptions and eco-nomic metaphors shapes the structure of science policy in undesirable ways. In particular, reli-ance on economic reasoning tends to shift the discourse about science policy away from politi-cal questions of “why?” and “to what end?” to economic questions of “how much?” Borrowing from the “public values failure framework”, we examine public values criteria for science policy, illustrated with case vignettes on such topics as genetically modified crops and the market for human organs.

Barry Bozeman (corresponding author) is at the School of Pub-lic Policy, Georgia Tech, Atlanta, GA 30332, USA; Tel: +1 404 894 0093; Fax: +1 404 385 0504; E-mail: barry.bozeman@ pubpolicy.gatech.edu. Daniel Sarewitz is at the Consortium for Science, Policy, and Outcomes, Arizona State University, Tempe, Arizona, USA.

The authors gratefully acknowledge David Guston’s helpful comments.

HE PROCESSES OF economic valuation do not easily accommodate public values (Sen, 1987). However, in part because of their con-

creteness, a focus on economic values sometimes displaces consideration of public values in political discourse. In science policy, market-based argu-ments and outcomes dominate political thinking and action, despite the fact that science and technology hugely influence public values in ways that are in-dependent of the marketplace.

In this paper, we apply a new analytical and rhe-torical framework — public-failure theory (Boze-man, 2002) — to science policy. Our goal is to make available an intellectual approach that can compete with market logic in thinking about, and making, science policy decisions.

We start out by briefly discussing the ways in which marketplace logic has become dominant in science policy. We then discuss the concept of pub-lic values and introduce the theory of public failure as a complement to the idea of market failure. Using brief case discussions, we illustrate six criteria of public failure in science policy. We then present a simple graphical tool for conceptualizing the rela-tionship between market failure and public failure, before concluding with a brief discussion of why public failure might succeed in breaking the hegem-ony of economic rationality in science policy.

Background

Domestic science policy in the United States is linked inextricably to economic thinking, making it no different from many other policy domains. When Americans consider institutional arrangements for

T

Page 2: Public values and public failure in US science policy

Public values and failure in US science policy

120 Science and Public Policy April 2005

delivery of public goods and services, they either begin with systematic economic reasoning or, more often, with less systematic assumptions filtered through the laissez-faire nostrums embedded deeply in US political culture (Lindblom, 2001). When Americans seek to determine the value of resources, goods and services, they reflexively look for price indices, eschewing more complex and indeterminate approaches.

The focus of our paper is on developing a practi-cal analytical and rhetorical framework that con-fronts the manifest problems of economic valuing for science and technology activities. We argue that pervasive use of market valuation, market-failure assumptions and economic metaphors shapes the structure of science policy in undesirable ways. In particular, reliance on economic reasoning tends to shift the discourse about science policy away from political questions of “why?” and “to what end?” to economic questions of “how much?”

If we assume that everyone is made better off by investments in science then the only sensible policy question is “how much science can we afford?” Yet, if we assume that science’s benefits and costs affect citizens in very different ways and to different de-grees, and that those benefits and costs are in turn affected by the composition of society’s science portfolio, then public value questions emerge as at least as important as economic ones.

Critiques of post-World War II US science policy have recognized in its operation the subservience of a broadly construed but difficult to measure public interest to more easily recognizable economic inter-ests (for instance, Dickson, 1984; Winner, 1986; Kleinman, 1993; Cozzens and Woodhouse, 1995; Sarewitz, 1996; Greenberg, 2001). Indeed, a central dilemma of post-War science policy has been this: the value of science is commonly promoted in terms of concrete social outcomes (say, curing a disease),

but the process by which the products of science permeate society are largely through the market-place, and the tools for measuring value are largely economic.

Yet the tension between the public value em-bodied in promise of science and the market value realized through its commercialization is real and pervasive. For example, the doubling of the budget for the National Institutes of Health between 1998 and 2002 was justified by the promise of health benefits for Americans (for instance, FASEB, 2001), even as health care becomes increasingly unafford-able to increasing numbers of citizens, health disparities persist across socioeconomic and ethnic divides, and the public health returns on research investments remain difficult to document (for in-stance, Evans, et al, 1994; Wilkinson, 1996; Calla-han, 1998; Smedley et al, 2002).

The most resilient justification for publicly funded science has been job creation and increased stan-dards of living (for instance, Bush, 1945; Clinton and Gore, 1994; Lindsay, 2001). Yet GDP (gross domestic product) per capita is the coarsest possible proxy for quality of life; indeed the affluent-world experience of the past 30 shows that science- and technology-based economic growth is accompanied by increasing inequality in distribution of economic benefits, including increasing unemployment or un-deremployment, decreasing real wages, increasing wage inequality, and increasing wealth concentra-tion within nations and between nations (for in-stance, Noble, 1995; Sen, 1997; UNDP, 1999; Castells, 1998; Galbraith and Berner, 2001; Arocena and Senker, 2003).

We might imagine that such tensions between public values and economic values in science policy would strengthen political motives to ensure that publicly funded science effectively served public values. Yet scholars in disciplines ranging from eco-nomics of technology to history of science have been documenting a 50-year increase in the influ-ence of the marketplace on the agendas of public science.

To some extent this rising influence can be seen as an explicit transition in national priorities from Cold War competition with the USSR to economic competition with emerging economies such as Ja-pan. Yet even during the Cold War, science policies justified in terms of national defense were tightly coupled to private-sector innovation processes (for instance, Mansfield, 1968; Noble, 1984; Leslie, 1993; Norburg and O’Neill, 1996).

More subtly still, the dynamics of science’s role in economic growth seems to promote very strong coupling between public agendas and private mo-tives. For example, technological frontiers (generally pursued in the private sector) have a strong influence on academic research by revealing new phenomena and problems for scientists to confront (for instance, Rosenberg, 1982, chapter 7; Narin and Olivastro, 1992; Rosenberg and Nelson, 1994; Pavitt, 1998).

Barry Bozeman is Regents Professor of Public Policy andFounding Director of the Research Value Mapping Programat Georgia Tech’s School of Public Policy. His researchfocuses on science and technology policy and management.Current projects examine: distributional equities in scienceand technology outcomes; social dynamics of collaborationin science technology; and impacts of markets on publicdomain science. He received his degree in political sciencefrom Ohio State University in 1973.

Daniel Sarewitz is Professor of Science and Society andDirector of the Consortium for Science, Policy, and Out-comes at Arizona State University. His work focuses onunderstanding the connections between scientific researchand social benefit, and on developing methods and policies to strengthen such connections. His most recent book isLiving with the Genie: Essays on Technology and the Questfor Human Mastery (co-edited with Alan Lightman andChristina Desser; Island Press, 2003) He is also the co-editor of Prediction: Science, Decision-Making, and theFuture of Nature (Island Press, 2000) and the author ofFrontiers of Illusion: Science, Technology, and the Politics ofProgress (Temple University Press, 1996). He received hisPhD in geological sciences from Cornell University in 1986.

Page 3: Public values and public failure in US science policy

Public values and failure in US science policy

Science and Public Policy April 2005 121

Also, a progressive increase in the ratio of private to public funding in the US, combined with chang-ing intellectual property regimes, is continually strengthening the linkages between private-sector priorities and publicly funded science (for instance, Press and Washburn, 2000; Nelson, in press). In some sense this is not surprising. If several decades of scholarship in science and technology studies (STS) have demonstrated the co-evolution of science and society (for instance, see Jasanoff et al, 1995), then in modern liberal democracies the place where we would most expect to see this co-evolution ex-pressed is in the economic realm.

There is nothing inherently wrong with the idea of science as catalyst for economic activity —a healthy economy is essential to the functioning of the mod-ern state. Rather our concern is with science’s insuf-ficiency in meeting public values. As Ezrahi (1990, chapter 11) has provocatively argued, the gradual delegitimation of science as a source of authority leaves the economic role as the only one that no longer demands a defense. All that matters is eco-nomic growth.

It is therefore sufficient to know that, as Hull (1988, page 31) observed, science “works,” and works quite well enough to contribute robustly to economic growth (for instance, Nelson et al, 1967; Rosenbloom and Spencer, 1996; Mowery and Rosenberg, 1989; Griliches, 1995; Denison, 1962). When it comes to economic value, the cultural or political construction of science is neither here nor there. Thus, the last bastion of “exceptionalism” for science may well be its role in economic growth (Bimber and Guston, 1995).

What is ‘science policy?’

Throughout the paper we use the term ‘science policy’ in its broadest and most encompassing sense. Indeed, most ‘science policies’ are actually subsidi-ary to other policies, such as higher-education policy, defense policy, agricultural policy, energy policy, and space policy. For many purposes, distin-guishing among these is vital, but not for our pur-poses. When we use ‘science policy’ we refer to all these efforts to bring technical knowledge to fruition and application.

We do not distinguish among basic research, applied research and development, nor among discovery, invention, and innovation. As important as these distinctions may be for some types of policy decision, our analysis is at the highest level of ag-gregation. We also note that these distinctions are far from clear in many situations.

We are concerned about the values of this enter-prise, writ large. Since all acts of knowledge crea-tion, use and distribution have the potential to have impact on people, all such acts are united in their normative implications if in no other way. Each element of the technical enterprise has implication for public value.

We also note that when we refer to ‘science policy’ we are not concerned with the hundreds of thousands of decisions and the hundreds (at least) of decision processes pertaining to specific research grants for individual scientists. We are concerned with the broad policies and policy processes that guide those micro-level behaviors, policies govern-ing allocation of research funding, and institutional arrangements for the conduct and delivery of re-search. We are also concerned with the mental mod-els and rhetorical constructs that help guide decision making throughout the R&D enterprise.

Public values and public failure

Social theory offers few alternative ways of thinking systematically about science policy. There is no so-cial theory of scientific outcomes; there is no social choice theory for science. There is a market-failure model that tells us conditions under which govern-ment should ‘intervene.’ What is missing in most contemporary rationales for science policy, at least as manifested in actual public policy deliberations, is a sense of the specific ways science does and does not serve public values.

One of the difficulties of previous work in public values and the public interest (for instance, Flath-man, 1966; Held, 1972; Cochran, 1974) is that these key concepts, important as they are, do not easily match with extant institutions and processes. The elegance of market solutions is that to achieve their result we, allegedly, need to do nothing other than restrain barriers to competition. The argument spills over easily into science: in a seminal paper, Michael Polanyi (1962) portrayed science as a self-regulating marketplace of ideas whose efficiency could not be improved, but could only be compromised, by out-side intervention. Such economic metaphors have had a powerful impact on public discourse about science policy (see discussion in chapter 3 of Gus-ton, 2000).

To enact public values, something more than phi-losophy and reticence is required. Thus, we examine not only the possible role of public values in science policy but also identify a set of criteria that can as-sist policy-makers in identifying public values issues

There is nothing inherently wrong with the idea of science as catalyst for economic activity — a healthy economy is essential to the functioning of the modern state: our concern is with science’s insufficiency in meeting public values

Page 4: Public values and public failure in US science policy

Public values and failure in US science policy

122 Science and Public Policy April 2005

in much the same way as they now use market-failure models to identify economic aspects of science policy. We provide an alternative schema, which we call the public-failure model (Bozeman, 2002). This model gives rise to criteria that can allow policy-makers to make rigorous judgments about the public value and distributional implica-tions of science policy, in addition to questions of economic efficiency and economic growth.

The public-failure model depends on the notion that there are such things as ‘public values,’ just as the market-failure model derives from an idealized notion of a market, generally defined in terms of efficiency. The term public value has many mean-ings (for instance, Schubert, 1960; Sorauf, 1957; Fuller, 1964). What we mean by ‘public values’ are those that embody the prerogatives, normative stan-dards, social supports, rights and procedural guaran-tees that a given society aspires to provide to all citizens.

This is not the same as a public good because public values are not goods, either tangible (dish-washers) or less tangible (for instance scientific in-formation). Public values are not Platonic ideals, rather they vary across cultures and time, depending on the common values prized in the culture. Public values can be identified in empirical inquiry into the particular public values of particular cultures, but, even absent an empirical anchor, public value is a useful enabling concept, not unlike that of the per-fectly competitive market.

A key assumption of our paper, and the public-failure model, is that market efficiency and public value are not closely correlated. Public failure can occur with market success, public failure and market failure can occur simultaneously, and in some happy circumstances, public success and market success can coincide.

To illustrate the disjunction between market effi-ciency and public value, we need only return to the case of AIDS drugs, an excellent illustration that market failure and public failure are not the end points of a single dimension. AIDS drugs represent a remarkable market success, where initial public in-vestment in research under conditions of market failure led to private-sector development of effective pharmaceutical interventions that also generate con-siderable profit. However, the vast majority of HIV and AIDS sufferers worldwide do not have access to these expensive drugs.

It is possible, that is, to have governments inter-vene effectively through R&D to correct a market failure, and to have markets operate with great effi-ciency once the market failure has been addressed, and yet still have unconscionable failure of public values. The case of AIDs medicines is an illustration of the constricting moral and operational knots we tie ourselves in when we have no frameworks or criteria to compete with market efficiency as a guide to science policy and its concomitant social outcomes.

An introduction to public-failure theory

Policy-makers of every ideological stripe have long been comfortable with the idea that the private sec-tor will never make sufficient investments in basic scientific research to generate the necessary knowl-edge to support robust innovation and economic growth. This market-failure argument has long been accepted as good sense (for instance, Bush, 1945), but since World War II it also been vindicated by economic scholarship (for instance, Nelson, 1959; Mansfield, 1980).

In this paper, we present an analytical and rhetori-cal counterbalance to the economic rationality that has so successfully justified and vindicated science policies for the past 50 years. Our approach, follow-ing Bozeman (2002), is to develop a framework that is conceptually symmetrical to the market-failure rationale used to justify government investment in science. We call this framework “public-failure theory.”

‘Public failure’ occurs when neither the market nor public sector provides goods and services re-quired to achieve core public values. A public-failure approach changes the discussion of public policy by making government (and public values) something other than a residual category or an issue of technical efficiency in pricing structures. With the public-failure model, the key policy question be-comes: “Even if the market is efficient, is there nonetheless a failure to provide an essential public value?”

Using a market-failure model, government in-volvement is thus justified only when market pro-cesses fail and “prices lie — that is, when the prices of goods and services give false signals about their real value, confounding the communication between consumers and producers” (Donahue, 1991, page 18). The causes of such failure are oft-articulated and well-understood: externalities; steep transaction costs; distortion of information or inhibition of in-formation flow about a good or service; and mo-nopolistic behavior or other competitive failures. (For a precise definition of market failure, see Ban-nock et al, 1998, page 117.) A public-failure model eschews assumptions about the correspondence of efficient markets and desirable social outcomes and focuses instead on social values criteria.

With respect to science policy, economics is about the private value of public things. Despite the rela-tive rigors of market-failure theory, no economic analysis can encompass the full range of policy choices facing society. Efficient pricing cannot solve all problems of public values, and indeed, there is a strong societal sense that, in some domains, economically optimal solutions are entirely inappro-priate, for example, in creating a market for blood or organ donors. Some political actions explicitly eschew economic efficiency in favor of social equity, for example, the Americans with Disabilities Act. Yet economic approaches have been applied,

Page 5: Public values and public failure in US science policy

Public values and failure in US science policy

Science and Public Policy April 2005 123

often controversially, in domains that encompass public values, such as in the valuation of natural re-sources, the creation of markets for pollution-permit trading, and the design of science and technology programs.

Even if we accept the market-failure rationale on its own terms, it is unavoidably incomplete, for two rather obvious reasons. First, just because markets fail, it does not mean that government action can eliminate or avoid the causes of failure. For exam-ple, environmental externalities may signal market failure, but a government decision to respond to these externalities is not an a priori demonstration that doing so will lead to more social benefit. While the US government invested billions in collaboration with the big three automakers in the Partners for a New Generation of Vehicles program, Japanese automakers on their own produced a first generation of hybrid-electric automobiles that outstripped any-thing the Americans had to offer.

Second, the absence of market failure does not a priori imply public success, and therefore is not a sufficient reason to eschew government investment. That is, market value does not equal public value. The situation with AIDS drugs in the developing world is archetypal: the market has succeeded and millions wait to die.

As obvious as these arguments may be, they are insufficiently available for purposes of public policy making. The aim of public-failure theory is to provide an analytical framework that can be set alongside market failure as an alternative set of policy-making criteria. The notion of public failure derives from the reality that sometimes neither the market nor the public sector is providing goods and services neces-sary to achieve certain core public values.

A public-failure approach to policy making changes the terms of debate by making government (and public values) something other than a subsidi-ary issue of efficiency in market performance. The key question in market-failure rhetoric is: “are prices distorted due to a failure of communication between consumers and producers?” The key question in public failure goes an essential step further: “regard-less of market efficiency, is there nonetheless a fail-ure to provide an essential public value?”

We recognize, of course, that ‘prices’ are more tangible, or at least quantifiable, than ‘public values’ as a unit of analysis. However, there is very little, if any, fundamental disagreement in the United States about the existence of a fairly comprehensive set of core public values, especially those embodied in the nation’s founding documents, such as the right to subsistence, the rule by consent of the governed, freedom of speech and religious practice, and habeas corpus. This basis is more than sufficient for us to proceed, because public-failure theory is not a decision-making tool (à la cost–benefit analysis), but rather a framework (à la market failure) to pro-mote rigorous deliberation about the relation be-tween economic value and public value.

Public-failure criteria

Public failure occurs when core public values are not reflected in social relations. Bozeman (2002) eluci-dates criteria for identifying public-values failure, criteria that to some extent mirror market-failure criteria, but that focus on public values rather than efficiency of market transactions. In applying this framework to science policy, we identify six public-failure criteria:

• inadequate values articulation • imperfect monopolies • scarcity of providers • short time horizons • non-substitutability of resources • benefit hoarding

While we do not claim this set of public value cri-teria to be canonical, it provides a starting point to enhance discourse and decisions for the allocation of responsibilities between public and private sectors, and for shaping allocations within the public sector. Table 1 (adapted from Bozeman, 2002) provides diagnostic criteria of the public-failure model, illus-trated by examples specific to science policy. In the next section, we discuss in greater detail specific illustrations of each criteria, to show the applicabil-ity of public-failure theory to real-world science-policy dilemmas.

Cases in public failure and science policy

Case 1. Public failure in values articulation and aggregation: priority setting in science

Public failure can occur when expression of public values is stifled or distorted. For example, if cam-paign financing procedures lead to conspicuous con-flict between public values (as elicited, say, through polling) and the values of elected officials, then there is a potential for public failure. This type of conflict may provide an incentive for private in-vestment in lobbying that is rational by economic standards yet counter to the larger public interest — as when efforts by the insurance industry and its allies to prevent health-care reform have overcome a broader public desire for a more affordable and equi-table health-care delivery system.

In science policy, a pervasive cause of values ar-ticulation public failures is an absence of mecha-nisms that allow non-scientists to have a significant say about public investments and priorities in most areas of science. The reasons for this failure are clear: How can a non-scientist be expected to make a sensible choice between, say, funding two distinct areas of health research, for instance, research aimed at understanding the contribution of polycystic ovar-ian syndrome, versus research on understanding the impact of bovine hormone in milk on breast cancer?

Page 6: Public values and public failure in US science policy

Public values and failure in US science policy

1

It would take a great deal of effort even for most scientists to have an informed opinion about the largely unknowable trade-off between two important lines of applied medical research. Moreover, politicians (and the general public) are as likely as scientists to understand their own preferences for, say, favorable outcomes in breast cancer research vs favorable outcomes in astronomy or in polar sci-ences or in mental health. This is not an overestima-tion of the ability of the ‘median voter’ to understand science trade-offs but, rather, a perhaps more realistic (and pessimistic) estimate of the rela-tive ability of experts to make high-level values choices (for discussions of basic questions involved in scientific choice, see, for instance, Fuller, 2000).

Science policy-makers have evinced some con-cern about the balance between scientific values (for instance, what constitutes ‘good’ science?) and so-cial values (for instance, how is the public good

being advanced?). In response to Congressional prodding, the National Science Foundation (NSF) has sought to use the peer-review process to enhance the societal value of its research portfolio. In addi-tion to standard criteria of scientific merit, NSF added in 1997 the criterion of social benefit to its peer review process.

NSF then commissioned the National Academy of Public Administration (NAPA) to evaluate how well this effort to incorporate public values into peer review was working. NAPA (2001) reported back that the changes, as implemented, were unlikely to have much positive effect. Problems with NSF’s approach ranged from a lack of “quantitative measures and performance indicators to track the objectives of the new merit review criteria” (page 7), to skepticism or even outright opposition on the part of reviewers to the inclusion of social impact criteria to begin with. NAPA went on to recommend a variety of actions that NSF could take to correct these problems, such as improving “the conceptual clarity of the objectives of the new criteria” (page 8), and ensuring “genuine attention to the goals of the new criteria throughout the entire review cycle” (page 9).

Yet neither the NSF social impacts criteria nor the NAPA report addressed the underlying source of public failure. Nobody denies that the scientific community has great skill in assessing technical quality of research, but who has vested it with spe-cial training, skill, or legitimacy in assessing the so-cial value of research? Moreover, there is no particular reason to believe that the social priorities

T

P

Mva

I

S

S

Scr

B

In science policy, a pervasive cause of values articulation public failures is an absence of mechanisms that allow non-scientists to have a significant say about public investments and priorities in most areas of science

able 1. Public failure and public policy: a general diagnostic model

ublic failure Failure definition Science policy illustration

echanisms for alues articulation nd aggregation

Political processes and social cohesion insufficient to ensure effective communication and processing of public values

Peer review, the favored means of making decisions of individual-level projects, is appropriated for decisions about huge scientific programs, resulting in the displacement of social goals for more easily resolved technical goals

mperfect monopolies Private provision of goods and services permitted even though Government monopoly deemed in the public interest

When public authorities abrogate their responsibility for overseeing public safety in clinical trials for medical research, there is potential for violation of public trust and public value

carcity of providers Despite the recognition of a public value and agreement on the public provision of goods and services, they are not provided because of the unavailability of providers

The premature privatization of the Landsat program shows that a scarcity of providers can create a public failure potentially remediable by Government action

hort time horizon A short-term time horizon is employed when a longer-term view shows that a set of actions is counter to public value

Policy for energy R&D, by considering the short term, fails to fully capture the costs of global climate change on future generations

ubstitutability vs onservation of esources

Policies focus substitutability (or indemnification) even in cases when there is no satisfactory substitute

‘No-net-loss’ policies fail to take into account the non-substitutability of many natural organisms ranging from wetlands protection to prohibiting the sale of human organs on the open market

enefit hoarding Public commodities and services have been captured A prime technical success of genetic engineering, the

24 Science and Public Policy April 2005

by individuals or groups, limiting distribution to the population

‘terminator gene,’ proves an excellent means of enhancing the efficiency of agricultural markets, to the detriment of millions of subsistence farmers throughout the world

Page 7: Public values and public failure in US science policy

Public values and failure in US science policy

Science and Public Policy April 2005 125

of scientists are representative of society. Indeed, surveys of scientists’ political opinions and values suggest there are often large differences between scientists and the general public (for instance, Bauer et al, 2000; Plutzer et al, 1998; Barke and Jenkins-Smith, 1993), though scientists are closer to other elite groups such as journalists.

Just as with other professions and demographic groups, scientists’ values probably resemble those of persons who have socioeconomic attributes and life experiences similar to their own. Also, citizens have not voted for scientists or in any way designated them as the ‘official’ articulators or judges of public value. Finally, the public value of science is rarely ascertainable at the individual project level, where peer review most often operates.

Public values operate at a level where science pol-icy receives very little attention: cross-science com-parisons, and the opportunity costs associated with resource allocation decisions. It is also at this level of broad preferences where the general public, or at least the attentive public, may plausibly contribute (for instance, Carnegie Commission), but the struc-ture of US science policy provides little opportunity to do so. Congress, because of its jurisdictional structure, is generally unwilling to make either cross-science choices or systematic choices among the public values that science serves. Likewise, the science bureaucracy works within scientific fields, disciplines or objectives, rarely among them.

Scientists themselves have even less incentive for the internecine warfare that would arise with a more systematic assessment of the values associated with a diverse set of desired scientific outcomes, thus militating against even expert-driven processes of choosing among disciplines (for instance, as origi-nally suggested by Weinberg (1963), and revisited by Fuller (2000)). This is a built-in public failure affording limited opportunity for values articulation (see Kitcher (2001) for a formal argument about en-hancing the public role in articulating public values for science).

When the putatively ‘value neutral’ science policy funding machine is temporarily thrown off kilter, it is generally because an issue emerges that grips the public imagination, or a highly motivated special interest group, to such a degree that public value issues simply cannot be ignored. The most recent case in point is the stem cell controversy, which provides an excellent proof that it is not the com-plexity of scientific issues that forestalls public par-ticipation and public values articulation. The scientific issues in the various stem cell debates are technical and esoteric, but the values issues are so fundamental and compelling that research cannot proceed apace (for instance, Fukuyama, 2002).

When the conduct of research requires one to con-sider such issues as “what constitutes a human be-ing?”, the momentum of the science policy funding machine slows down and the role of public values is brought to the fore. The results of the stem cell

controversy highlight the dearth of institutions and analytical frameworks available for values articula-tion and analysis.

Case 2. Public failure and ‘imperfect monopolies’: clinical trials

Whereas private-sector monopoly is an indication of market failure, in some cases the inability of a gov-ernment activity to protect its monopoly may lead to erosion of a public value. For example, foreign pol-icy is a legitimate government monopoly, and any competition from unauthorized envoys could damage the broader public interest.

Regulation of private-sector activities to protect public welfare is a widely accepted role of govern-ment, although the appropriate degree of regulation is often highly contentious. In the area of clinical trials to determine the efficacy of pharmaceuticals and other medical interventions, the government has granted research institutions and scientists consider-able autonomy. This autonomy is justified by faith in the self-policing capacity of the scientific commu-nity, especially as embodied in its claim to objectiv-ity — or at least disinterest — through the scientific method, and to quality control through peer review.

The protection of humans in scientific experi-ments is a well-established public value, enshrined in international law through the Nuremburg Code and Helsinki Declaration (Woodward, 1999), and nationally through such codified principles as in-formed prior consent. All experiments involving humans that are funded, in whole or part, with fed-eral dollars, are overseen by Institutional Review Boards — decentralized, self-policing oversight bodies aimed at “protecting the rights and welfare of human subjects of research” (CFR (Code of Federal Regulations, 1993) Title 45, Part 46.103) and ensur-ing that “[r]isks to the subjects are reasonable in re-lation to anticipated benefits” (CFR 45: 46.111). Such experiments include tests and trials to demon-strate the efficacy of new drugs, therapies, and pro-cedures, both as part of the process of gaining government approval for general use, and as a means of informing physicians about the relative value of available options.

In September 1999, 18-year-old Jesse Gelsinger died while undergoing gene therapy for a rare liver disease. Gene therapy had long been touted as a po-tentially miraculous emerging line of treatment for a wide variety of serious genetic disorders, but its promise had remained unfulfilled, and Gelsinger’s death made national news. Early reporting on his death suggested only that something had gone terri-bly wrong in the experiment, but that all appropriate processes and procedures had been followed to en-sure that risk was minimized and his participation was fully consensual (Wade, 1999).

Deeper investigations revealed irregularities. The consent forms that Gelsinger signed misrepresented the dosages that were to be administered, and did not

Page 8: Public values and public failure in US science policy

Public values and failure in US science policy

126 Science and Public Policy April 2005

include information about animal deaths from simi-lar treatments. Evidence of high toxicity and adverse side effects in earlier experiments was ignored. The doctor in charge of the experiments, and the univer-sity he worked for, had a financial stake in the com-pany that would have produced the new gene therapy (Nelson and Weiss, 1999).

Before Gelsinger’s death grabbed the headlines, academic studies of clinical trials had been painting a more dispassionate, less publicized picture of pub-lic failure. A number of studies revealed that clinical trials directly or indirectly supported by pharmaceu-tical companies often yielded more favorable as-sessments of new therapies than trials that were not tied to the private sector in any way. In one analysis, only 5% of company-sponsored studies on anti-cancer drugs yielded unfavorable assessments, while, for studies sponsored by non-profits, the unfavorable rate was 38% (Friedberg et al, 1999).

An investigation of calcium-channel antagonists, a class of drug used to treat cardiovascular disease, demonstrated “a strong association between authors’ published positions on the safety of calcium-channel antagonists and their financial relationships with pharmaceutical manufacturers” (Stelfox et al, 1998, page 101). An analysis of published symposia pro-ceedings showed that “[a]rticles with drug company support are more likely than articles without drug company support to have outcomes favoring the drug of interest” (Cho and Bero, 1996, page 485).

Few have argued that such results demonstrate scientific fraud. More likely, “[c]lose and remunera-tive collaboration with a company naturally creates goodwill [that] can subtly influence scientific judg-ment in ways that may be difficult to discern” (An-gell, 2000, page 1517). Some scientists have publicly claimed that they are not subject to such influences. James Wilson, the researcher in charge of the Gelsinger trial, said: “To suggest that I acted or was influenced by money is really offensive to me … You’ve got to be on the cutting edge and take risks if you’re going to stay on top [scientifically]” (Nelson and Weiss, 1999, page A1)

While such claims to immunity from human weakness may or may not ring true, researchers’ ties to industry by definition create a conflict of interest, which, if not revealed to patients in trials, under-mines the principle of informed consent, and, if not apparent to peer reviewers and publishers of re-search articles, can obscure the implications of research results. Bodenheimer (2000) catalogued a variety of ways in which a drug test can be carried out to favor one result or another without rendering the data itself invalid. For example, “[i]f a drug is tested in a healthier population … than the popula-tion that will actually receive the drug, a trial may find that the drug relieves symptoms and creates fewer adverse effects than will actually be the case” (Bodenheimer, 2000, page 1541).

In terms of public-failure theory, a particularly troublesome attribute of this problem lies in the

difficulty of actually documenting the threat to pub-lic values. The Gelsinger story was atypical in that the connection between the conflict of interest and the public consequence — Gelsinger’s death — was obvious (even if not demonstrably causal). However, when published studies comparing one drug to an-other influence a physician to prescribe one drug rather than another, the very existence of the public failure may be difficult to ascertain, and the public-failure consequences highly diffused. It is, indeed, a testimony to the transparency of the biomedical and medical–legal research enterprises that this problem did emerge.

Federal regulations for oversight of human sub-jects research do not explicitly require Institutional Review Boards to consider conflict of interest in the approval process (CFR 45:46), although the Food and Drug Administration, for example, requires that applications for drug approval must be accompanied by disclosure of investigator conflicts in all research supporting the application. The final report of the now-defunct National Bioethics Advisory Commis-sion (2000) recommended that the Government de-velop specific guidelines for defining and regulating conflict of interest in human-subjects research, and that conflict of interest should be disclosed to re-search participants as part of the prior consent pro-cess. Enforcement of these recommendations would help to reestablish the Government monopoly over protecting human subjects of medical research, and thus help to reverse a case of public failure in sci-ence (see Goldner (2000) for a comprehensive dis-cussion of conflict of interest in biomedical research).

Case 3. Public failure due to scarcity of providers: geographic information

Protection of a core public value may depend on the presence of a sufficient number of providers of that value. If market signals are insufficient to attract the necessary number of providers, and if the govern-ment fails to step in, then public failure may occur. Few would disagree that the number of high quality public school1 teachers is less than optimal, due to many factors including relatively low salaries and

When published studies comparing one drug to another influence a physician to prescribe one drug rather than another, the very existence of the public failure may be difficult to ascertain, and the public-failure consequences highly diffused

Page 9: Public values and public failure in US science policy

Public values and failure in US science policy

Science and Public Policy April 2005 127

other disincentives. This may be counted a public failure. When certain government activities are de-regulated, provider scarcity may follow. For exam-ple, when airline deregulation leads to decreased services for rural areas, significant portions of the population may be adversely affected. The market may be operating efficiently, but public failure has occurred.

The federal Government has long been recognized as the appropriate source of support for developing and disseminating data on the geographic and physi-ographic characteristics of the nation. The Lewis and Clark expedition was a famous early example of a Government research project aimed at garnering geographic information, and surveying and mapping activities were early mainstays of federal support for science prior to World War I (for instance, Dupree, 1986).

In recent decades, the importance of geographi-cally referenced, or geospatial, data has increased rapidly. This growth has been fueled by new technologies, from remote sensing and geographic positioning satellites to sophisticated computer graphics software, and also by societal demand for new capabilities to monitor and address complex challenges, ranging from environmental protection to emergency management.

At the same time, private-sector involvement in both the collection and use of geospatial data has increased, leading to a number of dilemmas regard-ing the appropriate allocation of public and private activities. Considerable attention has been focused on the need to ensure that this increasing private-sector role does not result in erosion of public access to information and products that are recognized as public goods.

The revolution in geographic information began with NASA’s launching of the first civilian remote sensing satellite — Landsat 1 — in 1972. Efforts to gradually privatize the Landsat program were initi-ated in 1979 by the Carter administration. Two years later, the Reagan administration began advocating a more rapid shift to privatization, which in turn led to legislation in 1984 (P.L. 98-365) that privatized the sale of Landsat data, and encouraged private-sector development of future satellites.

These actions took place despite studies indicating that privatization was not yet economically sustain-able. Indeed, for the next five years, the Landsat program was in a state of constant fiscal crisis. By the early 1990s, with the operational satellites (Landsats 4 and 5) beyond their design life, and no concrete plans for replacing them either with public or private satellites, the very existence of the Land-sat program was in jeopardy (NASA, 1998; NRC, 1991).

In 1992, Congress took action to ensure the con-tinued provision of satellite-based geospatial data. In doing so, it explicitly noted that the privatization effort interfered with the provision of public goods: “The cost of Landsat data has impeded the use of

such data for scientific purposes, such as for global environmental change research, as well as for other public-sector applications” (H.R. 6133). A new law was enacted (P.L. 102-555) to ensure the continuity of the Landsat program, but also to ensure that pub-licly supported scientists and others who depend on satellite imagery for non-commercial uses would have access both to archived data and to newly ac-quired data — access that had been compromised by high prices during the privatization effort.

The near debacle created by the premature priva-tization of Landsat demonstrates how a scarcity of providers can deprive society of a public good upon which it depends, and how such public failure can be corrected by appropriate government action. In this case, the obvious failure of a putative market solu-tion — privatizing Landsat — made it easy to rec-ognize the public failure. That is, public failure and market failure existed hand-in-hand.

The more interesting and problematic case, how-ever, occurs when the markets are functioning well, but the provision of the public good is not automati-cally preserved. Recent development of geospatial data policy exemplifies an awareness of this tension, and is in fact something of a success story: a case where policy intervention ensures that market success is combined with the support of public values.

As we have discussed, overcoming public failure depends on general agreement about the desirability of a particular public value, and indeed the idea of geospatial data as a public good is well accepted. For example, a study by the National Academy of Public Administration states that “[m]any believe that [geospatial] data should be made widely avail-able at no cost or at reasonable cost to the user, and that this will satisfy an almost infinite variety of governmental, commercial, and societal needs” (NAPA, 1998, page 2).

One National Research Council (NRC) committee asserted that “it is in the public interest for govern-ment to play a leading and facilitating role in coordi-nating the development of spatial data and to make those data available for public use and exchange” (NRC, 1995, page 1). Another NRC committee made the even more specific claim “that it is in the public interest and a federal responsibility for the … development of an interdisciplinary, multidatabase architecture that will allow disparate databases to become nondestructively interoperable in a common geospatial context” (NRC, 2001, page 77).

While such language cannot easily be derived from the Constitution, it can nonetheless be justified by “countless applications (for instance, facility management, real estate transactions, taxation, land-use planning, transportation, emergency services, environmental assessment and monitoring, and re-search)” (NRC, 1993, page 2), and the consequent public benefits that access to geospatial data can confer.

The major obstacle to ensuring such benefits has become the coordination of rapidly expanding

Page 10: Public values and public failure in US science policy

Public values and failure in US science policy

128 Science and Public Policy April 2005

private- and public-sector capabilities in acquiring, processing, and disseminating a wide variety of geographic information. The transition from analog (‘paper’) maps and photos to digital databases has enabled a thriving private-sector effort to apply spa-tial data to a diversity of public and private needs. In particular, the rise of ‘geographic information sys-tems’ (GIS) has created the capability of bringing together very different types of data to support deci-sion making. The challenge of assuring that data, software, and hardware capabilities arising from a multitude of providers did not create a sort of geo-graphic information tower of Babel even led to the formation, in 1994, of a non-profit organization, the Open GIS Consortium, “to address the lack of inter-operability among systems that process georefer-enced data” (Open GIS Consortium, 1999, page 2).

The situation had rapidly changed from one of a scarcity of providers of a single type of data — satellite imagery — to a scarcity of providers of an integrated product. The point is worth emphasizing because it illustrates the subtlety and power of pub-lic-failure theory: in the first case, the public failure of provider scarcity correlated with market failure; in the second, the correlation was with market suc-cess. A new need rapidly arose: to ensure “a com-mon spatial data foundation organized according to widely accepted layers and scales (or resolution) that is available for the entire area of geographic cover-age … to which other geospatial data can be easily referenced” (NAPA, 1998).

For example, if a municipality needed to develop geospatial data to support ecosystem management, it might require spatially referenced data about the loca-tion of wetlands and other sensitive areas, about demographic and land-use trends, groundwater chem-istry, surface water flow, sources of pollution, distri-bution of animal and plant species, power lines and pipelines, and of course traditional physiographic data. For these data to be useful, they must be com-bined as part of a single geospatial database, which means they must be available in compatible formats and coverage, and usable with one software package, on one computer. Such compatibility was not arising from private-sector providers acting individually to maximize profit and capture market share.

The need for Government intervention has been broadly accepted and recognized in both the public and private sectors. In 1994, President Clinton is-sued Executive Order 12906, “Coordinating geo-graphic data acquisition and access: the national spatial data infrastructure [NSDI]” (page 17671), to establish:

1) a Federal Geographic Data Committee to coordi-nate the development of the NSDI;

2) a national geospatial electronic data clearinghouse that would encompass all data collected by the public sector;

3) a process for developing standards to ensure com-patibility among public, private, and non-profit

profit sector sources of geospatial data; and 4) a framework of basic geospatial data — “data you

can trust” — for a variety of applications that would be available for all.

The framework represented a clear embodiment of public action to protect a public good: “basic geo-graphic data in a common format and an accessible environment that anyone can use and to which any-one can contribute … a nationwide community for data sharing” (Federal Geographic Data Committee, 2001).

Case 4. Public failures and short time horizons: en-ergy R&D

Human beings pay attention to unborn generations, but they do not do so out of economic rationality. Pricing will not account for consequences that are expected to emerge in the distant future. Thus, there is clearly a public role in guaranteeing the long-term perspective even if there is no short-term market failure.

This type of problem has emerged most conspicu-ously in the area of environmental protection. For example, the price of gasoline in the past did not reflect the public health costs associated with high levels of lead in the air; currently gas prices do not account for the long-term global environmental costs associated with climate change. In the case of lead, Government action created a public success — man-datory introduction of lead-free gasolines — in the absence of market signals. For climate change, the necessary regulatory and R&D investments have yet to be made, so here we see both public failure and market failure arising from short time horizons.

The market-failure paradigm has provided a po-litically robust rationale for long-term investment in research where no foreseeable application exists. Oddly enough, the paradigm has been less success-ful as a justification for public research investment where the long-term application is clear, but the short-term incentives for private-sector involvement are weak. This irony reflects the apparent repug-nance in market-failure dogma to choose ‘winners and losers,’ or at least to significantly alter the bal-ance between current winners and losers. The consequences are starkly illustrated in the case of energy R&D.

The energy crises of the 1970s demonstrated that the long-term US dependence on foreign sources of oil could have far-reaching economic and political consequences. More recently, the growing aware-ness of the connections between fossil fuel use and global climate change have created an additional long-term rationale to switch to other types of less-polluting energy technologies. In the face of these two realities, trends in both public and private in-vestment in energy R&D are quite amazing: they have declined by almost two-thirds, in real dollars, since the late 1970s (Dooley, 1999, Figure 4).

Page 11: Public values and public failure in US science policy

Public values and failure in US science policy

Science and Public Policy April 2005 129

The reasons for this lack of investment are clear: over short time horizons, the market has been work-ing quite well. In particular, and contrary to general expectations, oil prices have been fairly stable as a result of aggressive exploration, enhanced extraction technologies, open global markets, and the willing-ness of the US to intervene militarily to protect ac-cess to oil supplies. In market-failure thinking, continued low energy prices justify neither aggres-sive public funding of energy R&D, nor Govern-ment regulatory action to promote efficiency, which could stimulate private-sector R&D. Absent the price pressures of oil embargoes, there has been little motive to innovate. Annual energy technology pat-ents, for example, have been declining since the 1980s (Margolis and Kammen, 1999).

Reduced commitment to research seems to be re-flected in both energy production and consumption

trends. In 1970, fossil fuels accounted for 93% of all

US energy production. This proportion declined 5%

over the next five years (during the first energy crisis)

but only 8% more over the succeeding 15 years. (En-ergy Information Administration, 2000). On the con-sumption end, energy intensity (energy use per unit of

economic output) declined 2.4% per year in the 1980s;

in the 1990s, the average annual decline was only

1.5% (Energy Information Administration, 2001). Given the ongoing volatility of politics in the

Middle East, and the increasing evidence that carbon emissions influence the behavior of the global cli-mate, it is difficult not to see the declining public investment in research on alternative energy sources and more efficient energy technologies as deeply problematic. This declining investment documents a continued failure to overcome the short-term think-ing in Government energy R&D policy that is made possible by a well-functioning market that continues to deliver stable energy supplies and low prices.

Case 5. Over-reliance on substitutability of resources: organ sales as a medical public failure

Market mechanisms may indemnify against the loss of particular resources, or offer substitutes for lost resources. While such mechanisms may be efficient

from a market perspective, they may also represent public failure. An obvious example is the calculation that automobile manufacturers might use when de-termining how safe to make a vehicle. Part of this calculation includes the price of adequate indemnifi-cation against law suits for wrongful death.

While such trade-offs are unavoidable (a totally safe car would be either unaffordable, or immobile), they may still represent public failure, for example, if the manufacturer determines that the costs of fix-ing a known problem exceed the expected legal costs, as occurred when Ford failed to correct the exploding gas tanks in its Pinto model (Tietz, 1993). The idea that life is explicitly substitutable offends sensibilities of most non-economists, and may often imply public failure.

A related example comes from environmental pol-icy. ‘No-net-loss’ policies allow for developers who fill in existing wetlands to construct artificial wet-lands as a substitute. However, ecological research suggests that artificial wetlands tend not to have the same species diversity or ecological value as the natural ones that they replace (Kaiser, 2001). Simi-larly, when old-growth forests are clear-cut and re-placed by planting of monoculture forests, the ecological value of the original forest has not been retained, even if the economic value, as measured by board-feet of lumber, is maintained.

In market-failure-based policies, public-value failures are most often a result of the substitution of money for a tangible or natural resource. One espe-cially interesting case implicating science policies and pertaining to medical practice is money-for-body-parts transactions. In the United States, traf-ficking in human organs is illegal — a clear signal that public values should take precedence over mar-ket efficiency. Elsewhere in the world, evidence of a market in human organs to supply rapidly advancing capabilities in medical science continues to crop up. An active market in kidneys and corneas has been documented in India (Kumar, 1994), and an organ market is also thought to exist in the Philippines (Medical Industry Today, 1998).

Such transactions degrade humans, victimize the poor, and invariably occur under some type of economic or political duress; they exemplify public failure. Yet, from a market standpoint, money-for-body-parts transactions may be viewed as efficient, with money being an acceptable substitute for per-sonal health. From a consumer sovereignty stand-point this logic is unimpeachable. As one Harvard economist writes: “If a desperately ill individual who would die without a kidney is able to buy one from a healthy individual, both are made better off. Why … stand in the way of market transactions that will not only make those who engage in them hap-pier but also save lives?” (Shavell, 1999, page 22)

Perhaps whether or not one keeps one’s corneas is really a matter of one’s view about the substitutabil-ity of health for other assets. Nevertheless, in a world where millions live in abject poverty, the

Given the volatility of politics in the Middle East, and the evidence that carbon emissions influence global climate, the declining public investment in research on alternative energy sources and more efficient energy technologies is deeply problematic

Page 12: Public values and public failure in US science policy

Public values and failure in US science policy

130 Science and Public Policy April 2005

notion of consumer sovereignty and of rational choice of one good for another seems less about markets than about massive public-values failure.

Case 6. Public failure and benefit hoarding: terminator technology

In the marketplace, externalities may distort prices and thus skew costs or benefits toward particular consumers. For example, the costs of cleaning up pollution are rarely included in the price of the pol-luting good. Thus, those who produce and consume that good may benefit preferentially. Analogously, if the benefits of a public policy meant to aid a large group are captured preferentially by a much smaller group, public failure may be occurring.

Recent attention to the ‘digital divide’ may illustrate

such a failure. Disparities in health care may be another

example. Development of the internet and many medi-cal technologies was made possible by public support

of the necessary R&D. If only certain segments of the

population are benefiting from this investment, then

benefit hoarding may be taking place. In the early 1980s, following a decade of disap-

pointing economic performance, US policy-makers were anxious to find ways to stimulate economic growth. One area of action focused on creating incentives to transfer the results of Government-funded research to the private sector as a stimulus to technological innovation, and resulted in such laws as the Stephenson-Wydler Act of 1980, the Bayh-Dole Act of 1980, and the Federal Technology Transfer Act of 1986.

The Technology Transfer Act legalized public–private research partnerships, called ‘CRADAs (cooperative research and development agreements), meant to stimulate collaboration between Govern-ment and corporate laboratories. The ‘Findings’ that articulate the rationale for the Act include brief mention of “social well-being,” “increased public services” and “public purposes,” but in fact focus almost entirely on economic arguments, for exam-ple: “Increased industrial and technological innova-tion would reduce trade deficits, stabilize the dollar, increase productivity gains, increase employment, and stabilize prices” (15 USC Sec. 3701). While these are all laudatory goals, they make no mention of possible social impacts that could undermine pub-lic values.

On 3 March 1998, the US Patent Office granted a patent jointly to the US Department of Agriculture’s (USDA) Agricultural Research Service and the Delta and Pine Land Co, a breeder of cotton and soybeans, entitled “Control of Plant Gene Expres-sion.” This patent arose from joint work funded through a CRADA, and embodied the type of tech-nology transfer envisioned by legislators more than a decade earlier.

The patent covered a process, called the Technol-ogy Protection System (TPS), that would allow seeds to be genetically engineered so that they did

not give rise to fertile offspring. The intention was to protect the technological innovation embodied in new varieties of seeds (for example, resistance to drought or herbicides), by ensuring that farmers could not plant second-generation seeds produced by the first generation crop. Rather, they would have to buy new seeds for each planting. In the words of the USDA (ARS, 2001), the new technology “would protect investments made in breeding or genetically engineering these crops. It would do this by reducing potential sales losses from unauthorized reproduc-tion and sale of seed.” This economic argument was causally linked to a social-benefits argument via standard market logic (ARS, 2001):

“The knowledge that the seed companies could potentially recoup their investment through sales will provide a stronger incentive for the companies to develop new, more useful varie-ties that the market demands. Today’s emerg-ing scientific approaches to crop breeding — especially genetic engineering — could be crucial to meeting future world food needs, conserving soil and water, conserving genetic resources, reducing negative environmental ef-fects of farming, and spurring farm and other economic growth. TPS technology will con-tribute to these outcomes by encouraging de-velopment of new crop varieties with increased nutrition to benefit consumers and with stronger resistance to drought, disease and insects to benefit farmers for example.”

TPS technology does appear to hold considerable interest for plant-breeding companies, and TPS patents continue to be granted in the US and abroad (ETC Group, 2002). In essence, TPS makes protec-tion of intellectual property a biological process, rather than a legal one. At present, seed companies must count on the honest farmers to honor intellec-tual property by not ‘brown-bagging’ second-generation seeds, or the companies must resort to policing of farms to enforce their intellectual property. Indeed, in pursuing the latter course, Mon-santo suffered a public relations disaster when they sued a Saskatchewan rape-seed farmer for patent infringement (for instance, Margoshes, 1999).

TPS is a testimony to amazing progress in genetic engineering. The process described in the original patent involves enormously complex, integrated ma-nipulation of transgenic components that are inserted into the DNA of the plant to be protected. A plant gene “normally activated late in seed development” must be fused with a “promoter to the coding se-quence for a protein that will kill an embryo going through the last stages of development” and then cou-pled to a mechanism to repress the promoter until it is treated with a specific chemical (Crouch, 1998).

Less than two years after the TPS patent was granted, M S Swaminathan, one of the founders of the Green Revolution and an advocate of biotechnology

Page 13: Public values and public failure in US science policy

Public values and failure in US science policy

Science and Public Policy April 2005 131

in the service of global agriculture, declared that if TPS was widely adopted, “[s]mall farmers will then experience genetic enslavement since their agricul-tural destiny will be in the hands of a few compa-nies” (Swaminathan, 1999). The Consultative Group on International Agricultural Research (CGIAR, the organization that provided much of the science for the Green Revolution) banned TPS from their re-search agenda (Service, 1998) and Monsanto, which was attempting to acquire Delta and Pine Land Company (co-holder of the original patent), pledged, under pressure from public interest groups and phil-anthropic foundations, “not to commercialize sterile seed technologies” (Shapiro, 1990).

The Rural Advancement Foundation (RAFI, later renamed ETC Group), which mobilized opposition to TPS, coined the phrase “terminator technology” and asserted that the “seed-sterilizing technology threatens to eliminate the age-old right of farmers to save seed from their harvest and it jeopardizes the food security of 1.4 billion people — resource poor farmers in the South — who depend on farm-saved seed” (ETC Group, 1998). RAFI also argued that TPS would further contribute to diminution of global agricultural genetic diversity, especially for plant varieties of importance to developing countries.

The argument against TPS is multifaceted (our summary is drawn from: Visser et al, 2001; Eaton et al, 2002; Service, 1998; ETC Group, 1998; 1999). At the heart of the issue is the practice by many farmers, especially (but not only) in the developing world, to continually seek better plant varieties for local growing conditions, through careful selection of kept seed, as well as purchase of new varieties from both private and public seed distributors.

TPS was alleged to threaten this process in many interconnected ways. First, it would allow commer-cial breeders to capture markets for crops that are not amenable to hybridization, including wheat, rice, and cotton. (Commercial breeders do not focus on such crops precisely because they cannot control farmers’ use of kept seed. Hybrid seed, on the other hand, tends not to express its engineered attributes homogene-ously in the second generation, and thus offers some inherent protection of intellectual property.)

This commercialization of seed varieties in turn would inevitably reduce the available commercial sources of such seed because of advantages con-ferred to larger breeders and seed purchasers by economies of scale. Local plant breeders’ access to new genetic materials would thus become increas-ingly restricted, and their ability to select for improved seed varieties would be impaired.

Secondly, because commercial plant breeders would be aiming their products at the most profit-able markets — those of the rich countries — they would be unlikely to engineer plant varieties to meet the needs of poorer farmers, as has been the case generally with hybrid products. At the same time, publicly funded plant breeding organizations, such as CGIAR, might be blocked from using engineered

traits developed by private breeders unless they also accepted TPS. Such trends would exacerbate agricultural technology gaps between rich and poor.

Thirdly, because poor farmers would find it in-creasingly difficult to acquire seed without termina-tor technology, their exposure to year-to-year food-supply disruption as a result of economic, political, climatic, or other factors would increase. Finally, genetic diversity of agricultural varieties would de-cline, because the largest source of such diversity is the seed-production activities of farmers themselves. Large breeding companies tend to reduce, not increase, genetic diversity.

In defense of TPS, USDA focuses on market arg-uments (ARS, 2001):

“[L]oss of cost savings from brown-bagging [kept seed] also must be weighed against the productivity gains to the farmer from having superior new varieties that could increase crop values such as yield and quality, input cost re-ductions such as for fertilizers and pesticides, and reduced losses such as those due to pests or adverse soils and weather.”

Such arguments assume a level playing field, where the attributes of new, engineered seed varieties will be those needed by small farmers and poor farmers, where such farmers will be able to afford the new varieties, and where they will, therefore, no longer be dependent on their own seed selection skills to optimize crops for particular local growing condi-tions. Even should such an optimistic scenario tran-spire, it ignores the effects of reduced genetic diversity on the resilience of agricultural systems worldwide.

Terminator technologies thus create a possibility for corporations to gain control of a process — seed selection — and a product — plant varieties — that have been in the hands of farmers for millennia. The effect is a private hoarding of what had been a public good — the plant-genetic commons. This effect is less troubling in the context of affluent nations, where agriculture has become increasingly industrialized, than for poor ones, where small farmers continue

Terminator technologies create a possibility for corporations to gain control of a process — seed selection — and a product — plant varieties — that have been in the hands of farmers for millennia: the effect is a private hoarding of what had been a public good

Page 14: Public values and public failure in US science policy

Public values and failure in US science policy

132 Science and Public Policy April 2005

to depend upon kept seed and selective breeding for crop improvement and adaptation, and for food se-curity from year-to-year.

A particularly conspicuous element of this story is that the original research was partly funded by pub-lic money and conducted at a public research labora-tory. As such, it is an exemplar of the way that market values displace public values in justifying public funding of science and technology.

Science policy and the ‘public value grid’

Among other implications of these examples and the criteria they illustrate is that public failure and mar-ket failure are not single poles on a dimension or even two orthogonal dimensions. Instead, it is best to view the two as axes of a grid, as in Figure 1, which provides a depiction of the ways in which science policies can have very different economic and public values outcomes.

The notion of setting market values against other values not easily encompassed by market framework is not new. For example, Page (1977) suggested con-trasting dimensions of market efficiency and conser-vation of resources. Norton and Toman (1997) speak of “two tiers,” one an efficiency criterion, the other a conservation criterion. Figure 1 provides a highly simplified depiction of public failure and market failure, illustrating the possibility of a virtually infi-nite range of outcomes among the extremes of com-plete public failure, complete public success, complete market failure, complete market success. One obvious point is that market failure and public failure need not be correlated at all.

Figure 1 is broadly illustrative. We might easily quibble with the exact location of the policies de-picted. Another obvious limitation is that such broad categories as “academic research policy” are little more than stand-in symbols for many diverse science

(and other) policies. However, the lack of precision should not undercut the chief point.

Take the case of Internet technology develop-ment. We might argue, for example, that the history of Internet technology development provides a happy example of a public values success and, at the same time, a market success. While the Internet, much as any ubiquitous social phenomenon, has not been a success in every possible respect (witness controversies about privacy, pornography, intellec-tual property, and ‘spam’), most would, on balance, assess the commercialization of the Internet as both a market and public values success.

Similarly, R&D policy for pharmaceuticals could be viewed as a considerable market success but in many respects a public failure. In neither instance is it a ‘pure’ success or failure and, thus, not at the ex-treme end of either pole.

One point worth noting is that over time policies move in ‘normative space,’ repositioning themselves in Figure 1. Thus, at the beginning of Internet tech-nology development, the best description would be ‘modest public values success combined with mar-ket failure.’ That is, in 1992, the Internet (not yet called that) served some significant if limited public values in linking scientists and a few other users. However, at that time, the Internet was a nearly per-fect market failure in the sense that no commercial value had yet been harvested, nobody was in a posi-tion exclusively to appropriate its benefits, and no price mechanisms were in place.

The public-failure grid provides a simple analyti-cal tool for thinking about the public values and economic values served by science policies. We need not be ‘right’ in the positioning of policies to find such deliberation useful. Likewise, we need not entertain each public failure (and each market fail-ure) diagnostic criterion to employ the grid and to obtain its value as a common sense check on delib-erations about the values served by science policies.

PF

PS

MF MS

Academic research policy

Internet technology development

Science and math education policy R&D policy for

pharmaceuticals

Figure 1. Public failure grid and science policy examples

Page 15: Public values and public failure in US science policy

Public values and failure in US science policy

Science and Public Policy April 2005 133

Public failure theory

For the past 50 years, questions of choice have been central to science policy discourse. In a world of finite resources, how shall we choose among various portfolios of scientific investments and research strategies, and who shall do the choosing? The famed Bush–Kilgore debates of the late 1940s, which sought in part to establish the degree to which publicly funded science should be linked to specific social goals (Kevles, 1987), set the stage for an ini-tial phase of intellectual debate that took as its start-ing point the assumption that science needed to be insulated from the vulgarities of democracy, and proceeded to ask how, then, choices might be made to serve the best interests of society (for instance, Polanyi, 1962; Weinberg, 1963; Brooks, 1968, chap-ters 1 and 2).

Yet the texture and depth of thinking about choice in science policy has had remarkably little impact on the actual science policy process, which remains, above all, a competition for incremental budget in-creases. Each year, science policy, as so much of the rest of federal policy, becomes riveted on the federal budget and, specifically, the amount of money avail-able for science. There is great deliberation about the actual amount, the rate of growth, the percentage of the budget, the amount for science as opposed to other outlay categories, and the amount (and per-centage and rate of growth) for particular agencies and particular programs within agencies, as com-piled annually by the American Association for the Advancement of Science (for instance, Intersociety Working Group, 2003), and biennially by the National Science Board (for instance, NSB, 2002).

Every year, even during periods of considerable growth in spending, there are at least a few ‘funding crises’ identified and these become grist for op eds (a page of special features usually opposite the edi-torial page of a newspaper), sound bites, lobbying, and anguish. (The latest example is the concerns voiced on behalf of the National Institutes of Health (NIH), whose budget doubled between 1998 and 2002, and must now suffer a reversion to rates of increase typical of its prior history of robust growth (Weiss, 2003).)

The fact is, few people (and more likely no peo-ple) have the breadth of understanding to even begin to provide a valid account of what will happen as a result of a 5% decrease or increase in, say, biochem-istry funding. Yet, in the world where science policy decisions are actually made, science funding acts as a proxy for public value; more of the former is as-sumed to yield more of the latter. Policy documents (for instance, Carnegie Commission on Science, Technology, and Government, 1992; Institute of Medicine, 1998) and scholarship (for instance, Pe-tersen, 1984; Kitcher, 2001) have made the case for greater diversity of voices in science policy priority-setting, a recommendation that we would support, yet one that does little to address the more complex

question of how successfully a given line of research may actually connect to advancing public values. It is telling that successful efforts to influence the in-ternal conduct of science to achieve particular socie-tal outcomes are relatively rare, and have mostly been driven by highly motivated and politically em-powered groups focused on changing the norms of clinical health research (for instance, Lerner, 2001; Epstein, 1996).

We suggest that, while democratizing choice is important, it is insufficient for enhancing the public value of science. Indeed, in one very real sense, choice in science policy has been highly democra-tized. In the words of the late Congressman George E Brown, Jr (1999), perhaps the most thoughtful of all practitioners of the politics of science: “Congress does have a rational priority-setting system. Unfor-tunately it is largely zip-code based: Anything close to my district or state is better than something farther away. Individual colleges and universities and many federal labs know this system well and have used it to their advantage for decades.”

The problem is that what really counts when it comes to the public values flowing from science pol-icy is not so much budgetary level of effort as the institutional, cultural, political, and economic con-texts in which science is produced and applied. So why, in the face of decades of critical STS scholar-ship about these complex contextual relations, does the formula that equates more-money-for-science with more-public-value assert itself with undimin-ished vigor? We suggest this is a consequence of two factors: first, the bipartisan power of economic thinking, bolstered by evidence of the key role for science in economic growth, and rationalized by the market-failure model for government intervention; and, second, the absence of analytical and rhetorical tools that can match the simplicity and potency of economic argumentation.

Here we have presented simple criteria that can be used to test claims that more science equals more public value. If values are not well articulated or aggregated; if public monopolies are imperfect; if providers are scarce; if time horizons are short; if resources are viewed as substitutable; if benefits can be captured by small groups; then the translation of science investments into public values may well be compromised, regardless of how well the market is operating, or how much money the science receives.

Note

1. In the US, public schools are free, tax-supported schools controlled by a local government authority.

References

Angell, M (2000), “Is academic medicine for sale?”, New England Journal of Medicine, 342(20), pages 1516–1518.

Page 16: Public values and public failure in US science policy

Public values and failure in US science policy

134 Science and Public Policy April 2005

Arocena, R, and P Senker (2003), “Technology, inequality and underdevelopment: the case of Latin America”, Science, Technology and Human Values, 28(1), pages 15–33.

ARS (USDA Agricultural Research Service) (2001),”Why USDA’s Technology Protection System (a.k.a. ‘Terminator’) benefits agriculture”, ARS News and Information website: <http://www. ars.usda.gov/is/br/tps/>, last accessed 22 September 2002.

Bannock, G, R Baxter and E Davis (1998), The Penguin Diction-ary of Economics (Penguin Books, New York).

Barke, R, and H Jenkins-Smith (1993), “Politics and scientific expertise: scientists, risk perception, and nuclear waste pol-icy”, Risk Analysis, 13(4), pages 425–439.

Bauer, M, K Petkova and P Boyadjieva (2000), “Public knowledge of and attitudes to science: alternative measures that may end the ‘science war’”, Science, Technology and Human Values, 25(1), pages 30–51.

Bimber, Bruce, and David Guston (1995), “Politics by the same means: government and science in the United States”, in Jasanoff et al (1995), pages 554–571.

Bodenheimer, T (2000), “Uneasy alliance — clinical investigators and the pharmaceutical industry”, New England Journal of Medicine, 342(20), pages 1539–1544.

Bozeman, B (2002), “Public value failure: when efficient markets may not do”, Public Administration Review, 62(2), pages 134–151.

Brooks, Harvey (1968), The Government of Science (MIT Press, Cambridge MA) pages 1–53.

Brown, George E, Jr (1999), “Past and prologue: why I am optimistic about the future”, in 1999 AAAS Science and Tech-nology Policy Yearbook (American Association for the Advancement of Science, Washington DC) chapter 3, avail-able at: <http://www.aaas.org/spp/yearbook/chap3.htm>, last accessed 17 October 2002.

Bush, Vannevar (1945), Science: The Endless Frontier (National Science Foundation, Washington DC).

Callahan, D (1998), False Hopes: Why America’s Quest for Per-fect Health is a Recipe for Failure (Simon and Schuster, New York).

Carnegie Commission on Science, Technology, and Government (1992), Enabling the Future: Linking Science and Technology to Societal Goals (Carnegie Corporation, New York).

Castells, M (1998), End of Millennium (Blackwell, Malden MA). Cho, M, and L Bero (1996), “The quality of drug studies published

in symposium proceedings”, Annals of Internal Medicine, 124(5), pages 485–489.

Clinton, William J, and Albert Gore Jr (1994), Science in the National Interest (Office of Science and Technology Policy, Washington, DC) page 1.

Cochran, C E (1974), “Political science and ‘the public interest’”, The Journal of Politics, 36, pages 327–355.

Code of Federal Regulations (1993), “Title 45, Public Welfare, Department of Health and Human Services, National Institutes of Health, Office for Protection from Research Risks, part 46, Protection of Human Subjects” (US Government Printing Of-fice, Washington DC) pages 112–113.

Cozzens, S, and E Woodhouse (1995), “Science, government, and the politics of knowledge”, in Jasanoff et al (1995), pages 533–553.

Crouch, M (1998), “How the terminator terminates: an explanation for the non-scientist of a remarkable patent for killing second generation seeds of crop plants” , Occasional Paper Edmonds, Edmonds Institute WA, revised edition.

Denison, Edward (1962), The Sources of Economic Growth in the United States and the Alternatives before Us (Committee for Economic development, New York).

Dickson, David (1984), The New Politics of Science (Pantheon, New York).

Donahue, J (1991), The Privatization Decision (Basic Books, New York).

Dooley, J J (1999), Energy Research and Development in the United States (Pacific Northwest National Lab, Washington DC) PNNL-12188, available at <http://energytrends.pnl.gov/ usa/documents/us005lpdf>, last accessed 17 October 2002.

Dupree, A Hunter (1986), Science in the Federal Government (The Johns Hopkins University Press, Baltimore MD).

Eaton, D, F van Tongeren, N Louwaars, B Visser and I Van der Meer (2002), “Economic and policy aspects of ‘terminator’ technology”, Biotechnology and Development Monitor, 49, page 19–22.

Energy Information Administration (2000), Annual Energy Review, Table 1.2, available at <http://www.eia.doe.gov/emeu/aer/txt/

tab0102.htm>, last accessed 21 October 2003. Energy Information Administration (2001), Annual Energy Out-

look, Table 7, available at <http://www.eia.doe.gov/oiaf/aeo/ tbl7.html>, last accessed 21 October 2003.

Epstein, Steven (1996), Impure Science: AIDS, Activism, and the Politics of Knowledge (University of California Press, Berkeley CA).

ETC Group (1998), “Terminator technology targets farmers”, Communique, 30 March, available at http:// www.etcgroup.org/ article.asp?newsid=188>, last accessed 27 April 2003.

ETC Group (1999), “The terminator’s wider implications”, Com-munique, 30 January, available at <http:// www.etcgroup.org/ article.asp?newsid+184>, last accessed 27 April 2003.

ETC Group (2002), “Sterile harvest: new crop of terminator pat-ents threatens food sovereignty”, News Item, 31 January, available at <http:// www.etcgroup.org/article. asp?newsid=290>, last accessed 27 April 2003.

Evans, R G, M L Barer and T R Marmor (editors) (1994), Why Are Some People Healthy and Others Not? The Determinants of Health of Populations (Aldine de Gruyter, New York).

Ezrahi, Yaron (1990), The Descent of Icarus: Science and the Transformation of Contemporary Culture (Harvard University Press, Cambridge MA).

FASEB (Federation of American Societies for Experimental Biol-ogy) (2001), “FASEB President says Bush’s budget for the NIH is a ‘Great Start,’ but vows to push for more on Capitol Hill”, FASEBnews (press release), 21 February, available at <http://www.faseb.org/opar/news/docs/nr2x23x1.html>, last accessed 22 July 2003.

Federal Geographic Data Committee (2001), “Overview: what the framework approach involves”, available at <http://www.fgdc. gov/framework/frameworkintroguide/overview.html>, last ac-cessed November 2001.

Flathman, R (1966), The Public Interest (Wiley, New York). Friedberg, M, B Saffran, T Stinson, W Nelson and C L Bennett

(1999), “Evaluation of conflict of interest in economic analyses of new drugs used in oncology”, Journal of the American Medical Association, 282(15), pages 1453–1457.

Fukuyama, F (2002), Our Post-human Future: Consequences of the Biotechnology Revolution (Allen and Unwin, New York).

Fuller, Lon (1964), The Morality of Law (Yale University Press, New Haven CT).

Fuller, S (2000), The Governance of Science (Open University Press, London).

Galbraith, J K, and M Berner (editors) (2001), Inequality and In-dustrial Change: a Global View (Cambridge University Press, Cambridge UK).

Goldner, J A (2000), “Dealing with conflicts of interest in biomedi-cal research: IRB oversight as the next best solution to the abolitionist approach”, The Journal of Law, Medicine, & Ethics, 28(4), pages 379–404.

Greenberg, Daniel (2001), Science, Money, and Politics (Univer-sity of Chicago Press, Chicago IL).

Griliches, Z (1995), “R&D productivity: econometric results and measurement issues”, in S Paul (editor), The Handbook of the Economics of Innovation and Technological Change (Black-well, Oxford).

Guston, D (2000), Between Politics and Science: Assuring the Integrity and Productivity of Research (Oxford University Press, Oxford).

Held, V (1972), The Public Interest and Individual Interests (Basic Books, New York).

Hull, David L (1988), Science as a Process: an Evolutionary Ac-count of the Social and Conceptual Development of Science (University of Chicago Press, Chicago IL).

Institute of Medicine (1998), Scientific Opportunities and Public Needs: Improving Priority Setting and Public Input at the National Institutes of Health (National Academies Press, Washington DC).

Intersociety Working Group (2003), AAAS Report XXVII: Re-search and Development FY 2004 (AAAS, Washington DC).

Jasanoff, S, G Markle, J Petersen and T Pinch (editors) (1995), Handbook of Science and Technology Studies (Sage Publica-tions, Thousand Oaks CA).

Kaiser, J (2001), “Wetlands restoration: recreated wetlands no match for original”, Science, 293(5527), page 25.

Kevles, Daniel J (1987), The Physicists: The History of a Scientific Community in Modern America (Harvard University Press, Cambridge MA).

Kitcher, Philip (2001), Science, Truth, and Democracy (Oxford University Press, New York NY).

Page 17: Public values and public failure in US science policy

Public values and failure in US science policy

Science and Public Policy April 2005 135

Kleinman, Daniel L (1993), Politics on the Endless Frontier: Post-ward Research Policy in the United States (Duke University Press, Durham NC).

Kumar, S (1994), “Curbing the trade in human organs in India”, Lancet, 344, page 750.

Lerner, Barron H (2001), The Breast Cancer Wars (Oxford Uni-versity Press, New York).

Leslie, S W (1993), The Cold War and American Science (Co-lumbia University Press, New York).

Lindblom, C E (2001), The Market System (Yale University Press, New Haven).

Lindsay, L (2001), Speech (untitled) to the American Association for the Advancement of Science, Science and Technology Pol-icy Colloquium, 3 May, available at <http://www.ostp.gov/html/ 010515.html>, last accessed 21 February 2001.

Mansfield, E (1968), The Economics of Technical Change (W W Norton, New York).

Mansfield, E (1980), “Basic research and productivity increases in manufacturing”, American Economic Review, 70, pages 863–873.

Margolis, R M, and D M Kammen (1999), “Underinvestment: the energy technology and R&D challenge”, Science, 285, 30 July, pages 690–692.

Margoshes, Dave (1999), “Saskatchewan farmer battles agro-chemical giant”, The Vancouver Sun, 14 August, page B1.

Medical Industry Today (1998), “Suspected murder for organ trade probed in Philippines”, 9 September, pages 127–128.

Mowery, D C, and N Rosenberg (1989), Technology and the Pur-suit of Economic Growth (Cambridge University Press, New York).

NAPA, National Academy of Public Administration (1998), Geo-graphic Information for the 21st Century — Building a Strategy for the Nation (NAPA, Washington DC), available at <http://www.gsdi.org/docs/canberra/naparepo.html>, last accessed 12 April 2005.

NAPA, National Academy of Public Administration (2001), A Study of the National Science Foundation’s Criteria for Project Selection (NAPA, Washington DC).

Narin, Francis, and Dominic Olivastro (1992), “Status report: link-age between technology and science”, Research Policy, 21, pages 237–249.

NASA, National Aeronautics and Space Administration (1998), “Landsat program chronology”, available at <http://geo.arc. nasa.gov/sge/landsat/lpchron.html>, last accessed 4 Novem-ber 1998.

National Bioethics Advisory Commission (2001), Ethical and Pol-icy Issues in Research Involving Human Participants (NBAC, Bethesda MD).

NSB, National Science Board (2002), Science and Engineering Indicators 2002 (National Science Foundation, Washington DC).

Nelson, Richard R (1959), “The simple economics of basic scien-tific research”, Journal of Political Economy, 67, pages 297–306.

Nelson, Richard R (in press), “The market economy and the sci-entific commons”, Research Policy.

Nelson, R, and R Weiss (1999), “Hasty decisions in the race to a cure?: gene therapy study proceeded despite safety, ethics concerns”, Washington Post, 21 November, page A1.

Nelson, R R, M J Peck and E D Kalachek (1967), Technology, Economic Growth, and Public Policy; a Rand Corporation and Brookings Institution Study (Brookings Institution, Washington DC).

Noble, David F (1984), Forces of Production (Alfred A Knopf, New York).

Noble, David F (1995), Progress without People. New Technol-ogy, Unemployment and the Message of Resistance (Between The Lines, Toronto).

Norburg, A L, and J E O’Neil (1996), Transforming Computer Technology: Information Processing for the Pentagon, 1962–1986 (Johns Hopkins University Press, Baltimore MD).

Norton, B and M Toman (1997), “Sustainability: ecological and eco-nomic perspectives”, Land Economics, 73(4), pages 553–568.

NRC (1991), Assessment of Satellite Earth Observation Programs 1991 (National Academy Press, Washington DC).

NRC (National Research Council) (1993), Toward a Coordinated Spatial Data Infrastructure for the Nation (National Academy Press, Washington DC).

NRC (National Research Council) (1995), A Data Foundation for the National Spatial Data Infrastructure (National Academy Press, Washington DC).

NRC (National Research Council) (2001), Future Roles and Op-portunities for the U.S. Geological Survey (National Academy Press, Washington DC).

NSB, National Science Board (2002) , Science and Engineering Indicators 2002 (National Science Foundation, Washington DC).

Open GIS Consortium (1999), Spatial Connectivity for a Changing World (promotional brochure) (Open GIS Consortium, Way-land MA).

Page, R Talbot (1977), Conservation and Economic Efficiency (Johns Hopkins University Press for Resources for the Future, Baltimore MD).

Pavitt, Keith (1998), “The social shaping of the National Science Base”, Research Policy, 27(12), pages 793–805.

Peterson, James C (editor) (1984), Citizen Participation in Sci-ence Policy (University of Massachusetts Press, Amherst MA).

Plutzer, Eric, A Maney and R O’Connor (1998), “Ideology and elites’ perceptions of the safety of new technologies”, Ameri-can Journal of Political Science, 42(1), pages 190–209.

Polanyi, Michael (1962), “The republic of science: its political and economic theory”, Minerva, 1, pages 54–73.

Press, Eyal, and Jennifer Washburn (2000), “The kept university”, The Atlantic Monthly, 285, March, pages 39–54.

Rosenberg, Nathan (1982), Inside the Black Box. (Cambridge University Press, New York) pages 141–159.

Rosenberg, Nathan, and Richard R Nelson (1994), “American universities and technical advance in industry”, Research Pol-icy, 23(8), pages 323–348.

Rosenbloom, R, and W Spencer (1996), Engines of Innovation: U.S. Industrial Research at the End of an Era (Harvard Busi-ness School Press, Boston MA).

Sarewitz, Daniel (1996), Frontiers of Illusion: Science, Technol-ogy, and the Politics of Progress (Temple University Press, Philadelphia PA).

Schubert, G (1960), The Public Interest (The Free Press, Glencoe IL).

Sen, Amartya (1987), The Standard of Living (Cambridge Univer-sity Press, Cambridge UK).

Sen, Amartya (1997), “Inequality, unemployment and contempo-rary Europe”, International Labour Review, 136(2), pages 155–171.

Service, Robert F (1998), “Seed-sterilizing ‘terminator technology’ sows discord”, Science, 282, 30 October, pages 850–851),

Shapiro, R B (1990), “Open letter from Monsanto CEO Robert B. Shapiro to Rockefeller Foundation President Gordon Conway”, 4 October, available at <www.biotech-info.net/monsanto_ letter.pdf>, last accessed 12 April 2005.

Shavell, S (1999), “Why not sell organs?, New York Times, 11 May, page 22.

Smedley, B D , A Y Stith and A R Nelson (editors) (2002), Un-equal Treatment: Confronting Racial and Ethnic Disparities in Health Care (National Academies Press, Washington DC).

Sorauf, F (1957), “The public interest reconsidered”, Journal of Politics, 19, pages 616–639.

Stelfox, H T, G Chua, K O’Rourke and A S Detsky (1998), “Con-flict of interest in the debate over calcium-channel antago-nists”, New England Journal of Medicine, 338(2), pages 101–106.

Swaminathan, M S (1999), “Genetic engineering and food, eco-logical livelihood security in predominantly agricultural developing countries”, speech to CGIAR/NAS Biotechnology Conference, 21 October, available at <http://www.cgiar.org/ biotechc/swami.htm>, last accessed 12 December 2002

Tietz, G F (1993), “Strict products liability, design defects and corporate decision-making: greater deterrence through stricter process”, Villanova Law Review, 38, pages 1361–1459.

UNDP, United Nations Development Program (1999), Human Development Report 1994 (Oxford University Press, New York).

Visser, B; I van der Meer, N Louwaars, J Beekwilder and D Eato (2001), “The impact of ‘terminator’ technology”, Biotechnology and Development Monitor, 48, pages 9–12.

Wade, N (1999), “Patient dies during a trial of therapy using genes”, New York Times, 29 September, page A24.

Weinberg, A M (1963), “Criteria for scientific choice” Minerva, 1(2), pages 159–171.

Weiss, Rick (2003), “NIH braces for slower funding growth: Presi-dent may apply brakes on 14% to 15% annual increase, and trim it to 2%”, Washington Post, 2 February, page A14.

Wilkinson, R G (1996), Unhealthy Societies: The Afflictions of Inequality (Routledge, New York).

Page 18: Public values and public failure in US science policy

Public values and failure in US science policy

136 Science and Public Policy April 2005

Winner, Langdon (1986), The Whale and the Reactor. A Search for Limits in an Age of High Technology (University of Chicago Press, Chicago IL).

Woodward, B (1999), “Challenges to human subject protection in US medical research”, Journal of the American Medical Asso-ciation, 282(20), pages 1947–1952.