52

American Public Opinion: Its Origins, Content and Impact

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: American Public Opinion: Its Origins, Content and Impact
Page 2: American Public Opinion: Its Origins, Content and Impact

CHAPTER 2Polling: The Scientific Assessment

of Public Opinion

Page 3: American Public Opinion: Its Origins, Content and Impact

• •

LEARNING OBJECTIVES

Identify ways that polls affect politiciansExplain probability sampling, sampling error, andconfidence levelIdentify potential biases in the wording of public opinionquestionsExplain how preelection polls are conducted and describethe accuracy of those polls

The use of scientific polling to gauge and analyze public opinion andelections has now been part of the political landscape for over seventyyears. And it continues to grow. This growth is evident in the media’sobsession with campaigns as horse races, where journalists’ interest liesmore in forecasting the outcome than the substance of the campaign.During the presidential election season, the news media is saturated withpolling reports. In recent presidential campaigns, one finds national pollsbeing reported at a rate of greater than one per day (Moore 2013), with evenmore state polls reported on a daily basis (Wlezien and Erikson 2002;Cohen 2008).1 Perhaps as many as two-thirds of the election stories that arereported on the network news in recent years are supported by polling data(Hess 2000). The news media’s preoccupation with the polls appears toagree with public demand. In 2012, 20 percent of all the traffic to the NewYork Times Internet site went to Nate Silver’s blog on election polling(Owen 2013).

Page 4: American Public Opinion: Its Origins, Content and Impact

2-1 The Value of PollsOf course, polling is not confined to election polls and the prediction of thenext election outcome. As discussed in Chapter 1, academic polls advanceour knowledge of public opinion. Commercial pollsters (e.g., Gallup)satisfy the public’s (and private clients’) curiosity regarding trends in publicopinion. Although not as highly publicized as the pollsters’ electionforecasts, polls routinely ascertain public preferences on a variety of policyissues and monitor the public pulse regarding such indicators as partyidentification, ideological identification, and the approval rating of thecurrent president. This is the stuff of public opinion analysis and the subjectof subsequent chapters.

Politicians and the PollsPoliticians have a particularly strong interest in the polls, as theirprofessional careers may depend on reading public opinion accurately. Likethe rest of us, politicians have access to commercial opinion polls such asthe Gallup Poll. But the politicians often seek further intelligence regardingpublic opinion by commissioning their own polls. The first president to pollthe public was Franklin Roosevelt, who did so in the late 1930s. While hisnext two successors, Truman and Eisenhower, had little use for privatepolling, presidential polling picked up in earnest with John Kennedy. Sincethen, virtually every major presidential decision has been made with thebenefit of private polls. Presidents seek updates of public attitudes towardcurrent policies, public reactions to future policy options, and especiallyinformation on their own standing with the voters (Jacobs and Shapiro2000; Fried 2012).

Politicians below the presidential level also do their own polling, as dothe major political parties. Senate and House members spend about 3percent of their sizable campaign war chests on polling their states ordistricts (Fritz and Morris 1992). Although the general purpose of thispolling is to monitor their electoral standing against current or futureopponents, private constituency polls can also register signals regardingshifts in constituency concerns.

Page 5: American Public Opinion: Its Origins, Content and Impact

A candidate’s poor showing in preelection polls makes fund-raisingdifficult and dampens volunteer enthusiasm. As observed by Mike Murphyand Mark Mellman (2007), two leading political consultants (one aRepublican and the other a Democrat), “Early national polling is used todeclare winners and losers. Those declarations affect the flow of money and[media] coverage.” Good poll numbers result in more media attention(Traugott and Lavrakas 2008). A strong showing in the polls can legitimizea candidate, ensuring that he or she is taken seriously, and being takenseriously is necessary to raise money.

Polls create expectations about who is the likely election winner. Inpresidential primaries with a number of more or less unknown candidates,expectations about who will win in the preelection polls help to frame thepostelection spin. Sometimes losers manufacture “wins” by beatingexpectations even when they fail to win the most votes. A classic exampleoccurred in the New Hampshire Democratic primary of 1992, when BillClinton came in second (to Paul Tsongas) and hailed himself as the“comeback kid.” In the 2004 election, this pattern was repeated when JohnEdwards received a bonanza of favorable press coverage when he beatexpectations in the Wisconsin primary, even though he lost to John Kerryby 6 percent of the vote (Ruttenberg 2005). Thomas Patterson (2002, 2005)makes a compelling argument that the standing in the polls of a presidentialcandidate determines the way the candidate’s personal qualities aredescribed by the media. When poll numbers are good, news stories arepositive; when poll numbers are bad, stories turn negative (see alsoFarnsworth and Lichter 2011). For example, in an analysis of Newsweek’sreporting of the 1988 presidential campaign, Patterson demonstrates thatwhen polls showed Michael Dukakis doing well in the 1988 presidentialrace, he was described as “relentless in his attack” and “a crediblecandidate.” As he later dropped in the polls, Newsweek described him as“reluctant to attack” and “trying to present himself as a credible candidate.”Interestingly, the change in descriptors coincided quite closely with thepoint he fell behind George H.W. Bush in the polls (Patterson 1989, 104–6).Finally, standing in the polls can determine whether or not a candidate isinvited to participate in the presidential debates. For example, while RossPerot’s poll numbers justified his appearance in the 1992 debates, his pollsupport was deemed insufficient for his inclusion in the 1996 presidentialdebates.

Page 6: American Public Opinion: Its Origins, Content and Impact

Presidential popularity, now gauged virtually daily, is the publicopinion indicator watched with perhaps the most interest by members ofCongress, others in the political community, and the media. Arguably, oneimportant aspect of presidential power, including influence over Congress,is the chief executive’s personal standing with the public (Canes-Wrone andde Marchi 2002; Bond, Fleisher, and Wood 2003; Canes-Wrone and Shotts2004; Lebo and O’Green 2011). While it is debatable how much a presidentcan leverage popular approval into public support for his favored projects,there is little doubt that members of Congress and Washington insidersbelieve that a popular president holds the power to persuade public opinion.It was the high job approval enjoyed by President Clinton (about 70percent) that in the opinion of many (including Clinton) saved hispresidency after he was impeached by the House of Representatives.2

President George W. Bush’s surge in popularity following 9/11 smoothedthe passage of the Patriot Act and other items from his conservative agendathrough the Congress.

The Public Looks at the PollsMore and more people say they are paying attention to the polls. In 1944,only 19 percent of the public said they regularly or occasionally followedpoll results; by 1985, it was 41 percent, and in 2000, the figure had risen to65 percent. In 2008, 89 percent said they had read or heard about the latestpolls in the presidential election contest.3 Yet few citizens have even arudimentary understanding of how a public opinion poll works. InSeptember 2005, the Gallup organization asked a national sample whetherinterviews with 1,500 or 2,000 people (typical sample sizes for nationalpolls) “can accurately reflect the views of the nation’s population” orwhether “it’s not possible with so few people.” Only 30 percent said asample of that size could yield accurate results, 68 percent said it was notpossible, and the remaining 2 percent had no opinion. While the public maynot understand the technicalities of polling, a sizable majority neverthelessbelieves polls are important guides for public officials (Traugott and Kang2000). In 2011, 68 percent said the nation would be better off if politicalleaders followed the views expressed in public opinion polls more closely.4

However, in recent years polling methodology has drawn far more thanthe usual public attention, thanks to the Internet. Bloggers from both the

Page 7: American Public Opinion: Its Origins, Content and Impact

political left and right complain about alleged biases in favor of onecandidate or another. Perhaps that is the reason why confidence in thepolling community has shown a recent decline. In 1998, 55 percent of thepublic said they “generally trust [pollsters] to tell them the truth.” By 2006,trust in pollsters had dropped to 34 percent, about the same level (at thetime) as trust in Congress (Gershenson, Glaser, and Smith 2011). On thepositive side, access to the Internet has allowed interested members of thepublic to learn more about the ins and outs of polling methodology thanever before.5

Page 8: American Public Opinion: Its Origins, Content and Impact

2-2 SamplingPublic opinion polls are based on samples. When the Gallup Poll reportsthat 50 percent of adult Americans approve of the way the president ishandling his job, it is obvious that the Gallup organization has not gone outand interviewed 225 million American adults. Instead, it has taken asample. The reasons for sampling are fairly straightforward. First, tointerview everyone would be prohibitively expensive. The Census for 2010is estimated to cost $11.8 billion. Second, to interview the entire populationwould take a very long time. Months might pass between the first and lastinterviews. Public opinion might, in that period, undergo real change.

Sampling provides a practical alternative to interviewing the wholepopulation—be it national, state, or local. Furthermore, when donecorrectly, sampling can provide accurate estimates of the political opinionsof a larger population. The theory of sampling is a branch of themathematics of probability, and the error involved in going from the sampleto the population can be known with precision. However, as the pollingenvironment changes, many surveys of public opinion are challenged by thedemanding requirements of sampling theory.

The history of scientific sampling can be divided into three eras(Goidel 2011). In the first era, between 1935 and 1974, almost all surveyswere done in person, usually at the home of the respondent. Telephonepenetration was not sufficiently widespread to allow for the selection ofrepresentative telephone samples. In 1936, 33 percent of the population hadlandline telephones, by 1946 it raised to 50 percent, and by 1974 it was 95percent.6

In the second era, beginning around 1974, the near universalpenetration of landline telephones plus the development of statisticalmethods to select a representative sample of telephone numbers, along withmuch lower cost, led to a dramatic increase in telephone surveys. For themedia, this method mostly supplanted in-person surveys, althoughgovernment agencies and major academic survey units continued to usethem.

The third era began around the onset of the twenty-first century withthe rise of Internet surveys. One benefit of Internet surveys is the increasein penetration. According to the U.S. Census, in 2012, 80 percent of

Page 9: American Public Opinion: Its Origins, Content and Impact

Americans had Internet access.7 With the parallel rise in cell phone-onlyhouseholds, telephone surveys became more expensive and response ratesplummeted. Most reputable survey firms now use a mix of landline and cellphone numbers. But with the development of sophisticated statisticaladjustment routines, proponents of Internet surveys began to proclaim themas differing little in accuracy from other survey modes—and again oftenmuch less expensive.

Sampling TheoryThe population is that unit about which we want information. In mostpolitical surveys, the population is one of the following: (1) those agedeighteen and older, (2) registered voters, or (3) those who will (or do) votein the next election. These are three quite different groups. It is importantthat the population be clearly specified. When one sees a poll addressingabortion, presidential popularity, or voter intent in an upcoming election,those reporting the poll should provide a clear description of the populationabout which they speak. The sample is that part of the population selectedfor analysis. Usually, the sample is considerably smaller than thepopulation. National telephone surveys conducted by reputable firmsemploy samples of 1,000 to 1,500 respondents, although smaller samplesare sometimes used in state and local contests. But as samples get smaller,the probability of error increases. Sample size should always be reportedalong with the results of a survey. If that information is missing, the allegedfindings should be treated with a great deal of caution.

When samples accurately mirror the population, they are said to berepresentative. The term randomness refers to the only method by which arepresentative sample can be scientifically drawn. In a simple randomsample, each unit of the population has exactly the same chance of beingdrawn as any other unit. If the population is American attorneys, eachattorney is required to have exactly the same probability of being selectedin order for the sample to be random. Attorneys in big cities could not havea greater likelihood of getting into the sample than those from rural areas.This situation obviously requires a detailed knowledge of the population. Inthe case of attorneys, one could get a list from the American BarAssociation and then sample from that list. But suppose the population wasunemployed adults. To specify the population in a fashion to be able to

Page 10: American Public Opinion: Its Origins, Content and Impact

draw a random sample would be difficult. As a consequence, obtaining arepresentative sample of the unemployed is not easy.

A probability sample is a variant of the principle of random sampling.Instead of each unit having exactly the same probability of being drawn,some units are more likely to be drawn than would others—but this is aknown probability. For example, if one were sampling voter precincts in astate, it is of consequence that some precincts contain more people than doothers. To make the sample of people in those precincts representative, thelarger precincts must have a greater likelihood of being selected thansmaller ones.

At this point, we need to introduce the concepts of confidence leveland sampling error. A sample rarely hits the true population value right onthe nose. The confidence level tells us the probability that the populationvalue will fall within a specified range. The sampling error specifies thatrange. A commonly used confidence level is 95 percent (it is sometimeshigher, but almost never lower). For a sample of 600, the sampling error is4 percent. What all this means is that if we took 100 samples from ourpopulation, and each of these samples consisted of 600 marbles randomlydrawn, then 95 out of 100 times we would be plus or minus 4 percent of thetrue population value. Our sample came up 65 percent red. While this figuremay not be exactly correct, we at least know that 95 out of 100 times (our95 percent confidence level) we are going to be within four points—oneway or the other—of the true proportion of red marbles in the barrel. Thismuch can be proved mathematically.

Let us turn to a political example. The Gallup Poll reported thatPresident Barack Obama’s approval rating after his first 100 days in officewas 69 percent (Washington Post/ABC News Poll, April 26, 2009). Sincethe sample size in this survey was 1,072, we know that if the poll wererepeated 100 times, 95 of the 100 repetitions (the 95 percent confidencelevel) would produce results that are plus or minus 3 percent (the samplingerror) of what we would find if we interviewed all American adults.8 Thus,Obama’s popularity on or about April 29, 2009, could have been as high as72 percent or as low as 66 percent. The best estimate (according to thecentral limit theorem) is 69 percent. Sampling error decreases as we moveaway from a 50/50 split, and it increases as the population has moreheterogeneous political attitudes. But by far the most important factor is thesize of the sample.

Page 11: American Public Opinion: Its Origins, Content and Impact

The importance of sample size for sampling accuracy can be easilydemonstrated by flipping a coin. We know that, in theory, if a coin isflipped honestly, a large number of times it should come up about equalproportions of heads and tails. Suppose we flipped a coin ten times. Given alarge number of repetitions of ten flips, we might frequently get sevenheads, eight tails, and so on. But if instead of flipping the coin ten times, weflipped it 100 times, our large number of repetitions would onlyoccasionally yield 70 heads or 80 tails, and if we flipped it 1,000 times, ourrepetitions would tend to cluster around 500 heads and 500 tails, with 700heads or 800 tails being rare events indeed. The larger the size of thesample (i.e., the number of times the coin is flipped), the closer we willcome to the theoretical expectation of one-half heads and one-half tails.

Unlike the size of the sample, the size of the population is of littleconsequence for the accuracy of the survey.9 That is, it does not make muchdifference if we are surveying the city of Houston or the entire UnitedStates. With a sample size of 600, the sampling error would be identical forboth the city and the nation—all other things being equal. Table 2.1 presentsthe sampling errors associated with specific sample sizes.

Using a large barrel of marbles and a table of random numbers, wedrew a “perfect” sample—that is, the sampling method fit perfectly with themathematics of sampling theory. When we sample humans, we cannot drawa perfect sample. Although a marble cannot refuse to tell us if it is red orgreen, a person may refuse to be interviewed. Sampling theory does notallow for refusals. Consequently, surveys of public opinion onlyapproximate the underlying theory of sampling.

TABLE 2.1 Sampling Error and Sample Size Employing SimpleRandom Sampling*

Sample Size Sampling Error (plus or minus)*2,430 2.01,536 2.51,067 3.0784 3.5600 4.0474 4.5

Page 12: American Public Opinion: Its Origins, Content and Impact

384 5.0267 6.0196 7.0150 8.0

* This computation is based on the assumptions of a simple random sampling with adichotomous opinion that splits 50/50 and a 95 percent confidence level.

Source: Compiled by the authors.

One way to compensate for the workings of chance in probability

surveys and for survey nonresponse is through poststratification weighting.We know from the Census the distribution of key demographic factorsamong the American population. If the sample does not match the Census,it can be weighted to bring it into line. For example, we know 12 percent ofthe American population is African-American. If a sample of 1,000contained just 8 percent African-Americans, it can be weighted to bring thetotal up to 12 percent. If this were the only weighting consideration, wewould multiply African-Americans by 1.50 and everyone else by .957. Thiswould raise the percent of African-Americans in the sample to 12 percentbut still keep the sample size at 1,000. The demographics normally used forweighting include age, gender, education, race/ethnicity, andhomeownership, although the choice of weight factors varies among surveyprofessionals. Weighting is an effective, but still imperfect, solution foradjusting samples that depart from known demographics. Those who refuseto be interviewed may be different from those who consent, and weightingcannot correct for this difference.10

Bad Sampling: Two Historical ExamplesHistorically, there are many instances in which sampling theory wasignored or those conducting the polls were ignorant. If the poll simplyconcerns political opinions (favor/oppose abortion, approve/disapprove oftax reform, favor/oppose the registration of handguns, etc.), there is noreality test. The survey may have been done badly and be considerably offthe mark, but how would one know?11 On the other hand, pre–Election Daysurveys have a reality test: Election Day. In these surveys, sampling

Page 13: American Public Opinion: Its Origins, Content and Impact

mistakes have, in several dramatic cases, cast opinion pollsters in a highlyunfavorable light.

To see how not to sample, it is worth reviewing two classic pollingmistakes from the past. One, discussed in Chapter 1, was the LiteraryDigest fiasco of 1936. The other was when the seemingly scientific polls allfailed in the presidential election of 1948. The Literary Digest, 1936 Chapter 1 discussed the prescientific straw pollsthat were used to gauge candidate fortunes prior to 1936. The best known ofthe commercial publications conducting straw polls was the LiteraryDigest. The Digest accurately forecast the winner (if not the exactpercentage points) of each presidential election between 1920 and 1932. In1936, as in previous years, it sent out some 10 million postcard ballots“drawn from every telephone book in the United States, from the rosters ofclubs and associations, from city directories, lists of registered voters, [and]classified mail order and occupational data” (Literary Digest, August 1936,3). About 2.2 million returned their postal ballots. The result was 1,293,699(57 percent) for Republican Alf Landon and 972,867 (43 percent) forPresident Franklin Roosevelt. On Election Day, Roosevelt not only won butalso won by a landslide, receiving 62.5 percent of the vote and carryingevery state except Maine and Vermont. The Literary Digest was not onlywrong; it was also wrong by 19.5 percent. On November 14, 1936, theLiterary Digest published the following commentary:

What Went Wrong with the Polls? None of Straw Votes Got Exactly the Right Answer—Why? In 1920, 1924, 1928 and 1932, the Literary Digest Polls were right. Not only right in the sensethey showed the winner; they forecast the actual popular vote and with such a small percentageof error (less than 1 percent in 1932) that newspapers and individuals everywhere heaped suchphrases as “uncannily accurate” and “amazingly right” upon us. … Well this year we usedprecisely the same method that had scored four bull’s eyes in four previous tries. And we werefar from correct. Why? We ask that question in all sincerity, because we want to know.

Why did the poll fare so badly? One reason that can be discounted issample size. The Literary Digest claims to have polled 10 million people(and received 2.2 million responses). Thus, a large sample is no guaranteefor accuracy. Rather, their sampling procedure had four fundamentaldefects. First, the sample was drawn in a biased fashion. The Digest clearly

Page 14: American Public Opinion: Its Origins, Content and Impact

did not use random selection or anything approaching it. Even thoughquestionnaires were sent out to 10 million people, a large part of the samplewas drawn from telephone directories and lists of automobile owners—during the Depression, a decidedly upper-middle-class group,predominantly Republican in its political sentiments. In other words, theDigest did not correctly specify the population. A second factorcontributing to the Digest’s mistake was time. The questionnaires were sentout in early September, making impossible the detection of any late trendfavoring one candidate or the other. Third, 1936 marked the emergence ofthe New Deal Coalition. The Digest had picked the winner correctly since1920 using the same methods as in 1936, but in 1936 voting becamepolarized along class lines. The working class and the poor votedoverwhelmingly Democratic, while the more affluent classes votedpredominantly Republican. Since the Digest’s sample was heavily biased inthe direction of the more affluent, it is not surprising that their sampletended to favor the Republican, Alf Landon.

Finally, there was the problem of self-selection. The Literary Digestsent out its questionnaires by mail. Of the 10 million they mailed, only alittle over 2 million were returned—about 22 percent. Those people whoself-select to respond to mail surveys are often quite different in theirpolitical outlook from those who do not respond. They tend to be bettereducated, to have higher incomes, and to feel more strongly about the topicsdealt with in the questionnaire (Dillman 2008). So even if the sample of 10million had been drawn in an unbiased fashion, the poll probably still wouldhave been in error due to the self-selection factor (Squire 1988). Afundamental principle of survey sampling is that one cannot allow therespondents to select themselves into the sample.

Despite the failure of the Literary Digest, several public opinionanalysts did pick Franklin Roosevelt as the winner. Among them wasGeorge Gallup, who built his reputation on a correct forecast in 1936. Interms of percentage, Gallup did not get particularly close. He missed byalmost 7 percent. But he got the winner right, and that is what most peopleremember. “Dewey Defeats Truman”: The 1948 Polling Debacle Following Gallup’ssuccess at predicting the 1936 election, “scientific” polling surged inpopularity and became perceived within the political class as almost

Page 15: American Public Opinion: Its Origins, Content and Impact

infallible. Gallup and other pollsters predicted the winners of thepresidential elections of 1940 and 1944. Then, in 1948 the pollsters allpredicted that the Republican presidential candidate, Tom Dewey, woulddefeat Truman. The pollsters were so confident of Dewey’s victory that theydid not even poll during the last few weeks of the campaign, seeing it as awaste of time. Virtually all politically knowledgeable observers saw theoutcome as certain. Famously, Dewey began to name his cabinet.

The 1948 presidential election jarred the pollsters from theircomplacency. President Truman defeated Dewey in an upset. Gallup’sprediction of Truman’s vote was off by 5 percentage points. While this errorwas actually smaller than that in 1936, the crucial fact was that Gallup gotthe winner wrong.12

Why were Gallup and the other pollsters wrong in 1948? Thetechnique used by Gallup in 1936 and up until the Dewey-Truman disasteris called quota sampling. This technique employs the Census to determinethe percentage of certain relevant groups in the population. Importantly,quota sampling is not scientific sampling. Respondents are not randomlychosen. For quota sampling, the pollster determines, for example, whatpercentage of the likely voters are male, Catholic, white, and collegeeducated. Within these groups, interviewers are then assigned quotas. Theymust interview a certain percentage of women, a certain percentage withless than high school education, a certain percentage of blacks, and so on.But there are few, if any, constraints as to which individuals in these groupsare to be interviewed.

The principal problem with quota sampling is a variation on theproblem of self-selection. The interviewer has too much opportunity todetermine who is selected for the sample. An interviewer who must, forinstance, get a specified number of females and a specified number ofblacks may avoid certain areas of town or may get the entire quota from asingle block. Experience with quota samples demonstrates that theysystematically tend to underrepresent the poor, the less educated, and racialminorities.

Research on Gallup’s misprediction in 1948 reveals he had too manymiddle- and high-income voters. One irony in Gallup’s misprediction is thattwo years before the Literary Digest fiasco, a Polish statistician namedJerzy Neyman (1934) published a landmark article demonstrating thesuperiority of random sampling to quota samples in survey work. So

Page 16: American Public Opinion: Its Origins, Content and Impact

powerful was his mathematical demonstration that quota samples becamean object of ridicule in scientific circles. During World War II, virtually allthe voluminous government survey work was based on some variant ofrandom sampling. The problems with quota samples had been known forsome time. However, for reasons of cost, Gallup did not think he couldabandon the quota method. After the 1948 election, he changed his mind, orhis mind was changed for him. A blue ribbon panel of statistical expertsvery publically criticized the quota samples used by Gallup and othercommercial pollsters (notably Archibald Crossley and Elmo Roper) asbeing scientifically indefensible (Frankel and Frankel 1987; Fried 2012).

Sampling Methods for In-Person SurveysFrom the 1930s into the 1980s, the predominant survey methodology was tointerview respondents in person, in their own homes. Earlier we describedsimple random sampling (SRS) to illustrate the principles involved indrawing a scientific sample. However, in-person surveys rarely employSRS. There is no master list of all Americans who could be sampled. Evenif there were, the persons selected would be widely scattered throughout thecountry, making in-person interviews prohibitively expensive. For example,an interviewer might have to travel to Kerrville, Texas, just to talk to onerespondent. Rather, polls where respondents are personally interviewed usemultistage cluster samples.

The first step in drawing a multistage cluster sample of the Americanelectorate is to divide the country into four geographic regions. Within eachregion, a set of counties and standard metropolitan statistical areas(SMSAs) is randomly selected. To ensure a representative selection of thenational population, each county or SMSA is given a chance to be selectedproportional to its population (probability sampling). National surveystypically include about eighty primary sampling units (PSUs). Onceselected, the PSUs are often used several years before they are replaced.The reason is economic: Compact geographic areas save time and money.Interviewers are expensive to train, but once trained they can be used overand over.

About twenty respondents are chosen within each PSU. First, four orfive city blocks (or their rural equivalents) are randomly selected. Then,four or five households are sampled within each block. It is at the

Page 17: American Public Opinion: Its Origins, Content and Impact

household level that the continued use of probability methods becomesmost difficult. Ideally, the interviewer obtains a list of all persons living inthe household and samples randomly from the list. And indeed, someacademic polls (e.g., the General Social Survey [GSS] and the AmericanNational Election Studies [ANES]) attempt to meet this ideal. Many surveyorganizations, however, abandon probability methods at the block level andrely on some type of systematic method for respondent selection. Theproblem is that specific individuals are hard to locate. Interviewers aregiven a randomly drawn starting point and then instructed to stop at everynth household. The interviewer first asks to speak to the youngest man ofvoting age; if no man is at home, the interviewer asks to speak to theyoungest woman. If both refuse, or if no one is at home, the interviewergoes to the next adjacent dwelling. Each interviewer has a male/femalequota and an age quota, but this is not a quota sample because theinterviewer cannot choose who gets into the sample.

Multistage cluster samples work well and are an efficient compromise,given the expense involved in a simple random sample approach. Thedrawback is that cluster samples have a greater sampling error than SRS, aswould be expected because the respondents are clustered into 75 to 100small geographic areas. A simple random sample of 1,000 has a 3 percentsampling error. Gallup reports that its in-person national samples(multistage cluster samples) of 1,500 have a 4 percent sampling error.

Telephone SurveysToday, most surveys are conducted by telephone (or over the Internet, seelater) rather than in person. The principal advantages are the speed withwhich telephone (and Internet) surveys can be completed and their lowcosts. They also free one from having to use clusters, as the physicallocation of a respondent is irrelevant in telephone and Internet surveys.However, there are disadvantages as well.

Nationwide, as many as a third of all telephone numbers may beunlisted. Those most likely to have unlisted numbers are younger, lowerincome, renters, and non-white households with children. A solution to thisproblem is random digit dialing (RDD). A ten-digit telephone number iscomposed of an area code (the first three numbers), an exchange (the nextthree), a cluster (the next two), and the two final digits. Since the

Page 18: American Public Opinion: Its Origins, Content and Impact

geographic assignment of area codes, exchanges, and clusters is publiclyavailable, one can define and sample a national population. The following isan example:

713-496-78 __ __

The first eight digits are randomly sampled from the population of72,000 digital residential codes. Then the last two digits are chosen by acomputer program that generates random digits. These methods bringpersons with unlisted numbers into the sample.13 On the negative side, somecalls are made to nonworking numbers. But the sample of households isrepresentative. Once the household is selected, it is then necessary tosample residents within it. Sophisticated methods to ensure randomselection have been developed by statisticians, but they require anenumeration of all household members by age and sex—information manyrespondents are unwilling to provide over the telephone. One simplealternative is the last birthday method. Interviewers ask to speak to theperson in the household who had the last—that is, the most recent—birthday. Another technique is to ask to speak to the youngest male, and ifnot at home, to the youngest female. Both these techniques give a goodapproximation of random sampling within the household (Daniel 2012).The latter has the advantage of always having an eligible respondent athome (the person with the last birthday may not be at home). The Cell Phone Challenge Telephone surveys only interview people withtelephones. Until recently, almost all households had landline telephones.This is not so today. Many potential survey respondents rely solely on cellphones, and the number is growing. According to the National HealthInterview Survey (a government-funded in-person survey), more than oneout of three households (34 percent) had only wireless telephones in 2012.Those adults most likely to rely only on cell phones are between eighteenand twenty-four years old, renters, attending school, and living withunrelated roommates (Blumberg and Luke 2012).

Cell phone-only households are now so prevalent that most surveyorganizations use “dual frame” samples based on a mix of cell phone-onlyhouseholds and households with a landline (the occupants of which mayalso have cell phones).14 The percent cell phone in any sample varies by

Page 19: American Public Opinion: Its Origins, Content and Impact

survey organization, but it can be as many as one-third.15 However, this“dual frame” strategy is expensive. Sometimes cell phone respondents areoffered a monetary incentive to participate in the survey (Link and Lai2011; Peytchev and Neely 2013). Also, the Telephone Consumer ProtectionAct of 1991 prohibits the use of auto dialers when calling cell phones. Thenumbers have to be dialed manually, which further adds to the expense. InJanuary 2008, Gallup announced its national polls would be based on dualframe samples (Moore 2008, 133). Among the 2012 presidential preelectionpolls, all eleven national telephone polls that were included in New YorkTimes columnist Nate Silver’s analysis of poll accuracy were dual framesurveys.16

Refusals and Nonresponse A significant problem with telephone surveys,although it applies to in-person surveys as well, is refusal. The telephone,however, particularly lends itself to abuse, as persons selling aluminumsiding, upholstery, or can’t-miss real-estate deals sometimes attempt to gaina respondent’s confidence by posing as a pollster. The sales pitch comeslater. A study by the Roper Organization showed that 27 percent of anational sample had experienced these sales tactics (Turner and Martin1984, 73). People, especially those in large cities, are becoming wary ofcallers claiming to be taking polls. However, by 2005 both the state andfederal government had instituted “do not call” lists, which bartelemarketers from calling telephone numbers registered on the lists. Thenational list contains millions of telephone numbers. However, the effect ofthese lists on survey response rates appears to be inconsequential. Responserates have continued to drop since 2005.

Nonresponse involves more than refusals. It includes any reason whythe designated respondent is not interviewed, such as ring-no answer orcontacting an answering machine. The response rate is computed as apercentage. In simple terms, it is the number of people successfullyinterviewed divided by the total number of people called.17 In the 1960s andearlier, it was common to find response rates in the 80 percent range(Moore 2008, 121).

Most pollsters believe that survey nonresponse is a serious problemand seems to be getting worse (Berinsky 2009; Keeter 2011; Peytchev2013). Holbrook, Krosnick, and Pfent (2008), analyzing over a hundredsurveys conducted by major field services between 1996 and 2005, found

Page 20: American Public Opinion: Its Origins, Content and Impact

that the response rate declined by a rate of two tenths of a percent eachmonth. The highly regarded Pew Research Center surveys saw responserates for their standard five-day telephone surveys drop from 36 percent in1997 to 9 percent in 2012.18 Even well-funded and highly influentialgovernment surveys, such as the University of Michigan’s Survey ofConsumer Attitudes, have experienced a significant drop in response ratesfor its monthly telephone survey of consumer confidence. In 1978, theresponses rate was 72 percent; by 2003, it had dropped to 48 percent. Evenmaintaining a 48 percent response rate has come at the cost of more time inthe field, more interviewer training, more calls to ring-no answer numbers,and sophisticated attempts to convert refusals into successful interviews bycalling respondents a second time.19 The principal cause of the drop hasbeen an increase in refusals, but there was also a significant increase innoncontacts beginning around the mid-1990s. For surveys where resourcesare not abundant and time windows are short (as is often the case in mediasurveys), response rates are significantly lower than for the Pew Center orfor the Survey of Consumer Attitudes (Curtin, Presser, and Singer 2005;Peytchev 2013).

To many, therefore, it comes as a surprise that in the face of decliningresponses rates there has been no commensurate decline in the accuracy ofpublic opinion surveys. The reason is that those sampled but notinterviewed differ little on social and political opinions than those sampledand interviewed. A great deal of research demonstrates that low responserates do not bias survey results. Since 2000 the Pew Research Center hasbeen conducting studies of the consequence of low response rates. Evenwith a response rate of just 9 percent in its 2012 study, the survey datamatched up well with reliable benchmarks provided by the Census andother government agencies. For instance, there was no significant differencein home ownership, having children in the household, being a currentsmoker, receiving food stamps, unemployment benefits, or being married.20

Still, it begs the question of how can polls with a 9 percent responserate be essentially accurate? The answer is that the decision to respond to apoll is probabilistic. It is not the case that potential survey respondents canbe divided into two discrete class—those willing and those never willing toparticipate. Rather, nearly everyone has some probability of consenting toan interview, which is dependent on mostly idiosyncratic circumstances atthe time the call is made. If a potential respondent is watching a closely

Page 21: American Public Opinion: Its Origins, Content and Impact

contested athletic event, the probability that he or she would decline theinterview is quite high. On the other hand, if the athletic event is a blowoutby one team, the probability of consenting to the interview increases. Butthese circumstances at the time of the call are random.

In an analysis of the research on nonresponse, Groves (2006) foundlittle evidence of any direct link between response rates and biased surveyresults. Groves found occasional bias but, crucially, that its degree was notpredictable from the response rate. Rather, it was predicted by other factorsthat were unique to each survey, such as the correlation between the surveysubject matter and the willingness of people to participate. Grovesconcluded that “blind pursuit of high response rates in probability sample isprobably unwise” (668). Still, there is a limit as to how far we can take thisprinciple, with response rates now hovering in single digits.

Interactive Voice Response (Robotic) TelephonePollsGiven the rising cost of high-quality telephone surveys, some have turnedto still new methods of polling that reach more respondents for less money.One innovation is robotic polling, whereby respondents punch buttons ontheir telephone in response to the instructions of automated voices (robots).

Robotic polls are most prevalent at the state or local level, where theyare typically conducted for media clients who wish to acquire fast andcheap information about where the public is leaning in the next election. Inrobotic polls (or as their sponsors prefer to call them, “interactive voiceresponse” polls), the survey research firm samples telephone numbers byRDD with an auto dialer. The person who answers the telephone hears therecorded voice of a professional announcer describing the poll and receivesinstructions for answering questions via punching numbers on thetelephone. The key difference from regular polling—but a big difference—is that there is no human being at the controls on the pollster’s side of thecommunication.

The allure of automatic polls is the ability to obtain many interviews atlow cost. There are no interviewers to train or hire. A robotic poll canreadily conduct over a thousand surveys within three days or even soonerupon demand. At least one robotic poll firm—Survey USA—claims a

Page 22: American Public Opinion: Its Origins, Content and Impact

record of considerable accuracy with only about 2.33 percent as the averageerror for general election surveys. For the 2012 presidential election,Survey USA missed the true vote division by 2.2 percent.21 Still,considerable controversy follows robotic polls. Their most obviouschallenge is that they offer no control over the person in the household (achild? a visitor?), who actually answers the telephone and punches a set ofreplies. And it is not in everybody’s nature to respond to automated surveys.Response rates are at best in single digits and may even fall below 1percent. And, as previously noted, federal law forbids the use of auto dialersfor cell phone numbers. Another problem is the inability of robotic polls togather much contextual data on the respondent. Vote choice and a couple ofdemographics are usually the best these polls can provide. Therefore,robotic polls are mostly limited to the horse-race question and not muchelse. Also, one should not conclude simply because robotic polls haveshown some success in predicting election outcomes that they provideunbiased estimates in other public opinion domains. Many politicalobservers who follow the polls closely are reluctant to treat robotic pollsseriously, preferring to focus on those conducted by humans.22

Internet PollsAs noted earlier, the 2012 Census reports that 80 percent of adultAmericans have access to the Internet. Of course, one cannot simply sendout a survey over the Internet and expect to get meaningful results. Still,given that level of Internet penetration, polls over the Internet are abooming business, in both the commercial world and the academic world.According to Gary Langer, a public opinion consultant and former directorof polling for the ABC television network, in 2009 there were about asmany polls done over Internet as done by telephone.23 The ratio favoring theInternet is only likely to rise, given the increasing cost of telephone surveysand the often quite inexpensive cost of some Internet surveys.24

Polling via the Internet allows for a certain degree of sophistication, ascomplex questions can be combined with visuals. However, Internet pollshave a number of drawbacks when compared to in-person and telephonepolls. First, there is the problem of coverage error. The 20 percent ofAmericans without Internet access have a zero probability of getting intothe sample. These 20 percent are quite different from those with Internet

Page 23: American Public Opinion: Its Origins, Content and Impact

access, and the problem cannot be solved by simply weighting the data toCensus benchmarks (Moore 2008). Second, Internet samples are notprobability (or random) samples. Unlike RDD, there is no finite master listof e-mail addresses from which to draw a sample. Consequently, Internetsamples are not scientific. Thus, strictly speaking, it is not permissible toreport sampling error or associated measures like “statistical significance.”Third, most recruiting for Internet surveys is done over the Internet. Thismethod introduces the problem of self-selection, the same problem thatoccurred with the Literary Digest in 1936. In the case of polls aboutpolitics, those with an interest in politics are more likely to “self-select” toparticipate in an Internet survey than are those with little or no interest. Anextensive analysis of Internet polling methods by the American Associationfor Public Opinion Research reached the conclusion that “researchersshould avoid nonprobability online panels when the one of the researchobjectives is to estimate population values.”25 Because they are notscientific, reputable organizations like the Pew Research Center refuse touse them, and the New York Times refuses to publish the results of anysurvey based on Internet polling that is not probability based.

There have, however, been some serious attempts to use the Internetapplying either scientific methods or methods that while not being strictlyscientific, proponents claim are sufficiently rigorous as to yield reliableresults. We shall review two of these methods. The first of these is theInternet probability panel.26 Address-based sampling is used to select arepresentative sample of respondents (Link and Lai 2011). That is, a sampleof tens of thousands is scientifically drawn from all mailing addresseshoused by the post office, which is then randomly sampled to form theInternet panel. Those selected for the sample are contacted by telephone ormail and offered payment to participate in the Internet surveys (usuallyabout one survey a week). Those without Internet access, who agree toparticipate, are given web–TV connections and equipment so they too arepart of the sample. The sample is, therefore, scientific. The principaldownsides are first the method is expensive. It is as expensive as or moreexpensive than dual frame telephone surveys. Second, critics claim there isa problem with “panel conditioning.” Respondents are interviewed up tofifty times a year. After numerous interviews, it is alleged that panelrespondents differ significantly from the general population (Dillman2008).

Page 24: American Public Opinion: Its Origins, Content and Impact

A second approach, which is not scientific, but employs a rigorousstatistical methodology to minimize bias, is an opt-in Internet survey using“sample matching.”27 A large number of people (a million plus) arenonscientifically recruited for an online panel through a variety of methods,including online advertising (pop-up ads) as well as targeted advertising torecruit people typically underrepresented in opt-in Internet surveys (Groveset al. 2009). Respondents are offered an incentive to participate. From thismillion-plus member database, a sample for an individual study of publicopinion is randomly drawn so that it matches the characteristics of the targetpopulation—such as age, race, gender, education, and region.28 Thismethodology works well in predicting the presidential vote. Based onsurveys conducted twenty-one days before the election, one vendor(UGOV) was off the mark by 2.6 percent in 20 1 2.29 Data from Internetsurveys have now essentially replaced telephone surveys in the mostprestigious peer-reviewed journals in political science. Between 2010 and2012, thirty-nine articles using Internet data were published in the AmericanPolitical Science Review, the American Journal of Political Science, andthe Journal Politics.30

Page 25: American Public Opinion: Its Origins, Content and Impact

2-3 Question WordingIt should surprise no one that in survey research, as in everyday life, theanswers received are often dependent on the questions asked. This realityhas both an upside and a downside for opinion polling. The upside is thatpeople are sensitive to the way survey questions are phrased. That meansthey are paying reasonably close attention to what they are being asked, andtheir answers can be taken seriously. If changes in question wording had noimpact, the opinion poll would likely be too blunt an instrument to be ofmuch use. The downside is that because the distribution of public opinioncan be a function of question wording, incorrect or inappropriate inferencesabout its true nature are easily possible. Not all variations in questionphrasing affect the distribution of responses, but in some cases they clearlydo. In this section, we review some of the aspects of question wording thatcan affect the distribution of public opinion.

Question AsymmetriesTwo questions that seem logically equivalent can sometimes yielddiametrically opposite results. The best known example is greater supportfor free speech when respondents are asked if the United States should“forbid” speeches against democracy versus should the United States“allow” speeches against democracy. Although to support free speechmeans answering “no” to the forbid question and “yes” to the allowquestion, the two are logically equivalent. Typically, 25 percent or more say“yes” to the allow option than “no” to the forbid option (Hippler andSchwarz 1986). In 2012, Gallup found 51 percent of Americansdisapproved of the “financial bailout” of the banks, while in the same weekPew reported 66 percent of the public thought the “government loans” tothe banks were mostly “good” for the economy. The difference is likely dueto the harshness of the terms—forbid and bailout are more likely to grate onthe ears than are allow and government loans.

Multiple Stimuli

Page 26: American Public Opinion: Its Origins, Content and Impact

Many questions have more than one stimulus to which respondents mayreact. Sometimes these are just bad questions. Take, for instance, thefollowing question from a 2007 Newsweek Poll: “Do you think thescientific theory of evolution is well-supported by evidence and widelyaccepted in the scientific community?”31 How does the substantial numberof people who believe the theory of evolution is not well supported by theevidence but is widely accepted by the scientific community respond to thisquestion?

In other instances, such questions are not faulty but are often anecessary requirement for measuring opinions on complex topics. Examplesof commonly used single-stimulus questions are “Generally speaking, doyou consider yourself a Democrat, an Independent, a Republican, or what?”or “Do you approve or disapprove of the way the president is handling hisjob?” But suppose we were interested in public support for the FirstAmendment right of free speech. We could ask people to agree or disagreewith the statement, “I believe in free speech for all, no matter what theirviews might be.” In 1978, 89 percent of opinion-holders agreed (Sullivan,Piereson, and Marcus 1982). But that tells us little other than in the abstractthere is near consensus on a fundamental principle of the American creed.However, real controversies over free speech do not occur in the abstractbut in concrete circumstances. Thus, we might ask, “Suppose an admittedCommunist wanted to make a speech in your community. Should he beallowed to speak or not?” In 2008, only 66 percent of the public said theperson should be allowed to speak. If the admitted Communist was teachingin a college, 37 percent said he should be fired. It is quite clear that inspecific circumstances, people are responding to both the free speechstimulus and the Communist stimulus. As we shall see, the level of supportfor free speech is a matter of considerable controversy (see Chapter 6).

Many less obvious instances of multiple stimuli can affect thedistribution of public opinion. One way is to preface the question with thename of an authoritative or admired person or institution. Consider theresponses of white Americans to different phrasings of a question about thecontroversial policy of minority set-asides (Sniderman and Piazza 1993,113).

Form A Form B

Page 27: American Public Opinion: Its Origins, Content and Impact

Sometimes you hear it saidthere should be a law to

ensure that a certain numberof federal contracts go tominority contractors. Do

you favor or oppose such alaw?

The Congress of the United States—both the House of Representatives andthe Senate—has passed a law to ensure

that a certain number of federalcontracts go to minority contractors.Do you favor or oppose such a law?

Favor law 43% Favor law 57%Oppose law 57% Oppose law 43%

100% 100%

Invoking the symbol of the law in this question transforms the policy

of set-asides from one backed by a minority of white respondents to onebacked by a clear majority—a change that may not be trivial in the politicsof affirmative action. The use of authoritative symbols also provides a cuefor the uncertain, decreases the don’t-knows, and raises the percentage ofthose favorable and unfavorable among those who know little or nothingabout the issue in question (Smith and Squire 1990; Bishop 2005, 33–36).

Question Order EffectsBy order effects, we mean the effect the previous content of the interviewmight have on a specific question. For example, respondents usually desireto be consistent in their answers. When asked back in 1980 if “the UnitedStates should let Communist newspaper reporters from other countriescome here and send back to their papers the news as they see it,” 55 percentsaid yes. But when the question was preceded by one that asks whether “aCommunist country like Russia should allow American newspaperreporters to come in and send back the news as they see it,” then 75 percentfavored letting Communist reporters into the United States (Schuman andPresser 1996, 27; Schuman 2008). Since most answered yes to allowingAmerican reporters into (Communist) Russia, it was obviously inconsistentto bar Communist reporters from the United States. With respect to theabortion issue, general support for abortion will increase if a specific itemwith a high level of support is asked first. Thus, if respondents are asked ifabortion should be permitted “if there is a strong chance of serious defect in

Page 28: American Public Opinion: Its Origins, Content and Impact

the baby” (84 percent say yes), general support for abortion is 13 percentgreater than if the specific abortion item is not asked first (Schuman andPresser 1996, 37).

Other examples of order effects include the questionnaire placement ofpresidential popularity or presidential vote items. Incumbent presidents doworse on both if the question comes late in a survey as opposed to early.The reason is that content of the interview reminds respondents oftroublesome aspects of the incumbent’s presidency (Sigelman 1981).Pocketbook voting (i.e., voting on the basis of changes in one’s personalfinances) can be induced by asking questions about finances immediatelybefore ascertaining vote choice (Sears and Lau 1983). Finally, pollsterssometimes provide information on somewhat obscure topics before askingopinions about the topic. Suffice it to say that the content of the informationprovided in the stem of the question can easily affect the answer given(Bishop 2005).

Balanced ArgumentsIf a question mentions only one side of a controversy, that side will oftenget a disproportionate number of responses. An example is, “Do you favorthe death penalty?” Because only one option (“favor”) is given, we mightexpect that option would be more frequently chosen than if the question isbalanced, as in “Do you favor or oppose the death penalty?” The same istrue for questions that use an agree/disagree format (called Likert scales).The common practice is to inform the respondent of both agree anddisagree options in the stem of the question.

At best, however, agree/disagree questions present only one side of theissue. Respondents must construct the other side of the issue themselves. Inthe following example, providing balanced alternatives on a question aboutunemployment yields quite different results compared to an agree/disagreealternative (Cantril 1991, 126).

Form A Form BDo you agree ordisagree with thisstatement: “Any

Some people feel that any able-bodied personcan find a job and make ends meet. Othersfeel there are times when it is hard to get

Page 29: American Public Opinion: Its Origins, Content and Impact

able-bodied personcan find a job andmake ends meet”

along and some able-bodied people may notbe able to find work. Whom do you agreewith most?

Agree 65% Can make ends meet 43%No opinion 10% No opinion 18%Disagree 25% Sometimes to get along 39%

100% 100%

The Middle PositionThe public seldom splits into polar camps on any issue, and even when amiddle position is not offered, many will volunteer an intermediatealternative. With no middle position offered, a respondent might claim “noopinion,” a response that might actually mean that the respondent hasthought about the matter but is cross-pressured and is conflicted betweenthe two sides of the issue (Treier and Hillygus 2009).

When the question wording explicitly includes a middle position,many respondents seize this middle response as the safe alternative. In onestudy, Schuman and Presser (1996, 166) found that when respondents areasked on political issues if they “are on the liberal side or the conservativeside,” 16 percent volunteer that they are “middle of the road.” But if the“middle of the road” option is offered, it is taken by 54 percent. While thereis little doubt about the consequence of offering a middle position (or an“unsure” or “undecided” option for issue questions), there is no consensusabout the desirability of doing so. Some argue that most respondents dohave a preference, albeit in some cases a weak one. Offering the middlealternative as an in-between sort of response encourages respondents not toexpress their opinion (Gilljam and Granberg 1993; Schuman 2008).

When questions are framed to offer a middle or compromise position,public opinion can appear in a different light than when they are presentedwith only a pair of two stark alternatives. For instance, Gallup asked twoquestions about the death penalty in May 2003, less than three weeks apart.When the public was asked if it favored the death penalty, 70 percent saidyes. But when Gallup presented an option of life in prison without thepossibility of parole, support for the death penalty dropped to 53 percent.Similar patterns are found, for instance, on the issue of school prayer. When

Page 30: American Public Opinion: Its Origins, Content and Impact

asked about the matter, the vast majority of respondents voice favor forofficial prayers in public schools, despite their unconstitutionality. Butwhen respondents are given the added option of moments of silence (whichcould be used for private prayer), much of the support for official prayermoves to this new alternative.

Response AcquiescenceOften when people are asked whether they agree or disagree with anabstract statement, or one about which they know little, they tend to agreerather than disagree. This tendency is called response acquiescence. Itarises mostly because the public is ill informed and may not have genuineopinions on many issues (see Chapter 3). We can illustrate this with a 1986study of attitudes about pornography. Respondents were asked whether theyagreed or disagreed that “people should have the right to purchase asexually explicit book, magazine, or movie, if that’s what they want to do.”Evidently, people saw this statement as a referendum on individual rights,since an overwhelming 80 percent agreed with the statement. However, inthe same poll respondents were also asked whether they agreed or disagreedwith the opposite statement that “community authorities should be able toprohibit the selling of magazines or movies they consider to bepornographic.” Sixty-five percent agreed with this opposite view as well.32

Overall, a large number of respondents agreed that the public had a right topurchase pornographic materials and also agreed that the community hasthe right to stop people from purchasing the same material. The most likelyexplanation of this contradiction is response acquiescence. These examplesindicate the danger of using single Likert-type items to measure the level ofpublic opinion. The best strategy for dealing with response acquiescence isto use two or more items where half are worded positively and halfnegatively—that is, to support free speech the respondent must agree withone statement and disagree with another.

Filter QuestionsMost people believe that good citizens should have opinions on currentpolitical topics. Thus, when interviewed by polling organizations, they tendto offer opinions on subjects about which they know little or nothing. In an

Page 31: American Public Opinion: Its Origins, Content and Impact

otherwise conventional survey of the Cincinnati area, Bishop and hiscolleagues (1980) included a question asking respondents whether thenonexistent “Public Affairs Act” should be kept or repealed. Eighteenpercent thought it should be kept; 16 percent favored its repeal. Given thesefindings on a fictitious issue, it should not be surprising that on real butsomewhat obscure issues many will offer opinions on matters about whichthey have no true attitude.

One way to reduce this problem is by using filter questions.Respondents are read a question and response alternatives and then asked,“Where would you place yourself on this scale, or haven’t you thoughtmuch about this?” Using the 2004 NES question that asks people to indicateon a 1–7 scale if they feel more or less money should be spent on defense,13 percent of the sample indicated they had not thought enough about theissue to have an opinion. But that does not mean that all the remaining 87percent have a true attitude. Another 23 percent took the middle position onthe scale. It seems likely some taking this position have either no attitude orone that is poorly developed.

One danger in using screening questions is that some people with realattitudes, but who are hesitant to express them, may be filtered out ofsurveys (Berinsky 1999). An alternative is to use branching questions,where those who say they are undecided or unsure are encouraged infollow-up probes to make a choice (Smith 1984; Krosnick and Kerent1993). But, of course, this strategy runs the risk of eliciting opinions whereno true underlying attitude exists.

Open-Ended QuestionsMost survey questions are “closed ended,” asking the respondent to choosefrom a fixed menu of options. Sometimes, however, survey researchers ask“open-ended” questions, without the prompt of a fixed set of alternatives.With open-ended questions, classifying the responses is a multistep processwith the interviewer recording the answer, and then coders classifying theanswers according to predetermined categories. With the closed-endedversion, almost all respondents choose one of the options that are offered.The open-ended version generates a wide range of responses, often with ahigh percentage offering no opinion due to the difficultly of instant recallrequired by surveys.

Page 32: American Public Opinion: Its Origins, Content and Impact

One frequently used open-ended question asks: “What is the mostimportant problem facing the country today?” Schuman and Presser (1996)compared this open-ended format to an identical closed-ended format,where the latter respondents were given a choice among seven possibleoptions. For alternatives such as “inflation,” the open and closed formatsyielded the same results. But twice as many people mentioned “crime andviolence” in the closed-ended format (35 percent) as in the open-endedformat (16 percent). The authors claim that listing “crime and violence” inthe closed-ended format legitimized the option, if only because the words“facing the country today” used in the stem of the question imply nationalproblems. Many think of crime as a local problem. In this instance, theextent to which crime and violence was seen by the public as the mostimportant problem facing the country is very much dependent on how thequestion was asked.

Naturally, any option that is offered in a closed-ended question islikely to be chosen more frequently than from the comparable open-endedquestion. One illuminating example is how the format affects the responsesfollowing 9/11 regarding who was responsible for the attacks. FewAmericans volunteered that Saddam was behind the 9/11 attacks whenasked the open-ended version. But when explicitly presented with thispossibility as an item from the menu of choices in a closed-ended format, asurprisingly large percentage—often more than 50 percent—saw Saddam asbeing personally involved (Althaus and Largio 2004). For some, beingspecifically given the option that Saddam may have been involved allowedfor a more efficient search of “considerations” in memory needed toconstruct an opinion.

Social DesirabilityRespondents want to make a good impression on those who interview them.Even though the interviewer is a total stranger, survey respondents oftenfeel pressured to give answers that conform to widely held social norms.This phenomenon, often called social desirability (or impressionmanagement), results in reported answers that are more socially acceptablethan the respondent’s true attitude or belief. For example, people overreportvoting. The 2008 election provides one example. Although only 62 percentof eligible citizens cast votes, over 80 percent claimed to have voted in

Page 33: American Public Opinion: Its Origins, Content and Impact

NBC News/Wall Street Journal polls after the election. Good citizens vote,and to avoid a negative impression, some nonvoters “recall” having voted.

Other examples abound of possible survey contamination from socialdesirability. Respondents often approach questions about religion as if thesurvey were a referendum on their faith and overreport the extent to whichthey attend church (Bishop 2005). During the 2008 Democratic presidentialprimary campaign, pollsters worried about respondents overreporting avoting preference for Obama on the grounds that it would appear sociallyundesirable to vote against a plausible black candidate. This concern wasshown to be unfounded (Hopkins 2009). And survey respondents mayoverreport their enthusiasm for platitudes such as free speech, a concern weaddress in Chapter 6.

One way to overcome social desirability bias and gain an accuratereading of public sentiment is via the “list experiment.” By this procedure,respondents are randomly assigned to two groups. The baseline group gets alist of statements and is asked the number of statements on the list that itagrees with. The experimental group gets the same list plus one morerepresenting the sensitive item of interest. The idea is that the differentialbetween the two groups in the mean number of statements they agree withunobtrusively measures the proportion who endorses the sensitivestatement. The reason this works is that respondents know that the surveyinvestigator can only know the number of statements agreed with, and nottheir views about any one actual statement (Glynn 2013).

Here is one example. A list experiment was conducted to ascertainwhat proportion would accept a woman as president. It turns out thatsupport plummeted 20 percent compared to the results of the usual questiondirectly asking people whether they would accept a woman president (Strebet al. 2008). Evidently, the usual survey support for a woman president isboosted somewhat by social desirability.

The Question Wording ExperimentFor scientific research, the gold standard is the experimental design.Average citizens most frequently encounter this method when the mediareports on trials involving the effectiveness of a new drug. Half of thesample randomly gets the drug; the other half randomly gets a placebo(sugar pill). If the half that gets the drug differs significantly on key

Page 34: American Public Opinion: Its Origins, Content and Impact

indicators from the half that gets the placebo, the drug is deemed effective—at least in the case of this one trial. While question wording experimentshave been around for a long time (J. Converse 1987), it is only with theadvent of computer-assisted telephone interviewing (CATI) that thepotential of this method has been fully realized. In a simple surveyexperiment, half of the sample might randomly get a question onaffirmative action for “blacks” and the other half would get exactly thesame question, but “women” would be substituted for blacks (Snidermanand Piazza 1993). This would allow the investigator insight into whetheropposition to affirmative action was based on principle or on racialresentment. CATI and Internet surveys allow for much more complexexperiments, where respondents are assigned multiple different versions ofan interview (Feldman and Huddy 2005; Huber, Hill, and Lenz 2012). Thesurvey experiment is a strong design because (1) the experiment is invisibleto the respondent, (2) there is the capacity to make strong causal inferencesowing to the experimental design, and (3) there is the ability to generalizefrom a sample to a population (Tedin and Weiher 2004).

Summary: The Pitfalls of Question WordingVariations in question wording can clearly influence the distribution ofpublic opinion. That fact should be recognized and considered when usingopinion data. Some questions are obviously biased (see next section). Readthe questions carefully. What are the stimuli? Are authorities beinginvoked? Are inferences being made about the level of opinion from singleLikert items? Are “no opinion” screens used on obscure issues? Arebalanced questions truly presenting both options in a fair manner? Is theformat open ended or closed ended? Are there reasons to think answers areinfluenced by impression management? All these possibilities are cause forconcern. Some types of inferences are more reliable than others. Multiplequestions on the same issue can give insight into the range of opinion. Butthe most reliable strategy is to compare responses to the same question.Thus, subgroup differences on the same question are usually quitemeaningful, as are differences when groups are randomly assigned differentquestions. Also effective is comparing change over time in response to thesame question. Because the stimulus is identical, variation is likely due to achange in the public’s view.

Page 35: American Public Opinion: Its Origins, Content and Impact

2-4 Polls and Predicting ElectionsHow accurate are public opinion polls? The one good test occurs onElection Day. Only when comparing the polls to actual election outcomescan the accuracy of the polls be tested objectively. Predicting electionspresents the pollsters with the usual challenges to polling accuracy, plusmore. Like all other public opinion polls, election polls must deal with theproblem of frequent refusals. What if, for instance, supporters of onecandidate or party were more likely to agree to be interviewed? Andelection polls are not immune from even the problem of question wording.There are many ways to ask respondents whom they will vote for, and thesesometimes have consequences (McDermott and Frankovic 2003). Whentrying to predict elections, pollsters face one further challenge that does notarise when simply trying to predict general opinion. That challenge is thatpollsters must take into account who votes and who does not. And votersare always prone to overestimate their likelihood of voting.

We will see below that preelection polls have a strong record ofaccuracy. Yet even polls that are correctly executed do not always predictthe actual Election Day winner. Different polls administered during thesame time interval sometimes show conflicting numbers. Occasionally pollnumbers swing wildly over a short period. These characteristics do notnecessarily mean the polls are defective, but they do require explanation.

The media tends to attribute more accuracy to preelection polls thaneven the best designed polls can possibly deliver. First, an obvious point.Preelection polls refer to sentiment at the time they are conducted, not onElection Day. The horse-race analogy commonly used to discuss electionsis, in this case, appropriate. The horse that is ahead going into the stretch isnot always the horse (or the candidate) that wins. But even given the overalltrend, there are still “house effects”—variations in survey results due toidiosyncratic ways in which survey organizations conduct their polling. Forexample, there is surprisingly great variation in the way the candidate-choice question is asked in presidential elections.33 Survey “houses” varyfurther in such matters as their frequency of callbacks to elicit responsesfrom refusers, how they poststratify and weight their respondents, how theyallocate “undecideds,” and how they measure and adjust for therespondent’s likelihood of voting. Thus, polls from different firms can be at

Page 36: American Public Opinion: Its Origins, Content and Impact

some variance with one another during the campaign for a variety ofreasons.

Most importantly, we should remember that polls will inevitably varybecause of ordinary sampling error. Commentators during electioncampaigns have a tendency to overinterpret shifts in the polls as meaningfulwhen they actually represent house effects or sampling error. One study ofthe preelection polls during the fall of 1996 showed that the evidentmovement of the polls, a subject of much interpretation by the popularmedia, was no greater than would be expected from chance alone, givenroutine sampling error (Erikson and Wlezien 1999). Preferencesundoubtedly did shift back and forth during the 2000 Bush-Gore campaign.But in the Bush-Kerry campaign of 2004, the polls were so stable within anarrow range where, arguably, virtually all the “changes” from one poll tothe next could have been due to nothing more than sampling error. Even in2008, when the presidential campaign was conducted during an economicstorm, the horse race changed surprisingly little.

Allocating UndecidedsTypically, surveys show 15 percent or more of the electorate to beundecided at any point during a presidential campaign (the figure is muchhigher for lower-level races), winding down to 5 to 8 percent a few daysbefore the balloting. Survey houses differ in the way they handle theseundecideds. While common sense might dictate a simple reporting ofcandidate preferences plus undecideds, most media polls do not follow thatstrategy. If voters claim to be unsure, they are pressed in a follow-upquestion to indicate which candidate they “lean to.” Voters who maintainthey are really undecided must be quite firm in their conviction in the faceof pressure to make a choice. The inevitable result of this strategy is asizable swing in candidate preferences, mostly accounted for by thoseforced to make a choice when their initial reaction was undecided (Crespi1988; Moore 1992).

Late movement by undecided voters can determine the election, yet goundetected if polling ceases before Election Day. The flow of theundecideds was one reason Truman beat Dewey (the final Gallup Pollshowed that 19 percent of the electorate was undecided). And it is thereason Ronald Reagan soundly defeated Jimmy Carter in 1980, when polls

Page 37: American Public Opinion: Its Origins, Content and Impact

showed the contest to be very close, and why Ross Perot performed betterin 1992 than almost all preelection polls indicated.

If undecideds in preelection polls do not split evenly on Election Day,how, then, is one to allocate them before the election? There is no onecorrect answer for all circumstances. Sometimes the undecideds arereported as just that—“undecided.” However, the poll will thenunderestimate the percentage of the vote going to a candidate who hasmomentum. Sometimes undecided respondents are asked to rate candidateson a 0–10 scale. The undecideds are then allocated to the candidate theylike best. If there is a tie on that question, party identification is used toassign the undecideds to a candidate. Other polls exclude the undecidedsaltogether and base their preelection forecasts only on those voters willingto make a choice.

Deciphering Who VotesUndoubtedly, the most fundamental problem with pre-Election Day polls aspredictors is an inability to define the population properly. Recall that arandom sample must be drawn from a finite population. However, it isimpossible to precisely define the Election Day population. Only about one-half of adult Americans vote in presidential elections. No one has yetdetermined a method that predicts with great accuracy who will vote. Yet itis those Election Day voters who constitute the population, not adultAmericans or even registered voters. In a preelection survey in 1996, thePew Research Center asked a national sample of those over age eighteen ifthey were “absolutely certain,” “fairly certain,” or “not certain” to vote inthe upcoming election for president.34 Here are the results:

Absolutely certain 69%Fairly certain 18%Not certain/unsure 13%

100%

Actual turnout in 1996 was about 49 percent.

Page 38: American Public Opinion: Its Origins, Content and Impact

Since respondents’ self-reports greatly inflate their likelihood ofvoting, survey researchers try to screen voters by objectively assessing theirlikelihood of voting. If all respondents who claimed they would vote wereallowed through the screen, the projected election results would be distortedin favor of the party whose supporters are least likely to vote. Historically,the poor and the less educated are the least likely to vote, and these peopletend to support Democratic candidates. Election results are also affected bythe level of excitement of each party’s supporters, which can vary fromelection to election. Pollsters seem to do a good job of screening in the daysleading up to election, when voter interest in the election is fairly easy toascertain. But screening becomes a challenge to pollsters in the early daysof the campaign. The danger is that shifts in the polls of likely voters areoften due not to people changing their mind but rather to change in thecomposition of who counts as a likely voter. Democrats’ and Republicans’relative excitement varies in the short term so that first one party and thenthe next will be overrepresented among likely voters in a way that has littlebearing on the vote later on Election Day, when their level of excitementmatters (Erikson, Panagopoulos, and Wlezien 2004).

To screen their preelection samples for nonvoting, pollsters first askrespondents whether they are registered and then ignore the choices of theunregistered. Registered respondents are asked a set of questions designedto measure their likelihood of voting. Pollsters use one of two methods atthis stage. By the cutoff method, they divide registered respondents intothose more and less likely to vote (above or below the cutoff value of votinglikelihood) and then count the preferences only of the most likely. (Forinstance, if they anticipate a 60 percent turnout, they count the preferencesof the 60 percent most likely to vote.) By the probable electorate method,the pollster weights registered respondents by their likelihood of voting. Forinstance, the preferences of a voter estimated at 30 percent likely to votewould be weighted only 0.30 as much as a certain voter. Which method isbetter is a matter of debate (Erikson 1993; Ferree 1993; Hugick andMolyneux 1993; Daves 2000). The important point is that the type of screenand its application can have an important consequence for a poll result.

How Accurate Are the Polls?

Page 39: American Public Opinion: Its Origins, Content and Impact

Given the daunting challenges that pollsters face, one might expect thatpolls do a poor job of predicting elections. But that would be wrong. Pollsconducted early in campaigns often give a misleading view of what willtranspire on Election Day. But that is because events change beyond thepollsters’ control. Similarly, primary elections are difficult to pollaccurately even in the final week of the campaign, as preferences leading upto primary elections are historically volatile. (Voters focus late, and theychoose within their party rather than between parties.) But general electionpolls conducted late in the campaign, such as in the final week, show anexcellent record of accuracy. This is reassuring not just because we mightsometimes want to know outcomes in advance. It assures us that opinionpolls should also be reasonably accurate as snapshots of public opinion atthe moment, even when we lack the feedback of an election to verify.

Figure 2.1 summarizes the accuracy of presidential polls over sixteenelections, from 1952 to 2012. This graph shows the actual vote (measuredas the Democratic percent of the two-party vote) as a function of theDemocratic share of the two-party vote in the polls during the final week ofthe campaign. (The Republican vote in the polls or on Election Day is 100percent minus the Democratic percent.) The vote in the polls is measured bypooling all polls available where the middate of the polling period is withinone week of Election Day.

In Figure 2.1, the polls would be perfectly accurate if the sixteenannual observations all fell on the diagonal line. They do not of course,since polls are not perfect. Rather, the late polls tend to be accurate withinabout 1.5 percentage points. This is less than one would expect by chancefrom sampling error alone, but we must keep in mind that the polls cannotregister the late movement in the very final days of the campaign.Historically, the late polls show no bias in favor of one party’s candidates orthe others. And, contrary to some claims, they do not inflate support forincumbents. One can see from Figure 2.1 that polls tend to exaggerate thesizes of large leads in landslide elections (e.g., 19 64, 19 72).35 But whenelections are close, presidential polls tend to be quite accurate.

Page 40: American Public Opinion: Its Origins, Content and Impact

FIGURE 2.1 Democratic vote for president (as share of two-partyvote) by Democratic strength (as share of two-partyvote) in all polls (combined) in the final week of thecampaign, 1952–2012. The 2012 result is highlighted.

Source: Compiled by Robert S. Erikson and Christopher Wlezien.

Presidential polls have been remarkably accurate in this century—in

2000, 2004, 2008, and 2012. In 2000, the late polls revealed Gore’s lateadvancement that led to a virtual tie in the popular vote. (Gore, of course,won the popular vote, whereas Bush won the contested and decisiveelectoral vote.) In 2004, the polls correctly predicted another close electionwith most tilting toward Bush, the winner. In 2008, the late polls correctlypredicted Obama’s win and reported the vote margin with impressiveaccuracy. In 2012, most polls predicted Obama’s victory while slightlyunderestimating its size.

Table 2.2 details the record of the final polls of the 2012 race. Here,the vote is measured as the relative strengths of candidates Obama andRomney, allowing votes for undecideds (in the polls) and votes for minor-party candidates (in the polls and the election), so that the sum of Obamaand Romney support does not add quite up to 100 percent. The margin ismeasured as Obama’s lead over Romney. The national polls did vary but ina range that is more than acceptable in terms of sampling error.

Page 41: American Public Opinion: Its Origins, Content and Impact

TABLE 2.2 Accuracy of Final Polls in the 2012 PresidentialElection

Poll Obama Romney Obama MarginABC/Washington Post 50 47 +3 Politico/GWU/Battleground 47 47 +0 CBS/New York Times 48 47 +1 CNN 49 49 +0 Fox 46 46 +0 Gallup 48 49 −1 IBD/Tipp 50 49 +1 NBC/Wall Street Journal 48 47 +1 Pew 48 45 +3 Average (Poll of Polls) 48.7 47.7 +1.0Actual Vote 51.1 47.2 +3.9 Source: Based on data from 2012 General Election: Romney vs. Obama, Huffington Post,November 2012, http://elections.huffingtonpost.com/pollster/2012-general-election-romney-vs-obama. Selected polls are prominent likely voter polls based on live telephone interviewers.

As the table reveals, the polls underestimated the size of Obama’s lead

by about 3 percentage points, probably because of a modest last-minutetrend toward the president and also because some pollsters underestimatedturnout among racial minorities loyal to Obama. Meanwhile, the pollstersdid very well at predicting the results in the key battleground states. Of theten states deemed to be the battleground states of 2012, Obama defeatedRomney in all but North Carolina. And indeed the averages of the polls hadthe winner right in all ten states.36 Over recent election years, state-levelpolls have a strong record of accuracy, for predicting not only how the statewill vote for president but also the important offices of senator andgovernor.

Given the handicaps that pollsters face and the fact that they do notalways converge toward a common solution to their common problems, it isremarkable that the polls perform as well as they do. The pollsters’ best

Page 42: American Public Opinion: Its Origins, Content and Impact

defense is that, despite possible criticisms of specific procedures, their pollsdo predict with considerable accuracy. What keeps the pollsters on theirtoes is the knowledge that if they were systematically off the mark, theywould be out of business.

Election Day Exit PollsNow over a quarter century old, the exit poll is an important if sometimescontroversial innovation in opinion polling. Representative precincts areselected, and voters are interviewed immediately after leaving the votingbooth. Questions go beyond vote choice to include demographic andattitudinal information. Exit polls are commissioned by the news media tohelp them to project the winner soon after the polls close on Election Dayand report on the demographic and issue correlates of the vote choice.While the National Election Pool (NEP) provides data to televisionnetworks and other news outlets, each news sources hires its own experts toanalyze the data and ultimately call the election night winners (Best andKrueger 2012).

When networks and other news organizations call elections, exit pollresults actually play only a partial role compared to such techniques asgetting early returns from sample precincts and comparing reported electionresults with past voting patterns from the same geographic area. The majorpurpose of exit polls is to provide background data on which types of votersprefer which candidates. As raw data from exit polls stream in throughoutElection Day, they are unreliable as indicators of the vote margins in thestates and the nation.

There are several reasons why raw exit results are unreliable. There isa small but consistent partisan skew. The “raw data” tend to overstate theDemocratic vote and understate the Republican vote. The reason is thatDemocrats are more willing to cooperate with exit pollsters than areRepublicans (Best and Krueger 2012). Compounding the problem is that thelarge number of polling places that need to be monitored requires that manyinterviewers must be hired on a temporary basis. They are ofteninexperienced and poorly trained. With mostly young interviewers seekinginterviews outside polling places, older voters are more likely to refusecooperation, which (currently) contributes to the Democratic skew in theraw data.

Page 43: American Public Opinion: Its Origins, Content and Impact

Early voting provides a further hurdle. As of 2012, about 37 percent ofall voters cast their vote in presidential elections before the first Tuesday inNovember.37 These early voters tend to be more partisan, ideological, andpolitically interested than Election Day voters. Election night analysts mustestimate this early vote and factor that along with the exit poll data andother considerations when making their Election Day projections (Biemer etal. 2003; Frankovic 2003; Stein and Vonnahme 2008).38

Although by themselves, the data from exit polls comprise at best onlya rough indicator for predicting elections results, they are useful forpostelection analysis of the vote. After the election outcomes are known,exit poll data can be properly weighted to adjust for such factors as precinctvoting and certain demographic patterns to the refusals. Processed in thisway, long after the election, exit polls become among the best sources forinvestigating the variables that affect the vote in different elections. Foranalytical purposes, exit polls have two advantages to offset theirdisadvantages. The respondents include only voters, and their responsesreflect their thinking on Election Day.

Page 44: American Public Opinion: Its Origins, Content and Impact

2-5 ConclusionFor more than sixty years, public opinion polls have been part of thepolitical landscape, and there is every sign that their influence is growing.While the scientific pollsters have an admirable track record, it must berecognized that polls are imperfect predictors. Even the best preelection pollis based on a probability model that on rare occasions can go wrong due toan unlucky dose of sampling error. Preelection polls can also go wrongwhen the behavior they are sampling changes underneath their feet. Theelectorate’s choice is often volatile in the days before an election, and thepollsters race to catch up by means of polling until the last minute. Polls cango wrong because people are increasingly likely to refuse requests to beinterviewed. And polls can go wrong when they miscalculate which eligiblevoters will actually show up to vote. Under the circumstances,contemporary polls do a surprisingly good job at predicting elections. Onereason may be that pollsters know they must do a good job because theylose credibility and their jobs if they do not.

In this book, our central interest is in polls regarding public opinion,not election outcomes. A challenge is that unlike with election polls, we donot know for sure how close opinion polls are to the correct result. Unlessthe target opinion is the outcome of an initiative or referendum election,opinion polls lack direct validation. And opinion polls are sensitive tomatters of question framing and question wording. The key tounderstanding polls is a realization that the numbers do not speak forthemselves; they require interpretation. Among the keys to thatinterpretation are how the sample was selected, how the questions werephrased, an appreciation for the context in which the survey was conducted,and a comparison to other surveys taken at the same time or an analysis oftrends over time.

Page 45: American Public Opinion: Its Origins, Content and Impact

1.

2.

3.

Critical Thinking Questions

Since a poll’s exact question wording can affect the answers peoplegive about political issues, what cautions should one employ wheninterpreting the latest public opinion polls?Some people express doubts about the accuracy of polls because theyoften only interview a few hundred people. How would you defendscientific polling against this criticism?While pollsters have made disastrous mispredictions about theoutcome of some past elections, the polls have been quite accuratelately. But what kinds of dangers might lurk that could possibly causemajor polling errors again?

Page 46: American Public Opinion: Its Origins, Content and Impact

1.

2.

3.

4. 5.

6.

7. 8.

9.

10.

EndnotesAccording to David Moore (2013), in the last twenty-one days beforevoting in 2012, ninety survey organizations conducted 550 polls on theupcoming presidential election.Following the Senate’s failure to convict Clinton on impeachmentcharges, he is reported to have said, “Thank God for public opinion”(Edwards 2003, 4).Pew Research Center and Opinion Research Corporation, October 24–27, 2008.Gallup Poll, September 8–11, 2011.The best source for informed commentary on contemporary pollpractices is www.pollster.com.https://www.census.gov/hhes/www/housing/census/historic/phone.html.Internet World Stats, www.internetworldstats.com/stats.htm.Strictly speaking, in terms of statistical theory, if the poll wererepeated an infinite number of times, in 95 percent of the trials theresult would be between 54 and 60 percentage points. We are assumingsimple random sampling, which is a useful simplification of Gallup’ssampling procedure.The exception is when the population itself is quite small (e.g., lessthan 10,000). With a small population, a correction can be made toproperly lower the sampling error. Table 2.1 shows that a sample of384 drawn from a very large population has a sampling error of 5percent. However, if our barrel contained only 1,000 marbles and werandomly drew 384, then, instead of a sampling error of 5 percent, thetrue sampling error would be 0.785 × 5.0 percent, or 3.9 percent. Thesampling error correction for small populations is the square root of N− n/N − 1, where N is the population size and n is the sample size.Some pollsters extend poststratification to weight by partyidentification. The idea is that the public is reasonably stable inpartisanship, so that when one interviews, say, more Republicans thannormal, then Republicans should be weighted down to adjust for thedistorted sample. This procedure, designed to add precision to thesample estimates, is controversial. The danger is that, unlike with

Page 47: American Public Opinion: Its Origins, Content and Impact

11.

12.

13.

14.

15.

16.17.

18.

19.

20.

demographics, partisanship can change in the short run. An unexpectedsurge of Republican respondents might indicate sampling error, but itcould also mean more people decide to call themselves Republicans.One’s suspicions are aroused, however, if polls on the same topic atabout the same time with similarly worded questions show quitedifferent results.In 1948, Gallup had Dewey at 49.5 percent and Truman at 44.5 percent(with 6 percent going to minor-party candidates). The vote, in fact, wasTruman 49.5 percent and Dewey 45.1 percent, with others getting 5.4percent. The Roper Poll in 1948 was further off the mark than Gallup.Roper had Dewey at 52.2 percent and Truman at 37.1 percent (with10.7 percent going to minor-party candidates).This description is somewhat of an oversimplification of modern RDDmethods. For a detailed explanation, see Lepkowski et al. (2008).Thus, each sample has (1) cell phone–only households, (2) householdswith cell phones and landlines, and (3) households with landlines only.Complex statistical weights must be used to ensure that this mix isrepresentative of the adult population (Daniel 2012).Recently, Peytchev and Neely (2013) have suggested that surveyaccuracy could be optimized by going to cell phone–only samplingframe, and not calling landlines at all.fivethirtyeight.blogs.nytimes.com/2012/11/10.Actual formulas, however, are a good deal more complex. Not everyunanswered call is included in the denominator, depending on thesurvey. A detailed discussion of methods for computing response ratescan be found at the Web site for the American Association for PublicOpinion Research, www.aapor.org.www.people-press.org/2012/05/15/assessing-the-representativeness-of-public-opinion-surveys.While the overall response rate for the Survey of Consumer Attitudesbetween 1977 and 1999 was 70 percent, it dropped to 66 percent in theperiod between 1995 and 1999.The one significant difference was a tendency for survey respondentsto be more engaged in civic activity than nonsurvey respondents. Thereason is likely that people who are civic volunteers are alsopredisposed to be survey volunteers. Thus, the standard telephone

Page 48: American Public Opinion: Its Origins, Content and Impact

21.

22.

23.

24.

25.

26.

27.

28.

29.30.

survey may be inappropriate for studies of volunteer activities andmore generally for studies of social capital.Based on likely voters in the last twenty-one days of the campaign.fivethirtyeight.blogs.nytimes.com/2012/11/10.Joshua Clinton and Steven Rogers (2013) present evidence that roboticpolls are most accurate when the election they are polling is also beingpolled by conventional pollsters. The inference is that robotic pollresults may sometimes be adjusted to be consistent with the pack.Gary Langer, “Study Finds Trouble for Opt-in Internet Surveys,” TheNumbers, September 1, 2009,http://abcnews.go.com/blogs/politics/2009/09/study-finds-trouble-for-internet-surveys/.One can in fact do Internet surveys essentially for free. Seewww.surveymonkey.com. However, the data one gets for free at thissite may be worth exactly what was paid for it.For the full report, see Public Opinion Quarterly 74 (Winter 2010):711–81. For a rebuttal of this position, see Doug Rivers, “SecondThoughts about Internet Surveys,” www.huffingtonpost.com/guest-pollster/doug_rivers_b_724621.html.For an example of this methodology, seewww.knowledgenetworks.com.Among the best known users of this methodology are Harris Onlineand UGOV/Polimetrix. The Web site for the latter can be found athttp://research.yougov.com. The methodology described in this sectionfor the opt-in, sample matching methodology is the one used byUGOV/Polimetrix.This is a simple example, as other information from the targetpopulation may be known as well, such as voter registration, being aborn-again Christian, and party identification. This information can beobtained from the American Community Survey conducted by the U.S.Census.fivethirtyeight.blogs.nytimes.com/2012/11/10.Many of these articles were based on data collected byUGOV/Polimetrix for the Comparative Congressional ElectionsProject. UGOV/Polimetrix uses opt-in, sample matching Internetsurveys. The statistics employed in these studies assume probability

Page 49: American Public Opinion: Its Origins, Content and Impact

31.

32.

33.

34.

35.

36.

37.

38.

sampling, even though UGOV/Polimetrix samples are not probabilitysamples.Newsweek Poll conducted by Princeton Survey Research Associates,March 28–29, 2007.A 1986 Time/Yankelovich Survey, cited in “Opinion Roundup,” PublicOpinion (September/October 1986): 32.For a list of the way the major polling firms ask this question, seeCrespi (1988, Ch. 5).The Pew Research Center for the People and the Public. NewsRelease, August 2, 1996.Note that Figure 2.1 effectively assigns the undecideds as eitherstaying home or voting Democratic or Republican in proportion to thechoices among those who make a choice in the final poll. Allocatingthe undecideds as voting, say, 50–50, would lead to slightly lowerpredictions in landslide elections.Mark Blumenthal, “2012 Poll Accuracy: After Obama, Models andSurvey Science Won the Day,”http://www.huffingtonpost.com/2012/11/07/2012-poll-accuracy-obama-models-survey_n_2087117.html.http://www.people-press.org/2012/11/15/section-3-the-voting-process-and-the-accuracy-of-the-vote/.While the analysts at television networks and other news organizationsuse exit polls as one tool for predicting state-level races and winnersfor president and important statewide offices, they have learned topredict cautiously. The most notorious misprediction was the early callthat Gore had won Florida in 2000 (and by implication an ElectoralCollege majority). Faulty exit polls played a part in this misprediction.Since the 2000 fiasco, network analysts have made no serious errors inprematurely calling state winners. It may be said that network analystsknow more about forecasting elections than they used to, includingknowing more about the limits of what they know.

Page 50: American Public Opinion: Its Origins, Content and Impact

1. 2.

3.

4.

5.

6.

7.

EndnotesWe adopt West’s (2001) timeline with some modifications.There is still debate over what actually sunk the Maine. The mostplausible explanation is that the ship hit a mine, causing its munitionsto explode. Others believe the cause was internal; perhaps itsmunitions were ignited by a spark from its boilers.As the size of the audience for traditional news sources declines, theremaining audience grows older. In Pew’s 2012 survey, only 6 percentof young respondents aged eighteen to twenty-four reported havingread a print newspaper the day before and only 29 percent reportedwatching news on television. The comparable numbers for seniorsover age 65 were 48 percent (newspapers) and 72 percent (television).The youngest group did, however, lead the seniors in getting news offthe Internet, 52 to 28 percent.In 2012, among those regularly watch CNN 50 percent identified asDemocrats,16 percent as Republicans. Among those who regularlywatch Fox 22 percent are Democrats and 40 percent are Republicans.Among those who regularly watch MSNBC 58 percent are Democratsand 16 percent are Republicans (Pew Research Center).According to Arceneaux and Johnson’s (2013, 44) analysis of Pewdata from 2010, regular Fox News viewers are more than six timesmore likely to call themselves Republicans than Democrats. RegularMSNBC viewers tilt Democratic by more than 3 to 1. Thesedifferences are also reflected in ideological identification and stands onspecific issues.In the new fragmented media, one strategy candidates increasingly useto reach those with low political interest is appearing on entertainmenttalk shows or even Jon Stewart’s fake news, The Daily Show. Onebonus for these appearances is that the tone of the coverage theyreceive is overwhelmingly positive.While households not wired for cable comprise a captive audience fortelevision networks, the amount of news presented by broadcastnetworks has declined appreciably. Much political news of the typeonce reported by broadcast networks (e.g., coverage of partyconventions) is now relegated to cable news networks.

Page 51: American Public Opinion: Its Origins, Content and Impact

8. 9.

10.

11.

12.

13.

14.

15.

16.

17.

http://www.goodreads.com/quotes/tag/flat-earth.See the American Presidency Project’s “2012 General ElectionEditorial Endorsements by Major Newspapers,”http://www.presidency.ucsb.edu/data/2012_newspaper_endorsements.php.We averaged the positive tone for the Democratic and Republicancandidates for each year and then averaged those over the fourelections shown in Table 8.3.Talkers Magazine, October 2013, http://www.talkers.com/top-talk-radio-audiences/.Still another contributing factor to the news media’s increasednegativity is that they take the lead from the increased negativity ofcandidate advertising. As candidates increasingly run negativecampaigns attacking their opponents, these attacks become part of thenews story and hence reported by the media.Pew has tracked the degree to which prime time news shows cover thehorse race. In 2008, 53 percent of all prime time news stories about thepresidential campaign focused on horse-race coverage. The proportionfell to 38 percent in 2012 according to Pew but was not matched bycompensatory increases in coverage of candidate policy positions.One downside of the media’s reliance on polls is inaccurate reporting,as reporters often misinterpret poll results. Stephanie Larson (2001)gives a number of examples of how reporters come to wrongconclusions based on polling data.A meta-analysis is a study of the findings of all work in a specific area.The results of each study are quantified and entered into a database.Standard statistical tests are used to determine if the studies taken as awhole support a particular conclusion.With controls for education (dummies not beyond high school and forcollege graduates) and income (richest and poorest tertile), theinformation gain from living in a battleground state is significant at the.005 level. The estimated knowledge gap increases with the number ofcontrols.In the 2012 presidential race, the two campaigns followed differentstrategies for timing their ads. The Obama campaign struck early andoften, persistently attacking Romney for the allegedly predatorpractices of the company he once headed, Bain Capital. The idea was

Page 52: American Public Opinion: Its Origins, Content and Impact

that this campaign tactic would have a cumulative effect to erodeRomney’s standing with the voters. Romney saved his fire for a blitzof ads late in the campaign in the hope that they would create a large ifshort-term effect that would last until Election Day. See Sides andVavreck’s (2013) interesting analysis of the ad wars and their effects in2012.