20 Questions A Journalist Should Ask About Poll Results

23 nov

Second Edition
Sheldon R. Gawiser, Ph.D. and G. Evans Witt

For journalists and for pollsters, questions are the most frequently used tools for gathering information. For the journalist looking at a set of poll numbers, here are the 20 questions to ask the pollster before reporting any results. This publication is designed to help working journalists do a thorough, professional job covering polls. It is not a primer on how to conduct a public opinion survey.

The only polls that should be reported are “scientific” polls. A number of the questions here will help you decide whether or not a poll is a “scientific” one worthy of coverage – or an unscientific survey without value.

Unscientific pseudo-polls are widespread and sometimes entertaining, if always quite meaningless. Examples include 900-number call-in polls, man-on-the-street surveys, most Internet polls, shopping mall polls, and even the classic toilet tissue poll featuring pictures of the candidates on each sheet.

The major distinguishing difference between scientific and unscientific polls is who picks the respondents for the survey. In a scientific poll, the pollster identifies and seeks out the people to be interviewed. In an unscientific poll, the respondents usually “volunteer” their opinions, selecting themselves for the poll.

The results of the well-conducted scientific poll can provide a reliable guide to the opinions of many people in addition to those interviewed – even the opinions of all Americans. The results of an unscientific poll tell you nothing beyond simply what those respondents say.

With these 20 questions in hand, the journalist can seek the facts to decide how to handle every poll that comes across the news desk each day.

The authors wish to thank the officers, trustees and members of the National Council on Public Polls for their editing assistance and their support.

1. Who did the poll?
What polling firm, research house, political campaign, corporation or other group conducted the poll? This is always the first question to ask.

If you don’t know who did the poll, you can’t get the answers to all the other questions listed here. If the person providing poll results can’t or won’t tell you who did it, serious questions must be raised about the reliability and truthfulness of the results being presented.

Reputable polling firms will provide you with the information you need to evaluate the survey. Because reputation is important to a quality firm, a professionally conducted poll will avoid many errors.


2. Who paid for the poll and why was it done?
You must know who paid for the survey, because that tells you – and your audience – who thought these topics are important enough to spend money finding out what people think. This is central to the whole issue of why the poll was done.

Polls are not conducted for the good of the world. They are conducted for a reason – either to gain helpful information or to advance a particular cause.

It may be the news organization wants to develop a good story. It may be the politician wants to be re-elected. It may be that the corporation is trying to push sales of its new product. Or a special-interest group may be trying to prove that its views are the views of the entire country.

All are legitimate reasons for doing a poll.

The important issue for you as a journalist is whether the motive for doing the poll creates such serious doubts about the validity of the results that the numbers should not be publicized.

Examples of suspect polls are private polls conducted for a political campaign. These polls are conducted solely to help the candidate win –and for no other reason. The poll may have very slanted questions or a strange sampling methodology, all with a tactical campaign purpose. A campaign may be testing out new slogans, a new statement on a key issue or a new attack on an opponent. But since the goal of the candidate’s poll may not be a straightforward, unbiased reading of the public’s sentiments, the results should be reported with great care.

Likewise, reporting on a survey by a special-interest group is tricky. For example, an environmental group trumpets a poll saying the American people support strong measures to protect the environment. That may be true, but the poll was conducted for a group with definite views. That may have swayed the question wording, the timing of the poll, the group interviewed and the order of the questions. You should examine the poll to be certain that it accurately reflects public opinion and does not simply push a single viewpoint.

3. How many people were interviewed for the survey?
Because polls give approximate answers, the more people interviewed in a scientific poll, the smaller the error due to the size of the sample, all other things being equal. A common trap to avoid is that “more is automatically better.” It is absolutely true that the more people interviewed in a scientific survey, the smaller the sampling error – all other things being equal. But other factors may be more important in judging the quality of a survey.

4. How were those people chosen?
The key reason that some polls reflect public opinion accurately and other polls are unscientific junk is how the people were chosen to be interviewed.

In scientific polls, the pollster uses a specific method for picking respondents. In unscientific polls, the person picks himself to participate.

The method pollsters use to pick interviewees relies on the bedrock of mathematical reality: when the chance of selecting each person in the target population is known, then and only then do the results of the sample survey reflect the entire population. This is called a random sample or a probability sample. This is the reason that interviews with 1,000 American adults can accurately reflect the opinions of more than 200 million American adults.

Most scientific samples use special techniques to be economically feasible. For example, some sampling methods for telephone interviewing do not just pick randomly generated telephone numbers. Only telephone exchanges that are known to contain working residential numbers are selected – to reduce the number of wasted calls. This still produces a random sample. Samples of only listed telephone numbers do not produce a random sample of all working telephone numbers.

But even a random sample cannot be purely random in practice since some people don’t have phones, refuse to answer, or aren’t home.
5. What area (nation, state, or region) or what group (teachers, lawyers, Democratic voters, etc.) were these people chosen from?
It is absolutely critical to know from which group the interviewees were chosen.

You must know if a sample was draw from among all adults in the United States, or just from those in one state or in one city, or from another group. For example, a survey of business people can reflect the opinions of business people – but not of all adults. Only if the interviewees were chosen from among all American adults can the poll reflect the opinions of all American adults.

In the case of telephone samples, the population represented is that of people living in households with telephones. For most purposes, telephone households may be similar to the general population. But if you were reporting a poll on what it was like to be poor or homeless, a telephone sample would not be appropriate. Remember, the use of a scientific sampling technique does not mean that the correct population was interviewed.

Political polls are especially sensitive to this issue.

In pre-primary and pre-election polls, which people are chosen as the base for poll results is critical. A poll of all adults, for example, is not very useful on a primary race where only 25 percent of the registered voters actually turn out. So look for polls based on registered voters, “likely voters,” previous primary voters, and such. These distinctions are important and should be included in the story, for one of the most difficult challenges in polling is trying to figure out who actually is going to vote.

6. Are the results based on the answers of all the people interviewed?
One of the easiest ways to misrepresent the results of a poll is to report the answers of only a subgroup. For example, there is usually a substantial difference between the opinions of Democrats and Republicans on campaign-related matters. Reporting the opinions of only Democrats in a poll purported to be of all adults would substantially misrepresent the results.

Poll results based on Democrats must be identified as such and should be reported as representing only Democratic opinions.

Of course, reporting on just one subgroup can be exactly the right course. In polling on a primary contest, it is the opinions of those who can vote in the primary that count – not those who cannot vote in that contest. Each state has its own rules about who can participate in its primaries. Primary polls should include only eligible primary voters.

7. Who should have been interviewed and was not?
No survey ever reaches everyone who should have been interviewed. You ought to know what steps were undertaken to minimize non-response, such as the number of attempts to reach the appropriate respondent and over how many days.

There are many reasons why people who should have been interviewed were not. They may have refused attempts to interview them. Or interviews may not have been attempted if people were not home when the interviewer called. Or there may have been a language problem or a hearing problem.

8. When was the poll done?
Events have a dramatic impact on poll results. Your interpretation of a poll should depend on when it was conducted relative to key events. Even the freshest poll results can be overtaken by events. The President may have given a stirring speech to the nation, the stock market may have crashed or an oil tanker may have sunk, spilling millions of gallons of crude on beautiful beaches.

Poll results that are several weeks or months old may be perfectly valid, but events may have erased any newsworthy relationship to current public opinion.

9. How were the interviews conducted?
There are three main possibilities: in person, by telephone or by mail. Most surveys are now conducted by telephone, with the calls made by interviewers from a central location. However, some surveys are still conducted by sending interviewers into people’s homes to conduct the interviews.

Some surveys are conducted by mail. In scientific polls, the pollster picks the people to receive the mail questionnaires. The respondent fills out the questionnaire and returns it.

Mail surveys can be excellent sources of information, but it takes weeks to do a mail survey, meaning that the results cannot be as timely as a telephone survey. And mail surveys can be subject to other kinds of errors, particularly low response rates. In many mail surveys, more people fail to participate than do. This makes the results suspect.

Surveys done in shopping malls, in stores or on the sidewalk may have their uses for their sponsors, but publishing the results in the media is not among them. These approaches may yield interesting human-interest stories, but they should never be treated as if they represent a public opinion poll.

Advances in computer technology have allowed the development of computerized interviewing systems that dial the phone, play taped questions to a respondent and then record answers the person gives by punching numbers on the telephone keypad. Such surveys have a variety of severe problems, including uncontrolled selection of respondents and poor response rates, and should be avoided.

10. What about polls on the Internet or World Wide Web?
The explosive growth of the Internet and the World Wide Web has given rise to an equally explosive growth in various types of online polls and surveys. Many online polls may be good entertainment, but they tell you nothing about public opinion.

Most Internet polls are simply the latest variation on the pseudo-polls that have existed for many years. Whether the effort is a click-on Web survey, a dial-in poll or a mail-in survey, the results should be ignored and not reported. All these pseudo-polls suffer from the same problem: the respondents are self-selected. The individuals choose themselves to take part in the poll – there is no pollster choosing the respondents to be interviewed.

Remember, the purpose of a poll is to draw conclusions about the population, not about the sample. In these pseudo-polls, there is no way to project the results to any larger group. Any similarity between the results of a pseudo-poll and a scientific survey is pure chance.

Clicking on your candidate’s button in the “voting booth” on a Web site may drive up the numbers for your candidate in a presidential horse-race poll online. For most such efforts, no effort is made to pick the respondents, to limit users from voting multiple times or to reach out for people who might not normally visit the Web site.

The 900-number dial-in polls may be fine for deciding whether or not Larry the Lobster should be cooked on Saturday Night Live or even for dedicated fans to express their opinions on who is the greatest quarterback in the National Football League. The opinions expressed may be real, but in sum the numbers are just entertainment. There is no way to tell who actually called in, how old they are, or how many times each person called.

Never be fooled by the number of responses. In some cases a few people call in thousands of times. Even if 500,000 calls are tallied, no one has any real knowledge of what the results mean. If big numbers impress you, remember that the Literary Digest’s non-scientific sample of 12,000,000 people said Landon would beat Roosevelt in the 1936 Presidential election.

Mail-in coupon polls are just as bad. In this case, the magazine or newspaper includes a coupon to be returned with the answers to the questions. Again, there is no way to know who responded and how many times each person did.

Another variation on the pseudo-poll comes as part of a fund-raising effort. An organization sends out a letter with a survey form attached to a large list of people, asking for opinions and for the respondent to send money to support the organization or pay for tabulating the survey. The questions are often loaded and the results of such an effort are always meaningless.

This technique is used by a wide variety of organizations from political parties and special-interest groups to charitable organizations. Again, if the poll in question is part of a fund-raising pitch, pitch it – in the wastebasket.

With regard to the Internet, methods are being developed to sample the opinions of those who have online access, although these efforts are just starting. Even a survey that accurately sampled those who have access to the Internet would still fall short of a poll of all Americans, since only a relatively small fraction of the nation’s adults have access to the Internet.

11. What is the sampling error for the poll results?
Interviews with a scientific sample of 1,000 adults can accurately reflect the opinions of nearly 200 million American adults. That means interviews attempted with all 200 million adults – if such were possible – would give approximately the same results as a well-conducted survey based on 1,000 interviews.

What happens if another carefully done poll of 1,000 adults gives slightly different results from the first survey? Neither of the polls is “wrong.” This range of possible results is called the error due to sampling, often called the margin of error.

This is not an “error” in the sense of making a mistake. Rather, it is a measure of the possible range of approximation in the results because a sample was used.

Pollsters express the degree of the certainty of results based on a sample as a “confidence level.” This means a sample is likely to be within so many points of the results one would have gotten if an interview were attempted with the entire target population. They usually say this with 95% confidence.

Thus, for example, a “3 percentage point margin of error” in a national poll means that if the attempt were made to interview every adult in the nation with the same questions in the same way at about the same time as the poll was taken, the poll’s answers would fall within plus or minus 3 percentage points of the complete count’s results 95% of the time.

This does not address the issue of whether people cooperate with the survey, or if the questions are understood, or if any other methodological issue exists. The sampling error is only the portion of the potential error in a survey introduced by using a sample rather than interviewing the entire population. Sampling error tells us nothing about the refusals or those consistently unavailable for interview; it also tells us nothing about the biasing effects of a particular question wording or the bias a particular interviewer may inject into the interview situation.

Remember that the sampling error margin applies to each figure in the results – it is at least 3 percentage points plus or minus for each one in our example. Thus, in a poll question matching two candidates for President, both figures are subject to sampling error.

12. Who’s on first?
Sampling error raises one of the thorniest problems in the presentation of poll results: For a horse-race poll, when is one candidate really ahead of the other?

Certainly, if the gap between the two candidates is less than the error margin, you should not say that one candidate is ahead of the other. You can say the race is “close”, the race is “roughly even”, or there is “little difference between the candidates.” But it should not be called a “dead heat” unless the candidates are tied with the same percentages.

And just as certainly, when the gap between the two candidates is equal to or more than twice the error margin – 6 percentage points in our example – and if there are only two candidates and no undecided voters, you can say with confidence that the poll says Candidate A is clearly leading Candidate B.

When the gap between the two candidates is more than the error margin but less than twice the error margin, you should say that Candidate A “is ahead”, “has an advantage” or “holds an edge”. The story should mention that there is a small possibility that Candidate B is ahead of Candidate A.

When there are more than two choices or undecided voters – in the real world – the question gets much more complicated. While the solution is statistically complex, you can fairly easily evaluate this situation by estimating the error margin. You can do that by taking the percent for each of the two candidates in question and multiplying it by the total respondents for the survey (only the likely voters if that is appropriate). This number is now the effective sample size for your judgement. Look up the sampling error in a table of statistics for that reduced sample size, and apply it to the candidate percentages. If they overlap, then you do not know if one is ahead. If they do not, then you can make the judgement that one candidate has a lead.

And bear in mind that when subgroup results are reported – women or blacks, or young people – the sampling error margin for those figures is greater than for results based on the sample as a whole.

13. What other kinds of factors can skew poll results?
The margin of sampling error is just one possible source of inaccuracy in a poll. It is not necessarily the source of the greatest source of possible error; we use it because it’s the only one that can be quantified. And, other things being equal, it is useful for evaluating whether differences between poll results are meaningful in a statistical sense.

Question phrasing and question order are also likely sources of flaws. Inadequate interviewer training and supervision, data processing errors and other operational problems can also introduce errors. Professional polling operations are less subject to these problems than volunteer-conducted polls, which are usually less trustworthy.

You should always ask if the poll results have been “weighted.” This process is usually used to account for unequal probabilities of selection and to adjust slightly the demographics in the sample. You should be aware that a poll could be manipulated unduly by weighting the numbers to produce a desired result. While some weighting may be appropriate, other weighting is not. Weighting a scientific poll is only appropriate to reflect unequal probabilities or to adjust to independent values that are mostly constant.

14. What questions were asked?
You must find out the exact wording of the poll questions. Why? Because the very wording of questions can make major differences in the results.

Perhaps the best test of any poll question is your reaction to it. On the face of it, does the question seem fair and unbiased? Does it present a balanced set of choices? Would most people be able to answer the question?

On sensitive questions – such as abortion – the complete wording of the question should probably be included in your story. It may well be worthwhile to compare the results of several different polls from different organizations on sensitive questions. You should examine carefully both the results and the exact wording of the questions.

15. In what order were the questions asked?
Sometimes the very order of the questions can have an impact on the results. Often that impact is intentional; sometimes it is not. The impact of order can often be subtle.

During troubled economic times, for example, if people are asked what they think of the economy before they are asked their opinion of the president, the presidential popularity rating will probably be lower than if you had reversed the order of the questions. And in good economic times, the opposite is true.

What is important here is whether the questions that were asked prior to the critical question in the poll could sway the results. If the poll asks questions about abortion just before a question about an abortion ballot measure, the prior questions could sway the results.

16. What about “push polls”?
In recent years, some political campaigns and special-interest groups have used a technique called “push polls” to spread rumors and even outright lies about opponents. These efforts are not polls, but are political manipulation trying to hide behind the smokescreen of a public opinion survey.

In a “push poll,” a large number of people are called by telephone and asked to participate in a purported survey. The survey “questions” are really thinly-veiled accusations against an opponent or repetitions of rumors about a candidate’s personal or professional behavior. The focus here is on making certain the respondent hears and understands the accusation in the question, not in gathering the respondent’s opinions.

“Push polls” are unethical and have been condemned by professional polling organizations.

“Push polls” must be distinguished from some types of legitimate surveys done by political campaigns. At times, a campaign poll may ask a series of questions about contrasting issue positions of the candidates – or various things that could be said about a candidate, some of which are negative. These legitimate questions seek to gauge the public’s reaction to a candidate’s position or to a possible legitimate attack on a candidate’s record.

A legitimate poll can be distinguished from a “push poll” usually by:

The number of calls made – a push poll makes thousands and thousands of calls, instead of hundreds for most surveys; The identity of who is making the telephone calls – a polling firm for a scientific survey as opposed to a telemarketing house or the campaign itself for a “push poll”; The lack of any true gathering of results in a “push poll,” which has as its only objective the dissemination of false or misleading information.

17. What other polls have been done on this topic? Do they say the same thing? If they are different, why are they different?
Results of other polls – by a newspaper or television station, a public survey firm or even a candidate’s opponent – should be used to check and contrast poll results you have in hand.

If the polls differ, first check the timing of the interviewing. If the polls were done at different times, the differing results may demonstrate a swing in public opinion.

If the polls were done about the same time, ask each poll sponsor for an explanation of the differences. Conflicting polls often make good stories.

18. So I’ve asked all the questions. The answers sound good. The poll is correct, right?
Usually, yes. However, remember that the laws of chance alone say that the results of one poll out of 20 may be skewed away from the public’s real views just because of sampling error.

Also remember that no matter how good the poll, no matter how wide the margin, no matter how big the sample, a pre-election poll does not show that one candidate has the race “locked up.” Things change – often and dramatically in politics. That’s why candidates campaign.

19. With all these potential problems, should we ever report poll results?
Yes. Because reputable polling organizations consistently do good work. In spite of the difficulties, the public opinion survey, correctly conducted, is still the best objective measure of the state of the views of the public.

20. Is this poll worth reporting?
If the poll was conducted correctly, and you have been able to obtain the information outlined here, your news judgment and that of your editors should be applied to polls, as it is to every other element of a story.

Por NCPP – National Council on Publics Polls

Comentários desativados em 20 Questions A Journalist Should Ask About Poll Results

Publicado por em 23/11/2010 em Uncategorized


Os comentários estão desativados.

%d blogueiros gostam disto: