# Response rate: Wikis

Note: Many of our articles have direct quotes from sources you can cite, within the Wikipedia article! This article doesn't yet, but we're working on it! See more info or our list of citable articles.

# Encyclopedia

Response rate (also known as completion rate or return rate) in survey research refers to the ratio of number of people who answered the survey divided by the number of people in the sample. It is usually expressed in the form of a percentage.

Example: if 1,000 surveys were sent by mail, and 257 were successfully completed and returned, then the response rate would be 25.7 %.

In direct marketing, the response rate refers to the number of people who responded to the offer.

In oncology, response rate (RR) is a figure representing the percentage of patients whose cancer shrinks (termed a partial response, PR) or disappears after treatment (termed a complete response, CR) . In simpler terms RR=PR+CR.

## Importance

A survey’s response rate is the result of dividing the number of people who were interviewed by the total number of people in the sample who were eligible to participate and should have been interviewed.[1]

A low response rate can give rise to sampling bias if the nonresponse is unequal among the participants regarding exposure and/or outcome.

For many years, a survey’s response rate was viewed as an important indicator of survey quality. Many observers presumed that higher response rates assure more accurate survey results (Aday 1996; Babbie 1990; Backstrom and Hursh 1963; Rea and Parker 1997). But because measuring the relation between non-response and the accuracy of a survey statistic is complex and expensive, few rigorously designed studies provided empirical evidence to document the consequences of lower response rates, until recently.

Such studies have finally been conducted in recent years, and they are challenging the presumption that a lower response rate means lower survey accuracy.

One early example of a finding was reported by Visser, Krosnick, Marquette and Curtin (1996) who showed that surveys with lower response rates (near 20%) yielded more accurate measurements than did surveys with higher response rates (near 60 or 70%).[2] In another study, Keeter et al. (2006) compared results of a 5-day survey employing the Pew Research Center’s usual methodology (with a 25% response rate) with results from a more rigorous survey conducted over a much longer field period and achieving a higher response rate of 50%. In 77 out of 84 comparisons, the two surveys yielded results that were statistically indistinguishable. Among the items that manifested significant differences across the two surveys, the differences in proportions of people giving a particular answer ranged from 4 percentage points to 8 percentage points.[3]

A study by Curtin et al. (2000) tested the effect of lower response rates on estimates of the Index of Consumer Sentiment (ICS). They assessed the impact of excluding respondents who initially refused to cooperate (which reduces the response rate 5-10 percentage points), respondents who required more than five calls to complete the interview (reducing the response rate about 25 percentage points), and those who required more than two calls (a reduction of about 50 percentage points). They found no effect of excluding these respondent groups on estimates of the ICS using monthly samples of hundreds of respondents. For yearly estimates, based on thousands of respondents, the exclusion of people who required more calls (though not of initial refusers) had a very small one.[4]

Holbrook et al. (2005) assessed whether lower response rates are associated with less unweighted demographic respresentativeness of a sample. By examining the results of 81 national surveys with response rates varying from 5 percent to 54 percent, they found that surveys with much lower response rates were only minimally less accurate.[5]

As a result of these and other such recent findings, it now seems clear that a low response rate does not guarantee lower survey accuracy and instead simply indicates a risk of lower accuracy. Consumers of survey results are therefore cautioned to view response rates as informative but to recognize that these rates “do not necessarily differentiate reliably between accurate and inaccurate data.”[6]

## References

1. ^ “Response Rates – An Overview.” American Association for Public Opinion Research (AAPOR). 29 Sept 2008. http://www.aapor.org/responseratesanoverview
2. ^ Visser, Penny S., Jon A. Krosnick, Jesse Marquette, and Michael Curtin. 1996. “Mail Surveys for Election Forecasting? An Evaluation of the Colombia Dispatch Poll.” Public Opinion Quarterly 60: 181-227.
3. ^ Keeter, Scott, Courtney Kennedy, Michael Dimock, Jonathan Best and Peyton Craighill. 2006. “Gauging the Impact of Growing Nonresponse on Estimates from a National RDD Telephone Survey.” Public Opinion Quarterly. 70(5): 759-779.
4. ^ Curtin, Richard, Stanley Presser and Eleanor Singer. 2000. "The Effects of Response Rate Changes on the Index of Consumer Sentiment." Public Opinion Quarterly 64(4): 413-428.
5. ^ Holbrook, Allyson, Jon Krosnick, and Alison Pfent. 2007. “The Causes and Consequences of Response Rates in Surveys by the News Media and Government Contractor Survey Research Firms.” In Advances in telephone survey methodology, ed. James M. Lepkowski, N. Clyde Tucker, J. Michael Brick, Edith D. De Leeuw, Lilli Japec, Paul J. Lavrakas, Michael W. Link, and Roberta L. Sangster. New York: Wiley.
6. ^ “Response Rates – An Overview.” American Association for Public Opinion Research (AAPOR). 29 Sept 2008. http://www.aapor.org/responseratesanoverview

# Study guide

Up to date as of January 14, 2010