Descarga la aplicación para disfrutar aún más
Vista previa del material en texto
International Journal of Public Opinion Research Vol. 25 No. 2 2013 � The Author 2012. Published by Oxford University Press on behalf of The World Association for Public Opinion Research. All rights reserved. doi:10.1093/ijpor/eds040 Advance Access publication 12 December 2012 Survey-Research Paradigms Old and New Tom W. Smith NORC at the University of Chicago Abstract Information gathering underwent a dramatic transformation in the 1930s with the rise of survey research. Now, a growing number of critics are arguing that survey research’s time has passed and it should and will be replaced by a combination of data mining from the Internet and the use of various administrative records. The promise and reliability of the new techniques are assessed, and although they show promise, they are not at present scientifically superior to survey research for most uses. A paradigm shift occurred almost 80 years ago in the mid-1930s when Gallup, Roper, Crossley, and a handful of other innovators pioneered the public opinion poll (Brick, 2011; Converse, 1987; Groves, 2011b). Before the advent of polling, politicians, journalists, social scientists, and others had turned to various sources to measure public opinion and other aspects of society. These included tracking election returns, the outcomes of referenda, crowd counts, straw polls, compilations of editorials and news articles collected by such publications as Public Opinion (taken over by Literary Digest in 1906), studies of letters to the editor, and, as George Gallup (1957) noted in 1957, such other evidence as ‘‘letters to congressmen, the lobbying of pressure groups, and the reports of political henchmen . . . . ’’1 These alternatives were All correspondence concerning this article should be addressed to Tom W. Smith, NORC at the University of Chicago, 1155 East 60th Street., Chicago, IL 60637. 1 In general see Albig, 1939 & 1956; Gallup and Rae, 1940; Robinson, 1932. On straw polls (Converse, 1987; Kernell, 2000; Robinson, 1932), on counts of letters, telegrams, etc. (Dexter, 1956; Herbst, 1993; Kriesberg, 1945), on counting crowds (Herbst, 1993; Splichal, 1999), on informants, especially local party workers (Gallup and Rae, 1940; Herbst, 1993; Holli, 2002; Robinson, 1932), on newspapers (Albig, 1956; Herbst, 1993; Splichal, 1999) and on letters to the editor (Gray and Brown, 1970; Wahl-Jorgensen, 2001; 2002). at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from at Pontificia U niversidad C atà ³lica de C hile on O ctober 3, 2013 http://ijpor.oxfordjournals.org/ D ow nloaded from http://ijpor.oxfordjournals.org/ http://ijpor.oxfordjournals.org/ http://ijpor.oxfordjournals.org/ http://ijpor.oxfordjournals.org/ http://ijpor.oxfordjournals.org/ http://ijpor.oxfordjournals.org/ http://ijpor.oxfordjournals.org/ http://ijpor.oxfordjournals.org/ http://ijpor.oxfordjournals.org/ http://ijpor.oxfordjournals.org/ http://ijpor.oxfordjournals.org/ http://ijpor.oxfordjournals.org/ supplanted by the polls, and soon, public opinion and poll results became considered to be almost synonymous with each other. The advent of polling was a complete game changer. As Elmo Wilson (1947), a researcher at Roper and other organizations, remarked in 1947, ‘‘25 years ago the possibility of measuring public opinion with any degree of precision was at least as remote from public consciousness as the atomic bomb.’’ Now, a rising chorus is asserting that polls are passé, a growingly anti- quated relic of the past century. They claim that public opinion, consumer behaviors, and other sociopolitical outcomes can be better measured (less ex- pensively, more quickly, more easily) by the analysis of Internet usage in general and of social media in particular, by the data mining of administrative databases (including the merging of disparate information sources through such techniques as data fusion), or by a combination of these two alternatives to traditional surveys. The low cost of mined data is often touted in contrast to the high and rising cost of traditional survey research (Brick, 2011; Groves, 2011b; O’Connor, B., Balasubramanyan, R., Routledge, B. R., & Smith, N. A., 2010; Schaeffer and Dykema, 2011). Another attractive aspect of social media and other Internet analysis is its speed (O’Connor, B., Balasubramanyan, R., Routledge, B. R., & Smith, N. A., 2010). For example, on January 20, 2012, Politico (Van Dongen, 2012) published daily counts covering January 12-18 of ‘‘Facebook chatter’’ about the Republican presidential candidates based on more than a million ‘‘status updates, postings, and comments.’’ Others emphasize the ease of amassing Internet data: that it can be compiled routinely, even automatically (Hillygus, 2011; O’Connor, B., Balasubramanyan, R., Routledge, B. R., & Smith, N. A., 2010; Savage and Burrows, 2007). As Myron Gutmann (2011), Assistant Director of the National Science Foundation, has pointed out, there is the promising ‘‘poten- tial of existing data being more easily accessed. These might be administrative data collected by federal and state agencies or commercial or opportunistic data sets, collected for a particular purpose or, like the social media sites, obtained in the course of business or captured from a device.’’ For example, Mike Savage and Roger Burrows (2007), British sociologists, argue that the ‘‘glory days’’ of survey research ‘‘are in the past.’’ They write, ‘‘It is unlikely, we suggest, that in the future the sample survey will be a particularly important research tool . . . . ’’ They believe surveys will be replaced by ‘‘digital data generated routinely as by-products of their own transactions . . . . ’’2 Similarly, Carnegie Mellon computer scientist Brendan O’Connor and others in ‘‘From Tweets to Polls: Linking Text Sentiments to Public Opinion Time Series’’ (2010) have argued that ‘‘expensive and 2 For follow-ups to Savage and Burrows (2007), see Compton (2007), Savage and Burrows (2009), and Webber (2009). S U R V E Y - R E S E A R C H P A R A D I G M S O L D A N D N E W 219 Ignacio Resaltado Ignacio Resaltado time-intensive polling can be supplemented or supplanted with simple-to- gather text data that is generated from online social networking.’’ Others are not so ready to consign surveys to the dustbin of history, but also see them as losing ground to newer techniques. D. Sunshine Hillygus (2011), a political scientist at Duke, states, ‘‘There has been a noticeable decline in the prominence of polls in election politics and scholarship. In forecasting elections, statistical models and prediction markets appear to beviable alternatives to polling predictions, especially early in the campaign. In understanding voting behavior, surveys are increasingly replaced by experi- mental designs or alternative measures of attitudes and behaviors.’’ A star exhibit for the ‘‘alternative measures’’ of electoral behavior is the research on the 2009 German national election by Andranik Tumasjan and others (2010) at the Technical University of Munich. They concluded that ‘‘the mere number of tweets mentioning a political party can be considered a plausible reflection of the vote share and its predictive power even comes close to traditional election polls.’’ They reported a mean absolute error (MAE) for vote share of 1.65% from their tweets’ analysis versus an average MAE of 1.14% for six polls. But their success has not been replicated in a number of studies of U.S. elections. Analysis using tweets, Facebook likes, and other social media and studies of Internet searches of races involving the U.S. Senate, governors, and other contests have typically found mixed to poor results. For example, Lui, Metaxas, and Mustafaraj (2011) found that Google Trends3 ‘‘was, in general, not a good predictor of both the 2008 and 2010 elections, as compared to incumbency, the NYT polls, and even chance.’’ Evaluating Twitter chatter volume and sentiment analysis of the U.S. election in 2010, Gayo-Avello, Metaxas, and Mustafaraj (2011) found that ‘‘the mean average error (MAE) was rather high: 17.1% for Twitter volume and 7.6% for the sentiment ana- lysis.4 By comparison, MAE for professional polling services is typically about 2-3%.’’ Similarly, Olson and Bunnett (2011) found that Facebook likes5 in 2010 explained 13% of the variation in Senate races, but were ‘‘very weak’’ predictors of the results of gubernatorial races, and for the U.S. House races, there was ‘‘a slight negative correlation.’’ Estimates of other political matters beside election outcomes have also been problematic. Green (2011) reported that Twitter followings were 3 Google Trends is a sampling of Google Web searches that computes how many searches have been done for selected terms relative to the total number of searches done on Google. 4 Volume counts tally how many tweets use certain selected terms such as candidate names. Sentiment analysis using standard algorithms automatically classifies the tweets as positive or negative toward the indexed target. 5 Facebook usage has been analyzed in various ways. Some focus on only specific pages (e.g. candidates in a particular election) versus overall Facebook activity. The component examined most frequently was posted ‘‘likes,’’ but other elements studied include links to external Web sites, within Facebook links, and other posts. I N T E R N A T I O N A L J O U R N A L O F P U B L I C O P I N I O N R E S E A R C H220 ‘‘a leading indicator of polling’’ early in the Republican presidential contest, but no statistical analysis was presented to back up this assertion. O’Connor, B., Balasubramanyan, R., Routledge, B. R., & Smith, N. A., (2010) found that sentiment ratios on Twitter in 2008 did not correlate with pre-election polls. As a roundtable discussion of political blogging during the 2008 U.S. presi- dential campaign indicated, ‘‘Studying these blogs is valuable, but not the equivalent of studying the body politic nor the political process as a whole’’ (Permutter, 2008). Success outside politics has been on balance better for the social media and Internet tallies. These include predicting Best Motion Picture Oscar winners (Bothos, Apostolou, & Mentzas, 2010), box-office sales for movies (Asur and Huberman, 2010), initial unemployment claims filed for (Wolfers, 2011), various retail sales figures (Choi and Varian (2009), and estimates of corrup- tion (Saiz and Simonsohn, 2007). However, although scoring some notable results, outcomes have been far from perfect. Thus far, the new alternatives to survey research in general and public opin- ion research in particular have not established their superiority either scientific- ally or empirically. There are many limitations of the new alternatives to surveys. First, there is the coverage issue (Berube, 2011). Most people do not regularly use any social media, and many have no access to, or make little use of, the Internet. Second, the coverage is not representative of the general population. The digital divide is still wide. As Couper (2011) has observed, ‘‘While Facebook users are of interest in their own right, few would suggest that those who use Facebook are representative of the entire population of any particular country (and it’s often difficult to attach a country to such users or even properly identify who such users are).’’ For example, in 2011, 45% of U.S. Facebook users were aged �25 years. Third, much social media activity is heavily concentrated in a few thumbs. For example, among tweets about the 2009 German election, ‘‘Only 4% of all users accounted for more than 40% of the messages’’ (Tumasjan, A., Sprenger T. O., Sandner P. G., & Welpe I. M., 2010). Not only is social media activity skewed, but it also can be intentionally distorted. Gayo-Avello, Metaxas, and Mustafaraj (2011) note that social media counts ‘‘allow manipulation by spam- mers and propagandists.’’ Fourth, many Internet data–mining techniques are relatively untested and clearly have not yet been optimized (Hong, J. I., Zhang, J., & Zimmerman, J., 2011). This includes the selection of search lexicons, polarity algorithms, and sentiment analysis; filtering techniques to isolate target populations of interest (e.g., by age, residence); and smoothing, rolling averages, and aggregation pro- cedures. For example, the automatic tabulating of tweets, blogs, and other text in sentiment analysis is crude and often has low reliability (Gayo-Avello, Metaxas, & Mustafaraj, 2011). Other keyword counts also have problems S U R V E Y - R E S E A R C H P A R A D I G M S O L D A N D N E W 221 Ignacio Resaltado Ignacio Resaltado Ignacio Resaltado Ignacio Resaltado due to incomplete lexicons, words with multiple meanings, and ambiguous terms (Saiz and Simonsohn, 2007). Greater use and more attention to valid- ation will of course improve and refine these techniques. Fifth, while much information about Internet and social media usage is available, much more is restricted by the operators to protect privacy, for business reasons, or otherwise. For example, Couper (2011) notes that it is not possible to use Facebook users as a sample frame. Sixth, documentation of Internet data is often sparse. It is often not clearly indicated how they are compiled and exactly what they count. For example, Google Trends uses a fluctuating sampling strategy that is both not explained and creates variability in measurements over time (Lui, Metaxas, and Mustafaraj, 2011). Other products are limited in what they provide. Google Trends does not report hard counts, but only the relative level of searches from a base point. It also excludes many low-incidence events. Seventh, the Internet is a rapidly expanding and ever-changing phenom- enon, and that changeability makes its use and especially its consistent use over time particularly difficult (Murphy, 2012). Companies, hardware, soft- ware, platforms, Web sites, and so forth are continually changing, and research designs using these must continually adapt. At the minimum, this creates technical challenges. At the maximum, it seriously questions whether valid reliable comparisons can be made over time. Eighth, much of the new Internet analysis is, by necessity, restricted to what routinely appears there and in accessible databases. Often, they do not include crucial measures of theoretical and empirical importance (Golder and Macy, 2012). In Groves’ (2011a, b) nomenclature, the Internet-based ‘‘organic data’’ are not ‘‘designed data’’ and almost always lacks the scope and depth that a well-designed survey created toexplore a specific research question can include. Designed research can more completely articulate a full model for testing hypotheses. Crompton (2008) argues that in contrast to what is possible with most Internet data, ‘‘the capacity of survey research to simultaneously explore the impact of a number of different variables on a particular phenom- enon and the relationships between different variables should not be underestimated.’’ Ninth, there is the flip side of the limited-access issue, the invasion of privacy danger. If the veil of inaccessibility is raised, won’t we let in scam- mers, identity thieves, and worse along with benign researchers? If the door to data is swung open, who knows who will come in and what will be taken out? While the push and pull between sharing and privacy continues to move back and forth, there is a reasonable probability that privacy concerns are growing and that access will become more restricted in the future (Couper, 2011; Hong, J. I., Zhang, J., & Zimmerman, J., 2011). As with using social media and other Internet data, using administrative records has both many advantages and serious impediments. Administrative I N T E R N A T I O N A L J O U R N A L O F P U B L I C O P I N I O N R E S E A R C H222 Ignacio Resaltado Ignacio Resaltado Ignacio Resaltado Ignacio Resaltado Ignacio Resaltado data can be useful both independently and when combined with survey data. Although administrative data are sometimes portrayed as official and definitive and as representing actual situations and behaviors rather than merely the recalled and reported information that is collected in surveys, their advantage over survey data is neither certain nor automatic. First, there are serious access issues. Much valuable administrative data are either inaccessible to researchers in general or restricted to a narrow band of users, such as only by government statistical agencies. Access to individual administrative records is (and should be) more limited than for anonymous Internet-use data. However, if access is severely limited and even scholars for legitimate research applications cannot acquire and use the data, it really does not matter whether such potentially valuable, but inaccessible, information exists. Second, administrative records are variable in quality. Although some records are carefully collected and checked, this is far from universal. All records contain some error, and some records are poorly organized, out of date, badly documented, and with little or no quality-control procedures. For example, Brick (2011) notes that administrative records may not have ‘‘the same level of quality as is needed for sampling purposes.’’ Third, administrative records are collected for specific programmatic or commercial purposes and not to address specific research questions. As a result, they often lack the needed information to address the research ques- tions under investigation. For example, voter registration records indicate who is registered to vote and who voted in particular elections. However, they contain little other information about the voters. For example, they do not include voters’ demographics, who they voted for, or any attitudinal informa- tion. They tell us who voted, but little about who the ‘‘whos’’ are and nothing about their motivations, reasons, or many other matters. Of course, administrative records do not have to stand on their own. Their value can be enhanced by combining together with other data sources. First, different administrative records can be merged together. This has been done in a number of European countries where various governmental databases have been linked together to create a much fuller profile of individuals (Smith, 2011b). However, often there are serious technical, legal, and ethical barriers to such merging (Stockwell, 2011). Also, while good from a data analysis perspective, this raises both serious privacy concerns and access issues. Second, survey data and administrative data can be linked together. Usually administrative data are appended to survey data (but the data exchanges could go in the opposite direction). Such data linkages have been done for decades and are fairly common and often valuable (Blumberg & Cynamon, 1999; Calderwood & Lessof, 2009; Hewat, 2011; Huynh, Rupp, & Sears, 2000; Moore & Marquis, 1987; Pedace & Bates, 2000; Rhodes & Fung, 2004). However, linking individuals across lists is often challenging and error prone. Even when a correct linkage is S U R V E Y - R E S E A R C H P A R A D I G M S O L D A N D N E W 223 Ignacio Resaltado Ignacio Resaltado Ignacio Resaltado Ignacio Resaltado Ignacio Resaltado Ignacio Resaltado Ignacio Resaltado made, results may be problematic because the timing of survey and administrative data are out of sync and because of errors in either source. Although the increased volume of computerized and individual-level data has facilitated this approach, that has not eliminated the problems of mis- and non-linkage and of underlying data errors, nor are synthetic merging proced- ures such as data fusion a simple answer (Van der Putten &Kok, 2010). While this emerging technique can combine different data sources and create usable new information, the product is more of a simulation than actual hard data. Overall, administrative records can be valuable research tools and have even greater potential in the future (Card, 2011), but as Brick (2011) notes, ‘‘Despite their promise, the use of administrative records has not lived up to its potential, at least not in US applications [and] (t)he future use of admin- istrative records, while still promising, faces many challenges.’’ But of course, perhaps things do not need to go as far as the complete data acquisition and analysis transformation discussed earlier in the text. Although well short of a paradigm shift, there are important mode and sample shifts to access panels and river-sampling studies on the Internet. These have already become ubiquitous in market research and are widely used in other areas as well (Brick, 2011). There is a continuum of research designs running from 1) traditional surveys using such samples and modes as random digit dialing (RDD) and computer-assisted telephone interviewing (CATI) or address-based sampling and face-to-face or mixed modes to 2) probability-based Internet survey such as Knowledge Networks or e-mail–list samples of more limited target popu- lations such as employees of a particular company or students at certain uni- versities to 3) non-probability Internet samples using access panels and/or river sampling to 4) the data mining of Internet activity (e.g., searches, sites, social media) and/or administrative records without any interviewing and often with- out any sampling (i.e., the entire population is analyzed). While useful for certain purposes, the less extreme, but more common, shift to non-probability, Internet surveys also has serious problems. The AAPOR Task Force on Online Panels (Baker et al., 2010) recommended that ‘‘Researchers should avoid non-probability online panels when one of the research objectives is to accurately estimate population values.’’ They concluded that ‘‘Non-probability samples are generally less accurate than probability samples.’’ Similarly, Smyth’s (2012) methodological review of Web surveys concluded that they were not yet ‘‘ready for prime time.’’ Also, Yeager et al. (2011) found Internet non-probability studies were inferior to both non-Internet probability samples and Internet-based probability samples. Similarly, Brick (2011) notes that ‘‘No generally accepted method of sampling from the Web has been established at this time.’’ However, despite the shortcomings and uncertainties about the new sources of information, they should not simply be ignored. Social media are an important and growing form of social interaction and should be studied in depth. However, thestudy of these social media should not be equated to the I N T E R N A T I O N A L J O U R N A L O F P U B L I C O P I N I O N R E S E A R C H224 analysis of general public opinion or of society in toto (Bhutta, 2012). They should be studied for what they are and not for what they are not. They provide information about public opinion and are a source of information about many other societal aspects, but are generally more limited than stand- ard surveys in capturing the full range and complexity of public opinion and measuring society’s values and behaviors. Similarly, analysis of Internet searches and Web sites and the traffic to same can also be valuable and can illuminate many research topics. However, they will not be the best source for answering all research questions. New technol- ogies, especially new communication technologies, are important to study in general, and in particular for the formation and measurement of public opinion and other social phenomenon. As Splichal (1999) has noted, ‘‘a new technology of forming and expressing opinions may commonly be associated with changes in the nature of public opinion.’’ He goes on to indicate that just as the rise of newspapers changed public opinion in the past, ‘‘computer mediated communication may again affect the formation of public opinion and its nature’’ in the future. This development is important to study, but it does not mean that using the ‘‘new technology’’ can simply replace existing and more reliable means of measuring public opinion and other matters. Administrative databases can also be useful for sampling some populations (especially of course in countries with population registers) and can be used to check and supplement survey-based data when record linkage can be reliably carried out. They can be especially valuable for contextualizing data (Smith, 2011a). However, the idea that using administrative databases alone could replace survey research is clearly misguided. The most promising future is not one that abandons survey research, but which augments traditional forms of sampling and interviewing with supple- mental data from various administrative sources and combines survey-based individual-level information with aggregate-level data from censuses, admin- istrative records, and other surveys (Smith & Kim, 2009). Multilevel analysis combining individual-level and aggregate-level data from various sources is especially promising (Smith, 2011a, b). Couper (2011) notes that ‘‘these rich data sources will add much to our understanding of public opinion, but will not replace surveys in the near future.’’ As Robert Groves (2011a), Director of the U.S. Census, has observed, ‘‘The combination of designed data [from surveys] with organic data [from the Internet and other automatic sources] is the ticket to the future.’’ Examples of hybrid, old-school, new-tech blends include recruiting samples from Facebook users and combining survey responses with data from their online pages (Baresch, B., Knight L., Harp D., Yaschur C., 2011), time-series analysis combining measures of Web activity with data from real-world shopping studies (Stacey, Pauwels, & Lackman, 2012), aug- menting individual-level survey-based political data with aggregate-level S U R V E Y - R E S E A R C H P A R A D I G M S O L D A N D N E W 225 information from administrative records on voter registration and electoral results (Smith, 2011b), and cross-validating analysis of online discourse on a topic, media-content analysis, and survey-based studies of the same issue. So we should not shift the paradigm and jettison survey research, but we should boost standard surveys of public opinion and other topics with auxil- iary data from alternative sources, supplement general population surveys with valuable, but more limited, studies analyzing social media and other Internet usages, and use the study designs and sources that are best suited for assessing the many different research questions that we seek to answer. References Albig, W. (1939). Public opinion. New York: McGraw-Hill. Albig, W. (1956). Modern public opinion. New York: McGraw-Hill. Asur, S., & Huberman, B. A. (2010). Predicting the future with social media. Unpublished report, Social Computing Lab, HP Lab. doi:10.1109/ WI-IAT.2010.63. Baker, R., Blumberg, S. J., Brick, J. M., Couper, M. P., Courtright, M., Dennis, J. M., Dillman, D., Frankel, M. R., Garland, P., Groves, R. M., Kennedy, C., Krosnick, J., Lavrakas, P. J., Lee, S., Link, M, Piekarski, L., Rao, K., Thomas, R. K., & Zahs, D. (2010). AAPOR report on online panels. Public Opinion Quarterly, 74, 711–781. doi:10.1093/poq/nfq048. Baresch, B., Knight, L., Harp, D., & Yaschur, C. (2011). Friends who choose your news: An analysis of content links on Facebook. Paper presented to the International Symposium on Online Journalism, Austin. Berube, D. M. (2011). Researching social media in the 21st century (White paper for Gutmann, Myron P.). Rebuilding the mosaic: Fostering research in the social, behavioral, and economic sciences at the national science foundation in the next decade. Arlington, VA: NSF. Bhutta, C. B. (2012). Not by the Book: Facebook as a sampling frame. Sociological Methods & Research, 41, 57–88. doi:10.1177/0049124112440795. Blumberg, S. J., & Cynamon, M. L. (1999). Misreporting medicaid enrollment: Results of three studies linking telephone surveys to state administrative records. Proceedings of the 7th Conference on Health Survey Research Methodology. DHHS Publication No. 01-1013. Hyattsville, MD. Bothos, E., Apostolou, D., & Mentzas, G. (2010). Using social media to predict events with agent-based markets. IEEE Intelligent Systems, 50–58. doi:10.1109/ MIS.2010.152. Brick, J. M. (2011). The future of survey sampling. Public Opinion Quarterly, 75, 872–888. doi:10.1093/poq/nfr045. Calderwood, L., & Lessof, C. (2009). Enhancing longitudinal surveys by linking to administrative data. In P. Lynn (Ed.), Methodology of longitudinal surveys. Chichester: John Wiley & Sons. Card, D., Chetty, R., Feldstein, M., & Saez, E. (2011). Expanding access to admin- istrative data for research in the United States (White paper for Gutmann, Myron I N T E R N A T I O N A L J O U R N A L O F P U B L I C O P I N I O N R E S E A R C H226 P.). Rebuilding the mosaic: Fostering research in the social, behavioral, and eco- nomic sciences at the national science foundation in the next decade. Arlington, VA: NSF. Choi, H., & Varian, Hal. (2009). Predicting the Present with Google Trends. Unpublished paper, Google Inc. doi:10.2139/ssrn.1659302. Converse, J. M. (1987). Survey research in the United States: Roots and emergence, 1890-1960. Berkeley, CA: University of California Press. Couper, M. P. (2011). The future of modes of data collection. Public Opinion Quarterly, 75, 889–908. doi:10.1093/poq/nfr046. Crompton, R. (2008). Forty years of sociology: Some comments. Sociology, 42, 1218–1227. doi:10.1177/0038038508096942. Dexter, L. A. (1956). What do congressmen hear: The mail. Public Opinion Quarterly, 20, 16–27. doi:10.1086/266594. Gallup, G. (1957). The changing climate of public opinion research. Public Opinion Quarterly, 21, 23–27. doi:10.1086/266682. Gallup, G., & Rae, S. F. (1940). The pulse of democracy: The public–opinion poll and how it works. Westport, CT: Greenwood press. Gayo-Avello, D., Metaxas, P., & Mustafaraj, E. (2011). Limits of electoral predictions using twitter. Paper presented at ICWSM, Barcelona. Golder, S., & Macy, M. (2012). Social science with social media. ASA Footnotes, 40, 5. Gray, D. L., & Brown, T. (1970). Letters to the editor: Hazy reflections of public opinion. Journalism Quarterly, 47, 450–456. doi:10.1177/107769907004700302. Green, Z. (2011). Twitter predicts polls. Retrieved from http://140elect.com. Groves, R. M. (2011a). ‘Designated Data’ and ‘Organic Data’. The Director’s Blog. Retrieved from http://blogs.census.gov. Groves,R. M. (2011b). Three eras of survey research. Public Opinion Quarterly, 75, 861–871. doi:10.1093/poq/nfr057. Gutmann, M. P. (2011). Rebuilding the mosaic: Fostering research in the social, behavioral, and economic sciences at the National Science Foundation in the next decade. Arlington, VA: NSF. Herbst, S. (1993). Numbered voice: How opinion polling has shaped American politics. Chicago: University of Chicago Press. Hewat, C. (2011, May). Using education databases to improve the quality of the LFS. Paper presented to the 6th Workshop on LFS Methodology, Wiesbaden. Hillygus, D. S. (2011). The evolution of election polling in the United States. Public Opinion Quarterly, 75, 962–981. doi:10.1093/poq/nfr054. Holli, M. G. (2002). The wizard of Washington: Emil Hurja, Franklin Roosevelt, and the birth of public opinion polling. New York: Palgrave. Hong, J. I., Zhang, J., & Zimmerman, J. (2011). Understanding human behavior atlarge scales through mobile devices (White paper for, M. P.Gutmann). Rebuilding the mosaic: Fostering research in the social, behavioral, and economic sciences at the National Science Foundation in the next decade. Arlington, VA: NSF. Huynh, M., Rupp, K., & Sears, J. (2000). The Assessment of Survey of Income and Program Participation (SIPP) benefit data using longitudinal administrative records (SIPP Report No. 238) on line at http://www2.census.gov/prod2/sipp/wp/ SIPP_WP_238.pdf. S U R V E Y - R E S E A R C H P A R A D I G M S O L D A N D N E W 227 Kernell, S. (2000). Life before polls: Ohio politicians predict the 1828 Presidential vote. Political Science and Politics, 33, 569–574. doi:10.2307/420860. Kriesberg, M. (1945). What Congressmen and administrators think of the polls. Public Opinion Quarterly, 9, 333–337. Lui, C., Metaxas, P. T., & Mustafaraj, E. (2011, March). On the predictability of the U.S. elections through search volume activity. Paper presented to the e-Society Conference, Avila, Spain. Moore, J. C., & Marquis, K. H. (2010). Using administrative record data to evaluate the quality of survey estimates, research report series, survey methodology 2010-10. Washington, DC: US Census Bureau. (Original work published 1987). Murphy, J. (2012, August). Déjà vu too: Statisticians don’t like non-probability. Survey Post. Retrieved from https://blogs.rti.org/surveypost/2012/08/09/ facebook-and-surveys-thoughts-on-going-not-by-the-book/. O’Connor, B., Balasubramanyan, R., Routledge, B. R., & Smith, N. A. (2010). From tweets to polls: Linking text sentiments to public opinion time series. Tepper School of Business Paper 559. Carnegie Mellon University. Olson, S., & Bunnett, W. (2011). Facebook predicts electoral victory, or not? Retrieved from http://trilogyinteractive.com. Pedace, R., & Bates, N. (2000). Using administrative records to assess earnings reporting error in the survey of income and program participation. Journal of Economic and Social Measurement, 26, 173–192. Permutter, D. D. (2008). Political blogging and campaign 2008: A roundtable. International Journal of Press/Politics, 13, 160–170. doi:10.1177/1940161208315742. Rhodes, A.E., & Fung, K. (2004). Self-reported use of mental health services versus administrative records: Care to recall? International Journal of Methods in Psychiatric Research, 13, 165–175. doi:10.1002/mpr.172. Robinson, C. E. (1932). Straw votes: A study of political prediction. New York: Columbia University Press. Saiz, A., & Simonsohn, U. (2007). Downloading wisdom from online crowds. Unpublished report, Wharton School, University of Pennsylvania. doi: 10.2139/ ssrn.990021. Savage, M., & Burrows, R. (2007). The coming crisis of empirical sociology. Sociology, 41, 885–899. doi:10.1177/0038038507080443. Savage, M., & Burrows, R. (2009). some further reflections on the coming crisis of empirical sociology. Sociology, 43, 762–771. doi:10.1177/0038038509105420. Schaeffer, N. C., & Dykema, J. (2011). Questions for surveys: Current trends and future direction. Public Opinion Quarterly, 75, 909–961. doi:10.1093/poq/nfr048. Smith, T. W. (2011a). Contextualizing social-science data (White paper for M. P. Gutmann). Rebuilding the Mosaic: Fostering research in the social, behavioral, and eco- nomic sciences at the national science foundation in the next decade. Arlington, VA: NSF. Smith, T. W. (2011b). The report on the international workshop on using multi-level data from sample frames, auxiliary databases, paradata, and related sources to detect and adjust for nonresponse bias in surveys. International Journal of Public Opinion Research, 23, 389–402. doi:10.1093/ijpor/edr035. Smith, T. W., & Kim, J. (2009). An assessment of the multi-level integrated database approach (MIDA) (GSS Methodological Report No. 116). Chicago: NORC. I N T E R N A T I O N A L J O U R N A L O F P U B L I C O P I N I O N R E S E A R C H228 https://blogs.rti.org/surveypost/2012/08/09/facebook-and-surveys-thoughts-on-going-not-by-the-book/ https://blogs.rti.org/surveypost/2012/08/09/facebook-and-surveys-thoughts-on-going-not-by-the-book/ http://trilogyinteractive.com Smyth, J. D. (2012, January). On the frontier of Web surveys. Paper presented to NORC, Chicago. Splichal, S. (1999). Public opinion: Developments and controversies in the twentieth cen- tury. Lanham, MD: Rowman & Littlefield. Stacey, E. C., Pauwels, K. H., & Lackman, A. (2012). Beyond likes and tweets: How conversation content drives store and site traffic. Unpublished report, The Center for Measureable Marketing. Retrieved from http://pages.stern.nyu.edu/�atakos/ centerresearch/likesandtweets.pdf. Stockwell, I. (2011). The role of data and analytical methods in public discovery and implementation (White paper for M. P. Gutmann). Rebuilding the mosaic: Fostering research in the social, behavioral, and economic sciences at the national science founda- tion in the next decade. Arlington, VA: NSF. Tumasjan, A., Sprenger, T. O., Sandner, P. G., & Welpe, I. M. (2010). Predicting elections with twitter: What 140 characters reveal about political sentiment. Proceedings of the Fourth International AAAI Conference on Weblogs and Social Media. Washington, DC, USA. Van der Putten, P., & Kok, J. N. (2010). Using data fusion to enrich customer databases with survey data for database marketing. In J. Casillas & F. J. M. Lopez (Eds.), Marketing intelligent systems using soft computing. Springer studies in fuzziness and soft computing, (Vol. 258, pp. 113–130). doi:10.1007/978-3-642-15606-9_11. Van Dongen, R. (2012, January 20). Paul beats Mitt in Facebook chatter. Retrieved from http://political.com. Wahl-Jorgensen, K. (2002). The construction of the public in letters to the editor: Deliberative democracy and the idiom of insanity. Journalism, 3, 183–204. doi:10.1177/146488490200300203. Wahl-Jorgensen, K. (2001). Letters to the editor as a forum for public deliberation: modes of publicity and democratic debate. Critical Studies in Media Communications, 18, 303–320. doi:10.1080/07393180128085. Webber, R. (2009). Response to ‘The Coming Crisis of Empirical Sociology’: An outlineof the research potential of administrative and transactional data. Sociology, 43, 169–178. doi:10.1177/0038038508099104. Wilson, E. C. (1947). The Measurement of Public Opinion. Annals of the American Academy of Political and Social Science, 250, 121–129. doi:0.1177/00027162472 5000117. Wolfers, J. (2011, May 25). Google’s new correlation mining tool: It works! Retrieved from http://www.freakonomics.com. Yeager, D. S., Krosnick, J. A., Chang, L., Javitz, H. S., Levendusky, M. S., Simpser, A., & Wang, R. (2011). Comparing the accuracy of RDD telephone sur- veys and Internet surveys conducted with probability and non-probability samples. Public Opinion Quarterly, 75, 709–747. doi:10.1093/poq/nfr020. S U R V E Y - R E S E A R C H P A R A D I G M S O L D A N D N E W 229 http://pages.stern.nyu.edu/atakos/centerresearch/likesandtweets.pdf http://pages.stern.nyu.edu/atakos/centerresearch/likesandtweets.pdfhttp://pages.stern.nyu.edu/atakos/centerresearch/likesandtweets.pdf http://political.com http://www.freakonomics.com
Compartir