What Makes News Sharable on Social Media?
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
What Makes News Sharable on Social Media? Cathy Xi Chen1, Gordon Pennycook2,3, David G. Rand1,4,5 1 Sloan School of Management, Massachusetts Institute of Technology, 2 Hill/Levene Schools of Business, University of Regina, 3 Department of Psychology, University of Regina, 4 Institute for Data, Systems, and Society, Massachusetts Institute of Technology, 5 Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology With the rise of social media, everyone has the potential to be both a consumer and producer of online content. As a result, the role that word of mouth plays in news consumption has been dramatically increased. Although one might assume that consumers share news because they believe it to be true, widespread concerns about the spread of misinformation suggest that truthfulness may actually not be a dominant driver of sharing online. Across two studies with 5,000 participants, we investigate what makes news sharable on social media. We find that sharing is positively predicted by two separate factors. One factor does involve the headline’s perceived accuracy, as well as its familiarity. The second, however, involves the headline’s perceived importance and emotional evocativeness. This second factor is negatively associated with the headline’s objective veracity, and less decision weight is put on the second factor by subjects with more cognitive reflection and political knowledge, and by subjects who are less politically conservative. These findings have important implications for news publishers, social media platforms, and society at large. Keywords: social media; news sharing; word of mouth; online consumer behavior; misinformation Posted 7/9/2021 This working paper has not yet been peer reviewed Introduction The advent of social media has drastically changed the way that consumers share information (e.g., Lamberton & Stephen, 2016). One area where this change is particularly salient is news, with more people than ever getting their news from social media (e.g., Pew Research Center, 2017). Traditionally, a small number of news outlets broadcast news to millions of passive consumers. On social media, however, everyone has the potential to be both a consumer and producer of content. Hence, the rise of social media dramatically increases the role word of mouth communication plays in news consumption. Among many other things, this dynamic has been argued to have fueled the proliferation of misinformation and “fake news” (Lazer et al., 2018), which has important consequences for a wide range of consumer behaviors, and major economic impacts. Therefore, in the current work, we use the perspective of word of mouth to investigate how consumers decide what true versus false or misleading news to share on social media.
A substantial body of research on word of mouth has examined why consumers talk about certain topic and content with others (e.g., Berger & Schwartz, 2011; Toubia & Stephen, 2013; Chen & Berger, 2013; De Angelis, et al., 2012; see Berger, 2014, for a review). In addition, research has also investigated how features of the content itself affects sharing. For example, content that elicits more positive feelings (such as interesting or humorous) is more likely to spread (e.g., Bakshy, et al., 2011; Warren & Berger, 2011). Beyond simply evoking positive feelings, emotional content that specifically evokes high arousal (such as awe, anger, or anxiety) is more likely to be shared (e.g., Berger, 2011; Berger & Milkman, 2012); and sentiment volatility is suggested to be important in making cultural products more successful (Berger, Kim, & Meyer, 2021). Furthermore, content that is seen as useful and informative is also more likely to be shared (e.g., Berger & Milkman, 2012; Heath, Bell, & Sternberg, 2001). In the current work, we shed new light on word of mouth by investigating the social media sharing of news that varies in objective truth. Although one might naively assume that consumers only share news that they believe to be true, in fact empirical evidence suggests that false content may spread as much (Grinberg et al., 2019) or more (Vosoughi, Roy, & Aral, 2018) than similar true content. Experimental evidence also demonstrates a disconnect between accuracy judgments and sharing intentions: despite saying that accuracy is important to them, people sometimes (even often) share news that they could identify as being false if they considered its accuracy (Pennycook, Epstein, et al., 2021; Epstein, et al., 2021; Pennycook, McPhetres, et al., 2020). A natural question raised by these observations, then, is: If accuracy is not a primary driver of choices about what news content to share, then what does predict sharing of accurate vs inaccurate news? From prior work, we can assemble a set of hypotheses. The work reviewed above suggests that emotional evocativeness, informational value, and usefulness/importance may predict news sharing intentions (e.g., Berger, 2014). Furthermore, people are more likely to share news that aligns with their political ideology, in both survey experiments (Pennycook, Cannon, & Rand, 2018; Pennycook & Rand, 2019; Pennycook, Epstein, et al., 2021) and panel studies on social media (Grinberg et al., 2019). Another line of work suggests that familiarity may influence sharing, with Twitter data finding that novel content is more likely to be shared (Vosoughi, Roy, & Aral, 2018) but survey data suggesting that familiarity is positively associated with sharing (Pennycook & Rand, 2020). Therefore, in the current work, we examine the relationship between sharing intentions and each of these content dimensions, as well as perceived accuracy. We also examine individual differences in the relationship between these content dimensions and sharing. For example, prior work has found that people who engage in less cognitive reflection (i.e., analytic thinking) have been found to share more false content in survey experiments (Pennycook & Rand, 2019; Pennycook & Rand, 2020; Bronstein et al. 2019; Ross, Rand, & Pennycook, 2021) and share more news from less reliable sources on Twitter (Mosleh, et al., 2021; although see Osmundsen et al., 2021). Furthermore, studies of Twitter and Facebook have found that Republicans share more false content than Democrats (Grinberg et al., 2019; Guess,
Nagler, & Tucker, 2019). Thus, we ask how these and other individual characteristics predict the weight put on different content dimensions when deciding what to share. Methods Participants Study 1 was conducted during June 2019. For robustness, we recruited American participants from two different online subject recruitment platforms, a convenience sample from Amazon’s Mechanical Turk (N = 2,006) and a sample from Lucid that was quota-matched to the national distribution on age, gender, ethnicity, and geographic region (N = 1,985). The two datasets were combined for the following analyses (N = 3,991). Study 2 was conducted during March 2020. We recruited American participants from Amazon’s Mechanical Turk (N = 1,009). We pre-registered Study 2, with the preregistration available at: https://osf.io/q6gwy/. Headlines Participants in both studies were shown a set of headlines which were randomly selected from a large pool of actual news headlines, of which 1/3 were true, 1/3 were false, and 1/3 were misleading headlines (see https://osf.io/q6gwy/ for full list of headlines in each study). In Study 1, a list of 210 headlines (70 true, 70 false, and 70 misleading) was created from a variety of online sources (e.g. social media, Snopes, Breitbart). Participants were randomly assigned to only see one type of news (e.g. true, false, or misleading), and they were shown an image of each headline as it would appear on Facebook. Each participant rated a random subset of 10 headlines. In Study 2, a list of 216 headlines (72 true, 72 false, and 72 misleading) was created from a range of sources as in Study 1. Some of the headlines were the same as those used in Study 1, but some were replaced with more recent content. Participants were not restricted to a single type of news as they were in Study 1, and each participant rated a random subset of 18 headlines. Content Ratings Given the work reviewed in the introduction section, which suggests a variety of links between content dimensions and sharing, we asked participants to rate each headline on the following dimensions. In Study 1, participants rated Truth (“What is the likelihood that the above headline is true?”), Importance (“Assuming the headline is entirely accurate, how important would this news be?”), Familiarity (“Are you familiar with the above headline (have you seen or heard about it before?)”, Partisanship (“Assuming the above headline is entirely accurate, how favorable would it be to Democrats versus Republicans?”; we constructed a Politically Concordant variable by reversing the scores for participants who were Democrats), and Emotionality (“How exciting is this headline?” and “How worrying is this headline?”). Study 2
used the same measures of Truth, Importance, Familiarity, and Partisanship, but measured Emotionality differently (“How anger provoking is this headline?” and “How anxiety provoking is this headline”?), and added measures for Humor (“How funny is this headline?”) and Informativeness (“How informative is this headline?”). Dimension questions were all elicited using Likert scales, and were asked in fixed order. Last, participants were asked about Sharing (“If you were to see the above article on social media, how likely would you be to share it?”). For full materials, see https://osf.io/q6gwy/. Individual Difference Measures In both studies, after the news headline tasks, participants completed a variety of individual difference measures including the Cognitive Reflection Test (Frederick, 2005; CRT) which measures their tendency to engage in analytic thinking versus simply going with their intuitive first response, a set of questions to test their political knowledge, a set of questions about social media usage, a set of political position and preferences questions, as well as demographics including age, gender, income, and education. Cognitive Reflection. A 6-item version of the cognitive reflection test (Pennycook & Rand, 2018) was measured. Correct responses were summed, resulting in a score ranging from 0 (less reflective) to 6 (more reflective). Political Knowledge. Participants were asked to respond to a 5-item political knowledge quiz. Each question was shown individually, and participants were given a 10-second window to respond before the page advanced automatically. Correct responses were summed. Social Media Frequency. Participants were asked a set of questions regarding social media, including whether they have a Facebook account (Yes or No) and a Twitter account (Yes or No). They indicated how frequently they use social media accounts, with options ranging from 1 (never) to 5 (daily). Political Position & Preferences. Participants were asked a set of questions regarding their political position and preferences: (1) “Which of the following best described your political position?” (Democrat, Republican, Independent, or other); (2) “Which of the following best described your political preference?” (1 = strongly Democratic, 3 = lean Democratic, 4 = lean Republican, 6 = strongly Republican); (3) they were asked to rate their political position on social issues and economic issues from 1 (Strongly liberal) to 5 (Strongly conservative); (4) they were also asked a variety of other questions, including who they voted for during the 2016 presidential election and the 2018 Congressional Election, how they feel towards President Donald Trump, whether they support or oppose his presidency, and if they plan to vote for him again in the 2020 election. Since the ratings on these questions were highly correlated with each other, we used Principal Component Analysis (PCA) to reduce dimensions and construct a more effective measure. We found the first principal component described most of the variation (about 74%), so this component was used to construct an overall measure of political conservatism.
Results The Latent Structure of the Content Ratings We begin by assessing how the various content dimensions relate to each other. That is, we ask to what extent the different dimensions actually reflect a smaller number of underlying latent dimensions. To do so, we conducted exploratory factor analysis (EFA), used parallel analysis (Horn, 1965) to determine the number of factors to retain in a principled way, and then used varimax rotation to determine loadings. In Study 1, we had 6 content ratings: “true”, “familiar”, “important”, “political concordant”, “worrying”, and “exciting”. Parallel analysis led us to retain two factors, shown in Table 1: Factor 1 included high loadings on important (0.784), worrying (0.685), and exciting (0.596); and Factor 2 included high loadings on familiar (0.577) and true (0.499). For robustness, we also checked a 3-factor solution (see Table A1 in Appendix), and found that Factor 1 kept the same structure, while Factor 2 was separated into two factors: one included high loadings on familiar, and the other included high loadings on true and concordant. Table 1. Exploratory Factor Analysis with Varimax Rotation (Study 1, 2- factor solution) 1. Loadings above 0.4 shown in bold. Item Factor 1 Factor 2 Important 0.784 0.058 Worrying 0.685 0.081 Exciting 0.596 0.251 Familiar 0.166 0.577 True 0.179 0.499 Concordant -0.008 0.259 In Study 2, we had 8 content ratings: “true”, “familiar”, “important”, “political concordant”, “anger provoking”, “anxiety provoking”, “funny”, and “informative”, and found a broadly similar pattern. Parallel analysis again indicated two factors, shown in Table 2. Factor 1 included high loadings on anger (0.785), anxiety (0.778), important (0.524); and Factor 2 included high loadings on familiar (0.876), funny (0.617), and true (0.489). Furthermore, Informative loaded heavily on both factors (0.451 and 0.482). In terms of 3-factor solution (see Table A2 in Appendix), similar to what we found in Study 1, while Factor 1 kept the same structure, Factor 2 was separated into two factors: one included high loadings on familiar and funny, while the other included high loadings on true and informative. 1 In Study 1, the proportion of variance explained by Factor 1 and Factor 2 are 25% and 12%, respectively.
Table 2. Exploratory Factor Analysis with Varimax Rotation (Study 2, 2- factor solution) 2. Loadings above 0.4 shown in bold. Item Factor 1 Factor 2 Anger 0.785 0.267 Anxiety 0.778 0.361 Important 0.524 0.132 Familiar 0.260 0.876 Funny 0.189 0.617 True 0.240 0.489 Informative 0.451 0.482 Concordant 0.054 0.181 In sum, we find a similar 2-factor structure in both studies1. In particular, Factor 1 captures the perceived importance and emotional evocativeness of the headlines in both studies, while Factor 2 captured perceived accuracy and familiarity in both studies; with the addition of humorousness and importance in Study 2 (not collected in Study 1). Relationships with Sharing Intentions and Headline Veracity In the previous section, we found the consistent existence of a factor of content dimensions unrelated to perceived accuracy (Factor 1). Next we ask how social media sharing intentions relate to the various content dimensions, and in particular to the 2 factors identified above. We find that all of the content dimensions assessed were positively correlated with sharing intentions (see Tables 3 and 4). Interestingly, in both studies, all of the correlations were greater than r=0.30 except for political concordance, which was consistently the least strongly correlated dimension. We then explored the relationships between the two factors and sharing intentions. To do so, we ran OLS regressions with the two factors’ scores and their interaction term as independent variables, and sharing intentions as the dependent variable (one observation per rating, standard errors clustered on participant). All variables were standardized. The results are presented in Table 5. Across both studies, the coefficient on Factor 1 (Study 1: β=.464, SE=.007, p
factors represent distinct routes that lead to increased willingness to share news content – and, most importantly, that Factor 1 (which was unrelated to perceived accuracy but had strong loadings on importance and emotional evocativeness) was a consistent predictor of sharing intentions. This observation suggests that these content dimensions may contribute to the spread of misinformation. Further support for this conclusion comes from examining the relationship with the objective veracity of the news, as opposed to sharing intentions. To do so, we conducted a set of post-hoc analyses where headline veracity was coded as 1 if the news headline was true and 0 if the news headline was false or misleading. We again started with examining the correlations between each content rating and headline veracity (see Tables 3 and 4). Turning to the factors, OLS regressions (with all variables standardized) were run to predict headline veracity with the same set of independent variables (see Table 5). Across both studies, the coefficient on Factor 1 (Study 1: β=-0.071, SE=.010, p
Table 4. The Pearson Correlation between sharing intentions, headline veracity, and content ratings in Study 2. 1 2 3 4 5 6 7 8 9 1. Sharing - Intention 2. Headline 0.05 - Veracity 3. Anger 0.45 -0.02 - 4. Anxiety 0.54 -0.01 0.72 - 5. Important 0.36 0.05 0.43 0.43 - 6. Familiar 0.65 0.06 0.44 0.52 0.24 - 7. Funny 0.56 -0.09 0.33 0.40 0.08 0.6 - 8. True 0.48 0.28 0.29 0.33 0.26 0.49 0.25 - 9. Informative 0.58 0.13 0.44 0.51 0.47 0.53 0.35 0.52 - 10. Concordant 0.18 0.04 0.08 0.10 0.09 0.15 0.16 0.17 0.15 Table 5. OLS regression analyses (standardized) predicting either sharing intention or headline veracity using the two factor scores and their interaction term. Standard errors (clustered on participants) are in parentheses. Dependent Variable: Sharing Intention Headline Veracity Study 1 Study 2 Study 1 Study 2 -0.015* -0.008 0.004 0.007 (Intercept) (0.009) (0.014) (0.015) (0.008) 0.464*** 0.363*** -0.071*** -0.031*** Factor 1 (0.007) (0.011) (0.010) (0.008) 0.287*** 0.570*** 0.226*** 0.103*** Factor 2 (0.007) (0.011) (0.010) (0.009) 0.095*** 0.054*** -0.028** -0.052*** Factor 1 × Factor 2 (0.006) (0.011) (0.009) (0.009) Observations 39,635 18,162 39,635 18,162 R2 0.366 0.535 0.048 0.009 Adjusted R2 0.366 0.535 0.048 0.009 Note. *p
Individual Differences in Decision Weights Put on the Two Factors We now investigate individual heterogeneity in the weights placed on these two factors. 3 First, we estimated the coefficients of the two factors for each participant when predicting sharing intentions. 4 In order to better account for both subject-level and headline-level random effects, instead of using OLS regressions as planned in Study 2’s pre-registration, we used Bayesian multilevel models to estimate each individual’s factor coefficients. Next, we ran two OLS regressions (one for each factor) to predict the participant’s coefficient for each of the two factors based on the individual difference measures: demographics (age, gender, education, income), CRT score, political knowledge score, political conservatism, and frequency of social media use. The results are shown in Table 6. For completeness, we also examined the Pearson correlations between the dependent variables (i.e. the coefficients on the two factors when predicting sharing intentions) and the individual difference measures (see Tables 7 and 8), as some of the individual difference measures showed fairly high level of correlation with each other (although multicollinearity was not especially high, with all VIFs .61; see Tables A3 and A4 in Appendix). Two results emerge consistently across both studies. First, in the pairwise correlations, CRT and political knowledge were significantly negatively associated with the coefficients on Factor 1. In the multiple regression including both variables together, only CRT remained significant in Study 2 (although because of the high correlation between CRT and political knowledge in Study 2, r=0.55, the non-significance of political knowledge in the multiple regression should be interpreted with caution). Second, conservatism was significantly positively associated with the coefficient on Factor 1 in both studies. The demographic variables (i.e., age, gender, and income) were significantly associated with either the coefficient on Factor 1 or Factor 2 in Study 1, but these associations did not replicate in Study 2. Social media frequency was not significant associated with the coefficient on either factor in either study. 3 In these participant-level analyses, we excluded participants who gave the same response on the sharing intention question to all headlines (24.9% in Study 1, 22.6% in Study 2), because for these participants there is no variation in the outcome variable to predict. In addition, participants who did not complete the additional measures were also excluded from the following analyses (2.6% in Study 1, 1.4% in Study 2). 4 To make estimation tractable – and given the very small magnitude of the interaction terms in Table 5 – we do not include the interaction term in the individual-level models.
Table 6. OLS regression analyses (standardized) predicting a) coefficient for Factor 1 and b) coefficient for Factor 2 using four additional measures including CRT, political knowledge, conservatism, social media frequency, and four demographic measures including age, gender (Female = 1, Male = 0), education (college and more = 1, others = 0), income (more than $50,000 = 1, others = 0). Standard errors are in parentheses. Dependent Variable: Coefficient for Coefficient for Factor 1 Factor 2 Study 1 Study 2 Study 1 Study 2 0.000 0.000 0.000 0.000 (Intercept) (0.018) (0.036) (0.019) (0.035) -0.040* 0.047 -0.038 0.006 Age (0.019) (0.037) (0.020) (0.037) -0.085*** -0.024 -0.001 0.021 Gender (0.018) (0.036) (0.019) (0.036) -0.030 -0.053 0.031 0.015 Education (0.019) (0.038) (0.020) (0.038) -0.001 -0.041 -0.041* -0.051 Income (0.019) (0.037) (0.020) (0.037) -0.184*** -0.123** 0.029 0.080 CRT (0.020) (0.043) (0.021) (0.043) Political -0.127*** -0.026 0.018 0.171*** Knowledge (0.020) (0.045) (0.021) (0.045) 0.049** 0.098* 0.028 0.057 Conservatism (0.019) (0.039) (0.019) (0.039) Social Media 0.006 -0.008 0.018 0.045 Frequency (0.018) (0.036) (0.019) (0.036) Observations 2,892 767 2,892 767 R 2 0.083 0.037 0.005 0.050 Adjusted R2 0.081 0.027 0.002 0.040 Note. *p
Table 7. The Pearson correlation for the two factor coefficients and the individual difference measures in Study 1 (All variables have been standardized; *p
Table 8. The Pearson correlation for the two factor coefficients and the individual difference measures in Study 2 (All variables have been standardized; *p
Discussion In the current work, we examined how various features of news headlines predict consumers’ social media sharing intentions, and how that varies based on individual characteristics. From two studies with 5,000 participants, we found a consistent 2-factor structure to how different content dimensions related to each other: Factor 1 was highly loaded with “important”, “worrying”, “exciting” in Study 1, and “important”, “anger-provoking”, “anxiety-provoking” in Study 2; Factor 2 was highly loaded with “true” and “familiar” in Study 1, and “true”, “familiar”, and “funny” in Study 2. This factor structure reveals relationships that were not necessarily obvious ex ante, highlighting how importance and emotional evocativeness are distinct from perceived accuracy, and how perceived accuracy and familiarity consistently track each out. These observations help to further illuminate how consumers think about the news they read. Furthermore, each content dimension, as well as both factors, were significantly positively associated with sharing likelihood. Perhaps the most parsimonious interpretation of Factor 1 is that it represents how engaging participants find the headlines to be. This is particularly striking, given that this engagingness factor was positively correlated with sharing intentions, but negatively correlated with headline veracity. False or misleading claims were more engaging – which may help to explain why inaccurate news is often shared as much, or even more, than accurate news (Grinberg et al., 2019; Vosoughi, Roy, & Aral, 2018). This observation coincides with the individual differences associated with weighting this factor more when deciding what to share. Specifically, we found that people with higher CRT scores and more political knowledge – as well as less conservative political inclinations – placed lower average weights on this engagingness factor. These results connect with the prior observation that CRT is associated with more truth discernment in sharing intentions (Pennycook, McPhetres, et al., 2020; Ross, Rand, Pennycook, 2019) as well as actually sharing news from more trustworthy sources on Twitter (Mosleh, et al., 2021). Research also suggests that conservatives share more fake news on Twitter (Grinberg et al. 2019) and Facebook (Guess et al. 2019). Our findings suggest that these differences in sharing behavior may result not from differences in how much people attend to the perceived truth of the headlines (which loaded on Factor 2) but instead from differences in how much they are influenced by the content’s engagingness. This observation resonates with recent findings that the sharing of misinformation is often driven by inattention to accuracy rather than the inability to discern accuracy when attending to it (Pennycook McPhetres, et al., 2020; Pennycook, Epstein et al., 2021). The interpretation of Factor 2 is somewhat less straightforward. Familiarity and perceived truth loaded on this factor in both studies, which might suggest that the factor captures how plausible a headline seems. It is because these results connect with the well-established “illusory truth effect”, whereby repetition (i.e., low novelty) increases perceived accuracy (Hasher, Goldstein, & Toppino, 1977; see Dechêne, et al., 2010 for a review), even for blatantly false fake news (Pennycook Cannon, & Rand, 2018). However, the high loading of “funniness” in Study 2 is somewhat inconsistent with this characterization. Furthermore, the lack of consistent individual difference correlations with the coefficient for Factor 2 also does not help to clarify the factor’s interpretation. Future work should investigate this second factor in more details.
Our findings make several contributions to the existing literature. Our results demonstrate how engagingness is separate from truth, and emphasize the role that engagingness is likely to play in social media consumers’ choice to share inaccurate news online. These results resonate with past research on word of mouth suggesting that content evoking high-arousal emotions is more viral (e.g. Berger & Milkman, 2012), extending those results by demonstrating the disconnect between such emotions and objective truth. This disconnect has important implications for understanding the dynamics of news on social media. In particular, our findings shed light on why false content often spreads online (e.g. Vosoughi, Roy, & Aral, 2018). One might imagine that people share misinformation because consumers purposeful downplay truth. However, our results suggest that it may instead be because consumers are drawn to content that is engaging irrespective of its accuracy – and inaccurate content is often more engaging (likely because content creators can create more engaging content if they are not constrained by the truth). This observation can help to inform attempts of social media companies to reduce the spread of misinformation, as well as news organizations that are seeking to generate news that succeeds online. Furthermore, our work brings together work relevant to online word of mouth from several different disciplines beyond just marketing, a topic that is receiving considerable attention from fields including cognitive psychology (see Pennycook & Rand, 2021 for a review), economics (e.g., Allcott & Gentzkow, 2017), communication and media studies (e.g., Metzger et al., 2021; AI-Rawi, 2019; Kümpel, Karnowski, & Keyling, 2015; Lee & Ma, 2012), and computational social science (e.g., Goel, et al., 2016; Rudat, Buder, & Hesse, 2014). In doing so, we hope to help foster cross- disciplinary work on the important topic of online news sharing. Finally, there are several limitations of the current work that are important to acknowledge, and various important directions for future research. In addition to clarifying the nature of the second factor we identified, future work should investigate the relationship between sharing, veracity, and other content dimensions not considered here. Second, the current work uses self-reported measures of sharing intentions instead of actual sharing decision data from field experiments. Although there is reason to believe that self-report sharing intentions are informative (Mosleh, et al., 2020), future research could design field experiments on the social media platforms such as Twitter to improve ecological validity. Third, since the current work focuses on news sharing decisions, future work could assess how our results generalize to other contexts such as digital marketing related decision tasks, or other online content sharing decisions.
Reference Al-Rawi, A. (2019). Viral news on social media. Digital journalism, 7(1), 63-79. Allcott, H., & Gentzkow, M. (2017). Social media and fake news in the 2016 election. Journal of economic perspectives, 31(2), 211-36. Bakshy, E., Hofman, J. M., Mason, W. A., & Watts, D. J. (2011, February). Everyone's an influencer: quantifying influence on twitter. In Proceedings of the fourth ACM international conference on Web search and data mining (pp. 65-74). Berger, J. (2011). Arousal increases social transmission of information. Psychological science, 22(7), 891-893. Berger, J. (2014). Word of mouth and interpersonal communication: A review and directions for future research. Journal of consumer psychology, 24(4), 586-607. Berger, J., & Milkman, K. L. (2012). What makes online content viral?. Journal of marketing research, 49(2), 192-205. Berger, J., & Schwartz, E. M. (2011). What drives immediate and ongoing word of mouth?. Journal of marketing research, 48(5), 869-880. Berger, J., Kim, Y. D., & Meyer, R. (2021). What Makes Content Engaging? How Emotional Dynamics Shape Success. Journal of Consumer Research. Bronstein, M. V., Pennycook, G., Bear, A., Rand, D. G., & Cannon, T. D. (2019). Belief in fake news is associated with delusionality, dogmatism, religious fundamentalism, and reduced analytic thinking. Journal of Applied Research in Memory and Cognition, 8(1), 108-117. Chen, Z., & Berger, J. (2013). When, why, and how controversy causes conversation. Journal of Consumer Research, 40(3), 580-593. Dechêne, A., Stahl, C., Hansen, J., & Wänke, M. (2010). The truth about the truth: A meta- analytic review of the truth effect. Personality and Social Psychology Review, 14(2), 238- 257. De Angelis, M., Bonezzi, A., Peluso, A. M., Rucker, D. D., & Costabile, M. (2012). On braggarts and gossips: A self-enhancement account of word-of-mouth generation and transmission. Journal of Marketing Research, 49(4), 551-563. Epstein, Z., Berinsky, A. J., Cole, R., Gully, A., Pennycook, G., & Rand, D. G. (2021). Developing an accuracy-prompt toolkit to reduce COVID-19 misinformation online. Harvard Kennedy School Misinformation Review. Frederick, S. (2005). Cognitive reflection and decision making. The Journal of Economic Perspectives, 19(4), 25-42.
Goel, S., Anderson, A., Hofman, J., & Watts, D. J. (2016). The structural virality of online diffusion. Management Science, 62(1), 180-196. Grinberg, N., Joseph, K., Friedland, L., Swire-Thompson, B., & Lazer, D. (2019). Fake news on Twitter during the 2016 US presidential election. Science, 363(6425), 374-378. Guess, A., Nagler, J., & Tucker, J. (2019). Less than you think: Prevalence and predictors of fake news dissemination on Facebook. Science advances, 5(1), eau4586. Hasher, L., Goldstein, D., & Toppino, T. (1977). Frequency and the conference of referential validity. Journal of Verbal Learning and Verbal Behavior, 16, 107–112. http://dx.doi.org/10.1016/S0022-5371(77) 80012-1 Heath, C., Bell, C., & Sternberg, E. (2001). Emotional selection in memes: the case of urban legends. Journal of personality and social psychology, 81(6), 1028. Horn, J. L. (1965). A rationale and test for the number of factors in factor analysis. Psychometrika, 30(2), 179-185. Kümpel, A. S., Karnowski, V., & Keyling, T. (2015). News sharing in social media: A review of current research on news sharing users, content, and networks. Social media+ society, 1(2), 2056305115610141. Lamberton, C., & Stephen, A. T. (2016). A thematic exploration of digital, social media, and mobile marketing: Research evolution from 2000 to 2015 and an agenda for future inquiry. Journal of Marketing, 80(6), 146-172. Lazer, D. M., Baum, M. A., Benkler, Y., Berinsky, A. J., Greenhill, K. M., Menczer, F., ... & Schudson, M. (2018). The science of fake news. Science, 359(6380), 1094-1096. Lee, C. S., & Ma, L. (2012). News sharing in social media: The effect of gratifications and prior experience. Computers in human behavior, 28(2), 331-339. Metzger, M. J., Flanagin, A. J., Mena, P., Jiang, S., & Wilson, C. (2021). From Dark to Light: The Many Shades of Sharing Misinformation Online. Media and Communication, 9(1), 134-14 Mosleh, M., Pennycook, G., Arechar, A. A., & Rand, D. G. (2021). Cognitive reflection correlates with behavior on Twitter. Nature communications, 12(1), 1-10. Mosleh, M., Pennycook, G., & Rand, D. G. (2020). Self-reported willingness to share political news articles in online surveys correlates with actual sharing on Twitter. Plos one, 15(2), e0228882. Osmundsen, M., Bor, A., Vahlstrup, P. B., Bechmann, A., & Petersen, M. B. (2021). Partisan polarization is the primary psychological motivation behind political fake news sharing on Twitter. American Political Science Review, 1-17.
Pennycook, G., Cannon, T. D., & Rand, D. G. (2018). Prior exposure increases perceived accuracy of fake news. Journal of experimental psychology: general, 147(12), 1865. Pennycook, G., & Rand, D. G. (2019). Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition, 188, 39–50 Pennycook, G., & Rand, D. G. (2020). Who falls for fake news? The roles of bullshit receptivity, overclaiming, familiarity, and analytic thinking. Journal of personality, 88(2), 185-200. Pennycook, G., & Rand, D. G. (2021). The psychology of fake news. Trends in cognitive sciences. Pennycook, G., McPhetres, J., Zhang, Y., Lu, J. G., & Rand, D. G. (2020). Fighting COVID-19 Misinformation on Social Media: Experimental Evidence for a Scalable Accuracy-Nudge Intervention. Psychological Science, 0956797620939054. Pennycook, G., Epstein, Z., Mosleh, M., Arechar, A. A., Eckles, D., & Rand, D. G. (2021). Shifting attention to accuracy can reduce misinformation online. Nature, 1-6. Pew Research Center, September, 2017, “News Use Across Social Media Platforms 2017” https://www.journalism.org/2017/09/07/news-use-across-social-media-platforms- 2017/pi_17-08-23_socialmediaupdate_0-01/ Ross, R. M., Rand, D. G., & Pennycook, G. (2021). Beyond" fake news": Analytic thinking and the detection of false and hyperpartisan news headlines. Judgment & Decision Making, 16(2). Rudat, A., Buder, J., & Hesse, F. W. (2014). Audience design in Twitter: Retweeting behavior between informational value and followers’ interests. Computers in human behavior, 35, 132-139. Toubia, O., & Stephen, A. T. (2013). Intrinsic vs. image-related utility in social media: Why do people contribute content to twitter?. Marketing Science, 32(3), 368-392. Vosoughi, S., Roy, D., & Aral, S. (2018). The spread of true and false news online. Science, 359(6380), 1146-1151. Warren, C., & Berger, J. (2011). The influence of humor on sharing. ACR North American Advances.
Appendix Table A1. Exploratory Factor Analysis with Varimax Rotation (Study 1, 3-factor solution). Loadings above 0.35 shown in bold. Item Factor 1 Factor 2 Factor 3 Important 0.777 0.013 0.095 Worrying 0.701 0.112 -0.054 Exciting 0.599 0.145 0.173 Familiar 0.136 0.966 0.211 True 0.196 0.221 0.375 Concordant -0.012 0.040 0.404 Table A2. Exploratory Factor Analysis with Varimax Rotation (Study 2, 3-factor solution). Loadings above 0.5 shown in bold. Item Factor 1 Factor 2 Factor 3 Anger 0.787 0.247 0.151 Anxiety 0.770 0.335 0.203 Important 0.490 -0.083 0.423 Familiar 0.259 0.717 0.400 Funny 0.176 0.693 0.142 True 0.157 0.276 0.579 Informative 0.368 0.236 0.685 Concordant 0.023 0.141 0.169
Table A3. The VIF and Tolerance of the individual difference measures (Study 1). IVs VIF Tolerance Age 1.118 0.894 Gender 1.061 0.943 Education 1.140 0.877 Income 1.127 0.887 CRT 1.233 0.810 Political Knowledge 1.300 0.769 Conservatism 1.079 0.926 Social Media Frequency 1.021 0.979 Table A4. The VIF and Tolerance of the individual difference measures (Study 2). IVs VIF Tolerance Age 1.093 0.915 Gender 1.031 0.970 Education 1.158 0.863 Income 1.081 0.925 CRT 1.489 0.671 Political Knowledge 1.617 0.619 Conservatism 1.199 0.834 Social Media Frequency 1.036 0.966
You can also read