The echo chamber effect on social media (original) (raw)
Significance
We explore the key differences between the main social media platforms and how they are likely to influence information spreading and the formation of echo chambers. To assess the different dynamics, we perform a comparative analysis on more than 100 million pieces of content concerning controversial topics (e.g., gun control, vaccination, abortion) from Gab, Facebook, Reddit, and Twitter. The analysis focuses on two main dimensions: 1) homophily in the interaction networks and 2) bias in the information diffusion toward like-minded peers. Our results show that the aggregation in homophilic clusters of users dominates online dynamics. However, a direct comparison of news consumption on Facebook and Reddit shows higher segregation on Facebook.
Keywords: information spreading, echo chambers, social media, polarization
Abstract
Social media may limit the exposure to diverse perspectives and favor the formation of groups of like-minded users framing and reinforcing a shared narrative, that is, echo chambers. However, the interaction paradigms among users and feed algorithms greatly vary across social media platforms. This paper explores the key differences between the main social media platforms and how they are likely to influence information spreading and echo chambers’ formation. We perform a comparative analysis of more than 100 million pieces of content concerning several controversial topics (e.g., gun control, vaccination, abortion) from Gab, Facebook, Reddit, and Twitter. We quantify echo chambers over social media by two main ingredients: 1) homophily in the interaction networks and 2) bias in the information diffusion toward like-minded peers. Our results show that the aggregation of users in homophilic clusters dominate online interactions on Facebook and Twitter. We conclude the paper by directly comparing news consumption on Facebook and Reddit, finding higher segregation on Facebook.
Social media radically changed the mechanism by which we access information and form our opinions (1–5). We need to understand how people seek or avoid information and how those decisions affect their behavior (6), especially when the news cycle—dominated by the disintermediated diffusion of information—alters the way information is consumed and reported on. A recent study (7) limited to Twitter claimed that fake news travels faster than real news. However, a multitude of factors affects information spreading on social media platforms. Online polarization, for instance, may foster misinformation spreading (1, 8). Our attention span remains limited (9, 10), and feed algorithms might limit our selection process by suggesting contents similar to the ones we are usually exposed to (11–13). Furthermore, users show a tendency to favor information adhering to their beliefs and join groups formed around a shared narrative, that is, echo chambers (1, 14–18). We can broadly define echo chambers as environments in which the opinion, political leaning, or belief of users about a topic gets reinforced due to repeated interactions with peers or sources having similar tendencies and attitudes. Selective exposure (19) and confirmation bias (20) (i.e., the tendency to seek information adhering to preexisting opinions) may explain the emergence of echo chambers on social media (1, 17, 21, 22).
According to group polarization theory (23), an echo chamber can act as a mechanism to reinforce an existing opinion within a group and, as a result, move the entire group toward more extreme positions. Echo chambers have been shown to exist in various forms of online media such as blogs (24), forums (25), and social media sites (26–28). Some studies point out echo chambers as an emerging effect of human tendencies, such as selective exposure, contagion, and group polarization (13, 23, 29–31). However, recently, the effects and the very existence of echo chambers have been questioned (2, 27, 32). This issue is also fueled by the scarcity of comparative studies on social media, especially concerning news consumption (33). In this context, the debate around echo chambers is fundamental to understanding social media’s influence on information consumption and public opinion formation. In this paper, we explore the key differences between social media platforms and how they are likely to influence the formation of echo chambers or not. As recently shown in the case of selective exposure to news outlets, studies considering multiple platforms can offer a fresh view on long-debated problems (34). Different platforms offer different interaction paradigms to users, ranging from retweets and mentions on Twitter to likes and comments in groups on Facebook, thus triggering very different social dynamics (35). We introduce an operational definition of echo chambers to provide a common methodological ground to explore how different platforms influence their formation. In particular, we operationalize the two common elements that characterize echo chambers into observables that can be quantified and empirically measured, namely, 1) the inference of the user’s leaning for a specific topic (e.g., politics, vaccines) and 2) the structure of their social interactions on the platform. Then, we use these elements to assess echo chambers’ presence by looking at two different aspects: 1) homophily in interactions concerning a specific topic and 2) bias in information diffusion from like-minded sources. We focus our analysis on multiple platforms: Facebook, Twitter, Reddit, and Gab. These platforms present similar features and functionalities (e.g., they all allow social feedback actions such as likes or upvotes) and design (e.g., Gab is similar to Twitter) but also distinctive features (e.g., Reddit is structured in communities of interest called subreddits). Reddit is one of the most visited websites worldwide (https://www.alexa.com/siteinfo/reddit.com) and is organized as a forum to collect discussions on a wide range of topics, from politics to emotional support. Gab claims to be a social platform aimed at protecting freedom of speech. However, low moderation and regulation on content has resulted in widespread hate speech. For these reasons, it has been repeatedly suspended by its service provider, and its mobile app has been banned from both App and Play stores (36). Overall, we account for the interactions of more than 1 million active users on the four platforms, for a total of more than 100 million unique pieces of content, including posts and social interactions. Our analysis shows that platforms organized around social networks and news feed algorithms, such as Facebook and Twitter, favor the emergence of echo chambers.
We conclude the paper by directly comparing news consumption on Facebook and Reddit, finding higher segregation on Facebook than on Reddit.
Characterizing Echo Chambers in Social Media
Operational Definitions.
To explore the key differences between social media platforms and how they influence echo chambers’ formation, we need to operationalize a definition for them. First, we need to identify the attitude of users at a microlevel. On online social media, the individual leaning of a user i toward a specific topic, xi, can be inferred in different ways, via the content produced or the endorsement network among users (37). Concerning content, we can define the leaning as the attitude expressed by a piece of content toward a specific topic. This leaning can be explicit (e.g., arguments supporting a narrative) or implicit (e.g., framing and agenda setting). Let us consider a user i producing a number ai of contents, Ci={c1,c2,…,cai}, where ai is the activity of user i, and each content leaning is assigned a numeric value. Then the individual leaning of user i can be defined as the average of the leanings of produced contents,
Once individual leanings have been inferred, polarization can be defined as a state of the system such that the distribution of leanings, P(x), is concentrated in one or more clusters. A possible example is the case of a single cluster, distinguishable by a single, extreme peak in P(x). Another example is the typical case of topics characterized by positive versus negative stances, in which a bimodal distribution can describe polarization. For instance, if opinions are assumed to be embedded in a one-dimensional space (38), x∈[−1,+1] without loss of generality, as usual for controversial topics, then polarization is characterized by two well-separated peaks in P(x), for positive and negative opinions. In contrast, neutral ones are absent or underrepresented in the population. Note that polarization can happen independently from the structure or the very presence of social interactions. Homophily in social interactions can be quantified by representing interactions as a social network and then analyzing its structure concerning the opinions of the users (18, 39, 40). Social networks can be reconstructed in different ways from online social media, where links represent social relationships or interactions. Since we are interested in capturing the possible exchange of opinions between users, we assume links as the substrate over which information may flow. For instance, if user i follows user j on Twitter, user i can see tweets produced by user j, and there is a flow of information from node j to node i in the network. When the reconstructed network is directed, we assume the link direction points to potential influencers (opposite of information flow). Actions such as mentions or retweets may convey similar flows. In some cases, direct relations between users are not available in the data, so one needs to assume some proxy for social connections, for example, a link between two users if they comment on the same post on Facebook. Crucially, the two elements characterizing the presence of echo chambers, polarization and homophilic interactions, should be quantified independently.
Implementation on Social Media.
This section explains how we implement the operational definitions defined above on different social media. For each medium, we detail 1) how we quantify users’ leaning, and 2) how we reconstruct how the information spread.
Twitter.
We consider the set of tweets posted by user i that contain links to news outlets of known political leaning. Each news outlet is associated with a political leaning score ranging from extreme left to extreme right following the Materials and Methods classification. We infer the individual leaning of a user, i, xi∈[−1,+1], by averaging the news organizations’ scores linked by user i according to Eq. 1. We analyze three different datasets collected on Twitter related to controversial topics: gun control, Obamacare, and abortion. For each dataset, the social interaction network is reconstructed using the following relation so that there is a direct link from node i to node j if user i follows user j (i.e., the source). Henceforth, we focus on the dataset about abortion, and others are shown in SI Appendix.
Facebook.
We quantify the individual leaning of users considering endorsements in the form of likes to posts. Posts are produced by pages that are labeled in a certain number of categories, and, to each category, we assign a numerical value (e.g., Anti-Vax [+1] or Pro-Vax [–1]). Each like to a post (only one like per post is allowed) represents an endorsement for that content, which is assumed to be aligned with the leaning associated with the page. Thus, the user’s leaning is defined as the average of the content leanings of the posts liked by the user, according to Eq. 1.
We analyze three different datasets collected on Facebook regarding a specific topic of discussion: vaccines, science versus conspiracy, and news. The interaction network is defined by considering comments. In such an interaction network, two users are connected if they cocommented on at least one post. Henceforth, we focus on the dataset about vaccines and news, and others are shown in SI Appendix.
Reddit.
The individual leaning of users is quantified similarly to Twitter by considering the links to news organizations in the content produced by the users, submissions, and comments. We build the interaction network considering comments and submissions. There exists a direct link from node i to node j if user i comments on a submission or comment by user j (we assume that i reads the comment they are replying to, which is written by j).
We analyze three datasets collected on different subreddits: the_donald, Politics, and News. In the following, we focus on the dataset collected on the Politics and the News subreddits, and others are shown in SI Appendix.
Gab.
The political leaning xi of user i is computed by considering the set of contents posted by user i containing a link to news outlets of a known political leaning, similarly to Twitter and Reddit. To obtain the leaning xi of user i, we averaged the scores of each link posted by user i according to Eq. 1. The interaction network is reconstructed by exploiting the cocommenting relationships under posts in the same way as for Facebook. Given two users i and j, an undirected edge between i and j exists if and only if they comment under the same post.
Comparative Analysis
In the following, we perform a comparative analysis of four different social media. We select one dataset for each social media: Abortion (Twitter), Vaccines (Facebook), Politics (Reddit), and Gab as a whole. Results for other datasets for the same medium are qualitatively similar, as shown in SI Appendix. We first characterize echo chambers in the networks’ topology, and then look at their effects on information diffusion. Finally, we directly compare news consumption on Facebook and Reddit.
Polarization and Homophily in the Interaction Networks.
The network’s topology can reveal echo chambers, where users are surrounded by peers with similar leanings, and thus they get exposed, with a higher probability, to similar contents. In network terms, this translates into a node i with a given leaning xi more likely to be connected with nodes with a leaning close to xi (18). This concept can be quantified by defining, for each user i, the average leaning of their neighborhood, as xiN≡1ki→∑jAijxj, where Aij is the adjacency matrix of the interaction network, Aij=1 if there is a link from node i to node j, Aij=0 otherwise, and ki→=∑jAij is the out-degree of node i. Fig. 1 shows the correlation between the leaning of a user i and the leaning of their neighbors, xiN, for the four social media under consideration. The probability distributions P(x) (individual leaning) and PN(x) (average leaning of neighbors) are plotted on the x and y axes, respectively. All plots are color-coded contour maps, representing the number of users in the phase space (x,xN): The brighter the area in the plan, the larger the density of users in that area. The topics of vaccines and abortion, on Facebook and Twitter, respectively, show a strong correlation between the leaning of a user and the average leaning of their nearest neighbors. Similar behavior is found for different topics from the same social media platform; see SI Appendix. Conversely, Reddit and Gab show a different picture. The corresponding plots in Fig. 1 display a single bright area, indicating that users do not split into groups with opposite leaning but form a single community, biased to the left (Reddit) or the right (Gab). Similar results are found for different datasets on Reddit; see SI Appendix.
Fig. 1.
Joint distribution of the leaning of users x and the average leaning of their neighborhood xNN for different datasets. (A) Twitter, (B) Reddit, (C) Facebook, and (D) Gab. Colors represent the density of users: The lighter the color, the larger the number of users. Marginal distribution P(x) and PN(x) are plotted on the x and y axes, respectively. Facebook and Twitter present by homophilic clustering.
The presence of homophilic interactions can be confirmed by the community structure of the interaction networks. We detected communities by applying the Louvain algorithm (41), removing singleton communities with only one user. Then, we computed each community’s average leaning, determined as the average of individual leanings of its members. Fig. 2 shows the communities emerging for each social medium, arranged by increasing average leaning on the x axis (color-coded from blue to red), while the y axis reports the size of the community. On Facebook and Twitter, communities span the whole spectrum of possible leanings, but users with similar leanings form each community. Some communities are characterized by a robust average leaning, especially in the case of Facebook. These results are in accordance with the observation of homophilic interactions. Instead, communities on Reddit and Gab do not cover the whole spectrum, and all show similar average leaning. Furthermore, the almost total absence of communities with leaning very close to 0 confirms the polarized state of the systems.
Fig. 2.
Size and average leaning of communities detected in different datasets. A and C show the full spectrum of leanings related to the topics of abortions and vaccines with regard to communities in B and D, where the political leaning is less sparse.
Effects on Information Spreading.
Simple models of information spreading can gauge the presence of echo chambers: Users are expected to be more likely to exchange information with peers sharing a similar leaning (18, 42, 43). Classical epidemic models such as the susceptible–infected–recovered (SIR) model (44) have been used to study the diffusion of information, such as rumors or news (45–47). In the SIR model, each agent can be in any of three states: susceptible (unaware of the circulating information), infectious (aware and willing to spread it further), or recovered (knowledgeable but not ready to transmit it anymore). Susceptible (unaware) users may become infectious (aware) upon contact with infected neighbors, with a specific transmission probability β. Infectious users can spontaneously become recovered with probability ν. To measure the effects of the leaning of users on the diffusion of information, we run the SIR dynamics on the interaction networks, by starting the epidemic process with only one node i infected, and stopping it when no more infectious nodes are left.
The set of nodes in a recovered state at the end of the dynamics started with user i as a seed of infection, that is, those that become aware of the information initially propagated by user i, form the set of influence of user i, Ii (48). Thus, the set of influence of a user represents those individuals that can be reached by a piece of content sent by him/her, depending on the effective infection ratio β/ν. One can compute the average leaning of the set of influence of user i, μi, as
The quantity μi indicates how polarized the users are that can be reached by a message initially propagated by user i (18).
Fig. 3 shows the average leaning ⟨μ(x)⟩ of the influence sets reached by users with leaning x, for the different datasets under consideration. The recovery rate ν is fixed at 0.2 for every dataset. In contrast, the ratio between the infection rate β and average degree ⟨k⟩ depends on the specific dataset and is reported in the caption of each figure.
Fig. 3.
Average leaning ⟨μ(x)⟩ of the influence sets reached by users with leaning x, for different datasets under consideration. Size and color of each point represent the average size of the influence sets. The parameters of the SIR dynamics are set to (A) β=0.10⟨k⟩−1, (B) β=0.01⟨k⟩−1, (C) β=0.05⟨k⟩−1, and (D) β=0.05⟨k⟩−1, while ν is fixed at 0.2 for all simulations.
Again, one can observe a clear distinction between Facebook and Twitter, on one side, and Reddit and Gab on the other side. For the topics of vaccines and abortion, on Facebook and Twitter, respectively, users with a given leaning are much more likely to be reached by information propagated by users with similar leaning, that is, ⟨μ(x)⟩≈x. Similar behavior is found for different topics from the same social media platform; see SI Appendix. Conversely, Reddit and Gab show a different behavior: The average leaning of the set of influence, ⟨μ(x)⟩, does not depend on the leaning x. As expected, the average leaning in these media is not zero. Still, it assumes negative (positive) values in Reddit (Gab), indicating that the users of this platform are more likely to receive left (right)-leaning content.
These results indicate that information diffusion is biased toward individuals who share a similar leaning in some social media, namely Twitter and Facebook. In contrast, in others—Reddit and Gab in our analysis—this effect is absent. Such a latter configuration may depend upon two factors: 1) Gab and Reddit are not bursting the echo chamber effects, or 2) we are observing the dynamic inside a single echo chamber.
Our results are robust for different values of the effective infection ratio β/ν; see SI Appendix. Furthermore, Fig. 3 shows that the spreading capacity, represented by the average size of the influence sets (color-coded in Fig. 3), depends on the leaning of the users. On Twitter, proabortion users are more likely to reach larger audiences. The same is true for antivax users on Facebook, left-leaning users on Reddit, and right-leaning users on Gab (in this dataset, left-leaning users are almost absent).
News Consumption on Facebook and Reddit.
The striking differences observed across social media, in terms of homophily in the interaction networks and information diffusion, could be attributed to the different topics taken into account. For this reason, here we compare Facebook and Reddit on a common topic, news consumption. Facebook and Reddit are particularly apt to a cross-comparison since they share the definition of individual leaning (computed by using the classification provided by mediabiasfactcheck.org; see Materials and Methods for further details) and the rationale in creating connections among users that is based on an interaction network. Fig. 4 shows a direct comparison of news consumption on Facebook and Reddit along the metrics used in the previous sections to quantify the presence of echo chambers: 1) the correlation between the leaning of a user x and the average leaning of neighbors xN (Fig. 4, Top), 2) the average leaning of communities detected in the networks (Fig. 4, Middle), and 3) the average leaning ⟨μ(x)⟩ of the influence sets reached by users with leaning x, by running SIR dynamics (Fig. 4, Bottom). One can see that all three measures confirm the picture obtained for other datasets: On Facebook, we observe a clear separation among users depending on their leaning, while, on Reddit, users’ leanings are more homogeneous and show only one peak. In the latter social media, even users displaying a more extreme leaning (noticeable in the marginal histogram of Fig. 4 B, Top) tend to interact with the majority. Moreover, on Facebook, the seed user’s leaning affects who the final recipients of the information are, therefore indicating the presence of echo chambers. On Reddit, this effect is absent.
Fig. 4.
Direct comparison of news consumption on (A) Facebook and (B) Reddit. Joint distribution of the leaning of users x and the average leaning of their nearest neighbor xN (Top), size and average leaning of communities detected in the interaction networks (Middle), and average leaning ⟨μ(x)⟩ of the influence sets reached by users with leaning x, by running SIR dynamics (Bottom) with parameters β=0.05⟨k⟩ for A, β=0.006⟨k⟩ for B, and ν=0.2 for both. Facebook presents a highly segregated structure with regard to Reddit.
Conclusions
Social media platforms provide direct access to an unprecedented amount of content. Platforms originally designed for user entertainment changed the way information spread. Indeed, feed algorithms mediate and influence the content promotion accounting for users’ preferences and attitudes. Such a paradigm shift affected the construction of social perceptions and the framing of narratives; it may influence policy making, political communication, and the evolution of public debate, especially on polarizing topics. Indeed, users online tend to prefer information adhering to their worldviews, ignore dissenting information, and form polarized groups around shared narratives. Furthermore, when polarization is high, misinformation quickly proliferates.
Some argued that the veracity of the information might be used as a determinant for information spreading patterns. However, selective exposure dominates content consumption on social media, and different platforms may trigger very different dynamics. In this paper, we explore the key differences between the leading social media platforms and how they are likely to influence the formation of echo chambers and information spreading. To assess the different dynamics, we perform a comparative analysis on more than 100 million pieces of content concerning controversial topics (e.g., gun control, vaccination, abortion) from Gab, Facebook, Reddit, and Twitter. The analysis focuses on two main dimensions: 1) homophily in the interaction networks and 2) bias in the information diffusion toward like-minded peers. Our results show that the aggregation in homophilic clusters of users dominates online dynamics. However, a direct comparison of news consumption on Facebook and Reddit shows higher segregation on Facebook. Furthermore, we find significant differences across platforms in terms of homophilic patterns in the network structure and biases in the information diffusion toward like-minded users. A clear-cut distinction emerges between social media having a feed algorithm tweakable by the users (e.g., Reddit) and social media that don’t provide such an option (e.g., Facebook and Twitter). Our work provides important insights into the understanding of social dynamics and information consumption on social media. The next envisioned step addresses the temporal dimension of echo chambers, to understand better how different social feedback mechanisms, specific to distinct platforms, can impact their formation.
Materials and Methods
Here we provide details about the labeling of news outlets and the datasets considered.
Labeling of Media Sources.
The labeling of news outlets is based on the information reported by Media Bias/Fact Check (MBFC) (https://mediabiasfactcheck.com), an independent fact-checking organization that rates news outlets on the basis of the reliability and of the political bias of the contents they produce and share. The labeling provided by MBFC, retrieved in June 2019, ranges from Extreme Left to Extreme Right for political bias. The total number of media outlets for which we have a political label is 2,190. A detailed description of the source labeling process and political bias distribution can be found in SI Appendix.
Data Availability.
For what concerns Gab, all data are available on the Pushshift public repository (https://pushshift.io/what-is-pushshift-io/) at this link: https://files.pushshift.io/gab/. Reddit data are available on the Pushshift public repository at this link: https://search.pushshift.io/reddit/. For what concerns Facebook and Twitter, we provide data according to their Terms of Services on the corresponding author institutional page at this link: https://walterquattrociocchi.site.uniroma1.it/ricerca. For news outlet classification, we used data from MBFC (https://mediabiasfactcheck.com), an independent fact-checking organization. Anonymized data have been deposited in Open Science Framework (10.17605/OSF.IO/X92BR) (49). For further details about data, refer to the following section.
Empirical Datasets.
Table 1 reports summary statistics of the datasets under consideration. Due to the structural differences among platforms, each dataset has different features. For Twitter, we used tweets regarding three topics collected by Garimella et al. (16), namely Gun control, Obamacare, and abortion. Tweets linking to a news source with a known bias are classified based on MBFC. Facebook datasets were created by using Facebook Graph API and were previously explored in ref. 50 (Science and Conspiracy), (51) (Vaccines) and (11) (News). For the two datasets Science and Conspiracy and Vaccines, data were labeled in a binary way, respectively, provaccines/antivaccines and proscience/conspiracy, based on the page where they were posted. Posts in the dataset News were instead classified based on MBFC labeling. Reddit datasets have been obtained by downloading comments and submissions posted in the subreddit Politics, the_donald, and News and labeled according to the classification obtained from MBFC. The Gab dataset has been collected from https://files.pushshift.io/gab and contains posts, replies, and quotations. Posts were labeled according to MBFC classification. Further details can be found in SI Appendix.
Table 1.
Dataset details
Media | Dataset | T0 | T | C | N | nc |
---|---|---|---|---|---|---|
Gun control | June 2016 | 14 d | 19 million | 3,963 | 0.93 | |
Obamacare | June 2016 | 7 d | 39 million | 8,703 | 0.90 | |
Abortion | June 2016 | 7 d | 34 million | 7,401 | 0.95 | |
Sci/Cons | January 2010 | 5 y | 75,172 | 183,378 | 1.00 | |
Vaccines | January 2010 | 7 y | 94,776 | 221,758 | 1.00 | |
News | January 2010 | 6 y | 15,540 | 38,663 | 1.00 | |
Politics | January 2017 | 1 y | 353,864 | 240,455 | 0.15 | |
the_donald | January 2017 | 1 y | 1.234 million | 138,617 | 0.16 | |
News | January 2017 | 1 y | 723,235 | 179,549 | 0.20 | |
Gab | Gab | November 2017 | 1 y | 13 million | 165,162 | 0.13 |
Supplementary Material
Supplementary File
Acknowledgments
We thank Fabiana Zollo and Antonio Scala for precious insights for the development of this paper. We are grateful to Geronimo Stilton and the Hypnotoad for inspiring the data analysis and result interpretation.
Footnotes
The authors declare no competing interest.
This article is a PNAS Direct Submission.
References
- 1.Del Vicario M., et al. , The spreading of misinformation online. Proc. Natl. Acad. Sci. U.S.A. 113, 554–559 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Dubois E., Blank G., The echo chamber is overstated: The moderating effect of political interest and diverse media. Inf. Commun. Soc. 21, 729–745 (2018). [Google Scholar]
- 3.Bode L., Political news in the news feed: Learning politics from social media. Mass Commun. Soc. 19, 24–48 (2016). [Google Scholar]
- 4.Newman N., Fletcher R., Kalogeropoulos A., Nielsen R., “Reuters Institute Digital News Report 2019” (Rep. 2019, Reuters Institute for the Study of Journalism, 2019).
- 5.Flaxman S., Goel S., Rao J. M., Filter bubbles, echo chambers, and online news consumption. Publ. Opin. Q. 80, 298–320 (2016). [Google Scholar]
- 6.Sharot T., Sunstein C. R., How people decide what they want to know. Nat. Hum. Behav. 4, 14–19 (2020). [DOI] [PubMed] [Google Scholar]
- 7.Vosoughi S., Roy D., Aral S., The spread of true and false news online. Science 359, 1146–1151 (2018). [DOI] [PubMed] [Google Scholar]
- 8.Vicario M. D., Quattrociocchi W., Scala A., Zollo F., Polarization and fake news: Early warning of potential misinformation targets. ACM Trans. Web 13, 1–22 (2019). [Google Scholar]
- 9.Baronchelli A., The emergence of consensus: A primer. R. Soc. Open Sci. 5, 172189 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Cinelli M., et al. , Selective exposure shapes the Facebook news diet. PloS One 15, e0229129 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Schmidt A. L., et al. , Anatomy of news consumption on Facebook. Proc. Natl. Acad. Sci. U.S.A. 114, 3035–3039 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Cinelli M., et al. , The COVID-19 social media infodemic. Sci. Rep. 10, 16598 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Bakshy E., Messing S., Adamic L. A., Exposure to ideologically diverse news and opinion on Facebook. Science 348, 1130–1132 (2015). [DOI] [PubMed] [Google Scholar]
- 14.Jamieson K. H., Cappella J. N., Echo Chamber: Rush Limbaugh and the Conservative Media Establishment (Oxford University Press, 2008). [Google Scholar]
- 15.Garrett R. K., Echo chambers online?: Politically motivated selective exposure among Internet news users. J. Comput. Mediated Commun. 14, 265–285 (2009). [Google Scholar]
- 16.Garimella K., De Francisci Morales G., Gionis A., Mathioudakis M., “Political discourse on social media: Echo chambers, gatekeepers, and the price of bipartisanship” in Proceedings of the 2018 World Wide Web Conference, WWW ’18, Champin P. A., Gandon F., Médini L., Eds. (International World Wide Web Conferences Steering Committee, Geneva, Switzerland, 2018), pp. 913–922. [Google Scholar]
- 17.Garimella K., De Francisci Morales G., Gionis A., Mathioudakis M., “The effect of collective attention on controversial debates on social media” in WebSci ’17: 9th International ACM Web Science Conference (Association for Computing Machinery, New York, NY, 2017), pp. 43–52. [Google Scholar]
- 18.Cota W., Ferreira S. C., Pastor-Satorras R., Starnini M., Quantifying echo chamber effects in information spreading over political communication networks. EPJ Data Sci. 8, 35 (2019). [Google Scholar]
- 19.Klapper J. T., The Effects of Mass Communication (Free Press, 1960). [Google Scholar]
- 20.Nickerson R. S., Confirmation bias: A ubiquitous phenomenon in many guises. Rev. Gen. Psychol. 2, 175–220 (1998). [Google Scholar]
- 21.Del Vicario M., et al. , Echo chambers: Emotional contagion and group polarization on Facebook. Sci. Rep. 6, 37825 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Bessi A., et al. , Science vs conspiracy: Collective narratives in the age of misinformation. PloS One 10, e0118093 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Sunstein C. R., The law of group polarization. J. Polit. Philos. 10, 175–195 (2002). [Google Scholar]
- 24.Gilbert E., Bergstrom T., Karahalios K., “Blogs are echo chambers: Blogs are echo chambers” in 42nd Hawaii International Conference on System Sciences, (IEEE Computer Society, 2009), pp. 1–10. [Google Scholar]
- 25.Edwards A., (How) do participants in online discussion forums create ‘echo chambers’?: The inclusion and exclusion of dissenting voices in an online forum about climate change. J. Argumentation Context 2, 127–150 (2013). [Google Scholar]
- 26.Grömping M., ‘Echo chambers’: Partisan Facebook groups during the 2014 Thai election. Asia Pac. Media Educat. 24, 39–59 (2014). [Google Scholar]
- 27.Barberá P., Jost J. T., Nagler J., Tucker J. A., Bonneau R., Tweeting from left to right: Is online political communication more than an echo chamber? Psychol. Sci. 26, 1531–1542 (2015). [DOI] [PubMed] [Google Scholar]
- 28.Quattrociocchi W., Scala A., Sunstein C. R., Echo chambers on Facebook. SSRN [Preprint] (2016). 10.2139/ssrn.2795110 (Accessed 16 February 2021). [DOI]
- 29.Himelboim I., McCreery S., Smith M., Birds of a feather tweet together: Integrating network and content analyses to examine cross-ideology exposure on Twitter. J. Computer-Mediated Commun. 18, 154–174 (2013). [Google Scholar]
- 30.Nikolov D., Oliveira D. F., Flammini A., Menczer F., Measuring online social bubbles. PeerJ Comput. Sci. 1, e38 (2015). [Google Scholar]
- 31.Baumann F., Lorenz-Spreen P., Sokolov I. M., Starnini M., Modeling echo chambers and polarization dynamics in social networks. Phys. Rev. Lett. 124, 048301 (2020). [DOI] [PubMed] [Google Scholar]
- 32.Bruns A., Echo chamber? What echo chamber? Reviewing the evidence. QUT ePrints [Preprint] (2017). https://eprints.qut.edu.au/113937/ (Accessed 16 February 2021). [Google Scholar]
- 33.Barberá P., “Social media, echo chambers, and political polarization” in Social Media and Democracy: The State of the Field, Prospects for Reform, Persily N., Tucker J., Eds. (SSRC Anxieties of Democracy, Cambridge University Press, Cambridge, UK, 2020), pp. 34–55. [Google Scholar]
- 34.Gollwitzer A., et al. , Partisan differences in physical distancing are linked to health outcomes during the COVID-19 pandemic. Nat. Hum. Behav. 4, 1186–1197 (2020). [DOI] [PubMed] [Google Scholar]
- 35.Golovchenko Y., Buntain C., Eady G., Brown M. A., Tucker J. A., Cross-platform state propaganda: Russian trolls on Twitter and YouTube during the 2016 US presidential election. Int. J. Press/Politics 25, 357–389 (2020). [Google Scholar]
- 36.Zannettou S., et al. , “What is gab: A bastion of free speech or an alt-right echo chamber” in Companion Proceedings of the the Web Conference 2018 (International World Wide Web Conferences Steering Committee, Geneva, Switzerland, 2018), pp. 1007–1014. [Google Scholar]
- 37.Garimella K., De Francisci Morales G., Gionis A., Mathioudakis M., Quantifying controversy on social media. TSC ACM Transactions Soc. Comput. 1, 3 (2018). [Google Scholar]
- 38.DeGroot M. H., Reaching a consensus. J. Am. Stat. Assoc. 69, 118–121 (1974). [Google Scholar]
- 39.Kossinets G., Watts D. J., Origins of homophily in an evolving social network. Am. J. Sociol. 115, 405–450 (2009). [Google Scholar]
- 40.Bessi A., et al. , Homophily and polarization in the age of misinformation. Eur. Phys. J. Spec. Top. 225, 2047–2059 (2016). [Google Scholar]
- 41.Blondel V. D., Guillaume J. L., Lambiotte R., Lefebvre E., Fast unfolding of communities in large networks. J. Stat. Mech. Theor. Exp. 2008, P10008 (2008). [Google Scholar]
- 42.Garimella K., De Francisci Morales G., Gionis A., Mathioudakis M., “Reducing controversy by connecting opposing views” in WSDM ’17: 10th ACM International Conference on Web Search and Data Mining (Association for Computing Machinery, New York, NY, 2017), pp. 81–90. [Google Scholar]
- 43.Garimella K., De Francisci Morales G., Gionis A., Mathioudakis M., “Quantifying controversy in social media” in WSDM ’16: 9th ACM International Conference on Web Search and Data Mining (Association for Computing Machinery, New York, NY, 2016), pp. 33–42. [Google Scholar]
- 44.Anderson R. M., May R. M., Infectious Diseases in Humans (Oxford University Press, Oxford, United Kingdom, 1992). [Google Scholar]
- 45.Jalili M., Perc M., Information cascades in complex networks. J. Complex. Netw. 5, 665–693 (2017). [Google Scholar]
- 46.Zhao L., Cui H., Qiu X., Wang X., Wang J., Sir rumor spreading model in the new media age. Phys. Stat. Mech. Appl. 392, 995–1003 (2013). [Google Scholar]
- 47.Granell C., Gómez S., Arenas A., Dynamical interplay between awareness and epidemic spreading in multiplex networks. Phys. Rev. Lett. 111, 128701 (2013). [DOI] [PubMed] [Google Scholar]
- 48.Holme P., Network reachability of real-world contact sequences. Phys. Rev. E 71, 046119 (2005). [DOI] [PubMed] [Google Scholar]
- 49.Cinelli M., De Francisci Morales G., Galeazzi A., Quattrociocchi W., Starnini M., The echo chamber effect on social media. Open Science Framework. https://osf.io/x92br/?view_only=cdcc87d30e404e01b6ae316db15e3375. Deposited 22 January 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Bessi A., et al. , Users polarization on Facebook and YouTube. PloS One 11, e0159641 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Schmidt A. L., Zollo F., Scala A., Betsch C., Quattrociocchi W., Polarization of the vaccination debate on Facebook. Vaccine 36, 3606–3612 (2018). [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Supplementary File
Data Availability Statement
For what concerns Gab, all data are available on the Pushshift public repository (https://pushshift.io/what-is-pushshift-io/) at this link: https://files.pushshift.io/gab/. Reddit data are available on the Pushshift public repository at this link: https://search.pushshift.io/reddit/. For what concerns Facebook and Twitter, we provide data according to their Terms of Services on the corresponding author institutional page at this link: https://walterquattrociocchi.site.uniroma1.it/ricerca. For news outlet classification, we used data from MBFC (https://mediabiasfactcheck.com), an independent fact-checking organization. Anonymized data have been deposited in Open Science Framework (10.17605/OSF.IO/X92BR) (49). For further details about data, refer to the following section.