Social Media and Elections
Social Media and Elections
Social Media and Elections
Abstract
We study how social media affects election outcomes in the United States. We use variation
in the number of Twitter users across counties induced by early adopters at the 2007 South
by Southwest (SXSW) festival, a key event in Twitter’s rise to popularity. We show that this
variation is unrelated to observable county characteristics and electoral outcomes before the
launch of Twitter. Our results indicate that Twitter lowered the Republican vote share in
the 2016 and 2020 presidential elections, but had limited effects on Congressional elections
and previous presidential elections. Evidence from survey data, primary elections, and a
text analysis of millions of tweets suggests that Twitter’s relatively liberal content may have
persuaded voters with moderate views to vote against Donald Trump.
∗
We are grateful to Pablo Barbera, Levi Boxell, Matt Gentzkow, Ro’ee Levy, Alexey Makarin, Jesse Shapiro, James
Snyder, Ekaterina Zhuravskaya, as well as seminar participants at Princeton University, Imperial College, Warwick
University, Bocconi University, University of California San Diego, Insper, University of São Paulo, Inter-American
Development Bank, IPEG/Pompeu Fabra, University of Zurich, OPESS, Università di Bergamo, the 2021 ASSA
Conference, and the NBER Political Economy Spring 2021 Meeting for their helpful comments.
†
Princeton University, Department of Economics and SPIA, and NBER, fujiwara@princeton.edu.
‡
National University of Singapore, Business School, kmueller@nus.edu.sg.
§
Università Bocconi, Department of Economics and IGIER, and PERICLES, carlo.schwarz@unibocconi.it.
1 Introduction
Does social media affect election outcomes? A popular narrative holds that Twitter played a
decisive role in both recent American presidential elections and the United Kingdom’s “Brexit”
referendum. Many see this as part of social media’s broader influence on political polarization
and the re-emergence of populist politicians in many countries. The U.S. Federal Election
Commissioner, for example, has argued that Facebook “has no idea how seriously it is hurting
democracy” (NPR, 2020a).1
An alternative view suggests that social media platforms are biased against conservatives
(e.g., NPR, 2020b; Wall Street Journal, 2020) and that its younger, relatively left-leaning user
base is unlikely to tilt elections towards right-wing politicians (e.g., Boxell et al., 2017, 2018).
However, there is limited evidence that can be used to evaluate these contrasting (causal)
claims.
This paper focuses on the effects of Twitter, a platform used by almost a quarter of
American adults. We estimate how a county’s number of Twitter users affects election results
by exploiting a persistent network effect sparked by early Twitter adoption, building on
Müller and Schwarz (2019).2 Although it was launched in March 2006, Twitter’s popularity
increased rapidly after its advertising campaign at the South by Southwest festival (SXSW)
in March 2007. The SXSW festival was also key for Twitter’s geographical diffusion: counties
with more SXSW followers who joined during the 2007 festival saw disproportionately higher
growth of Twitter adoption compared to counties with SXSW followers who already joined
before the festival. Consistent with path dependence in technology adoption, this difference
in Twitter use across counties persists.
Our identification strategy leverages the 2007 SXSW festival as a shock to early Twitter
adoption that is uncorrelated with pre-existing election results. Conditional on geographic
controls and previous interest in the SXSW Twitter account, a county’s number of SXSW
followers who joined in March 2007 is essentially uncorrelated with a host of county charac-
teristics. It is also unrelated to election outcomes before Twitter’s launch (going back as far
as 1924) and during the period it had fewer users (between 2006 and 2012). However, the
number of SXSW followers who joined in March 2007 is correlated with Twitter usage in
2016, and has predictive power for the 2016 and 2020 presidential election results.
We estimate that a 10% increase in a county’s number of Twitter users lowered the vote
share of Republican presidential candidate Donald Trump by 0.2 percentage points (p.p.) in
1
See, for example, The New Yorker (2016); New York Times (2017); Allcott and Gentzkow (2017); The
Guardian (2018); UK Parliament (2019).
2
Enikolopov et al. (2020) use a similar empirical strategy based on spatial variation in early adopters of
the social media network VK in Russia to study its effects on protests.
1
both the 2016 and 2020 presidential elections. The implied persuasion rates are 8.6% and
9.4%, respectively. These estimates are smaller than the estimated pro-Republican effect of
Fox News (DellaVigna and Kaplan, 2007; Martin and Yurukoglu, 2017), the pro-Democrat
effect of the Washington Post (Gerber et al., 2009), or the effect of get-out-the-vote canvassing
on turnout (Gerber and Green, 2000), but larger than the effect of an independent anti-Putin
Russian TV channel on vote shares (Enikolopov et al., 2011) or the effect TV rollout on
turnout (Gentzkow, 2006).
For presidential elections before 2016, we find effects that are small and statistically
indistinguishable from zero. The same holds true for House and Senate races, including the
2016 and 2020 elections. Twitter adoption thus lowered Trump’s vote share but did not do so
for Republican candidates in congressional races in the same election. Together with other
“placebo tests,” this pattern bolsters confidence that our estimates are capturing the effect of
Twitter, which contains more content on presidential than congressional candidates.
An earlier draft of this paper, using only data up to the 2018 election, was posted online
on October 2020. When updating it to include November 2020 election results, we made
no revisions to the research design and regression specifications. In other words, the sample
selection, choice of controls, and variable definitions in the regressions we report were all
decided before the 2020 election results became available. Hence, our 2020 results can be
interpreted as a “pre-registered” research design.3
To shed light on the mechanisms behind these results, we estimate Twitter’s effect on
vote choices reported in the Cooperative Congressional Election Survey (CCES), primary
presidential candidates’ approval in the Gallup Daily Tracker, and county-level results in the
2016 and 2020 presidential primaries. Further, we explore data on the partisanship of political
content on Twitter.
These exercises yield three findings. First, the CCES results indicate that Twitter’s
effect is driven by independents and moderates switching their votes towards the Democratic
candidate (Hillary Clinton). This is consistent with Bayesian updating, since moderates
presumably have weaker priors and are thus more likely to be persuaded.
Second, we find that Twitter also lowered Trump’s vote share during the 2016 primaries,
a finding we confirm using individual-level Gallup candidate approval ratings. We find that
Twitter decreased Trump’s approval ratings and increased Clinton’s with only small effects
on relatively more moderate Republican candidates.4
3
The “pre-registration” document is the October 2020 draft, which is available and “time-stamped” at
Social Science Research Network (SSRN):
https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3719998.
4
We also estimate effects for the 2016 and 2020 Democratic primaries, detecting a (positive) effect for
Bernie Sanders in 2020.
2
Third, we document that political content on Twitter has a pro-Democratic slant. We
classify the slant of tweets based on two complementary approaches: one based on the network
users follow and one using the text of tweets in a machine learning approach in the spirit
of Gentzkow and Shapiro (2010). We apply these methods to the over 460 million tweets
mentioning the presidential candidates in the 2012, 2016, and 2020 elections. We find that
the number and attention (proxied by “likes”) of tweets mentioning Trump was substantially
larger than that of those mentioning Clinton and Joe Biden. Moreover, tweets about Trump
in 2016 and 2020 70% more likely to have Democratic rather than Republican slant. Overall,
our results are consistent with Twitter and its relatively pro-Democratic content persuading
voters with moderate views to not vote for Trump without inducing a more general negative
effect on Republicans.5
Our work contributes to the literature on the impact of media on political outcomes.
Expansions of traditional media such as newspapers, radio, broadcast television, and cable
news have been associated with changes in voter turnout, polarization, and electoral outcomes.6
While a set of papers studies the effect of overall internet access, the effects of social media
per se received less attention.7
A nascent literature studies the political effects of social media on protest participation
(Howard et al., 2011; Enikolopov et al., 2020; Acemoglu et al., 2017; Fergusson and Molina,
2021), xenophobia (Müller and Schwarz, 2020; Müller and Schwarz, 2019; Bursztyn et al., 2019)
and mental health (Braghieri et al., 2021).8 Additionally, a burgeoning field of experimental
research focuses on social media. Bond et al. (2012) and Jones et al. (2017) provide evidence
that online messages on social networks affect voter turnout. Allcott et al. (2020) and
Mosquera et al. (2020) find that individuals who deactivate Facebook react along many
dimensions, including some measures of political polarization. Levy (2021) studies the effect
of randomly assigning Facebook users subscriptions to conservative or liberal media outlets.
Bail et al. (2018) estimate the effect of paying Twitter users to follow a bot with messages
of the opposing political ideology. Most related to our paper is recent unpublished work by
Rotesi (2019), who finds social media negatively affected the Democratic vote share in the
5
Note also that Trump’s platform and impact on social media content was likely distinct from previous
prominent Republican candidates (Bursztyn et al., 2020).
6
See, for example, Gentzkow (2006); Huber and Arceneaux (2007); DellaVigna and Kaplan (2007); Gerber
et al. (2009, 2011); Gentzkow et al. (2011); Enikolopov et al. (2011); Campante and Hojman (2013); DellaVigna
et al. (2014); Larcinese and Miner (2017); Martin and Yurukoglu (2017); Spenkuch and Toniatti (2018); Chen
and Yang (2019).
7
There is evidence that broadband internet (Falck et al., 2014; Gavazza et al., 2019; Campante et al., 2017;
Lelkes et al., 2017) and mobile internet (Manacorda and Tesei, 2016; Guriev et al., 2020) exert political effects.
8
For reviews, see DellaVigna and Gentzkow (2010), Napoli (2014), Strömberg (2015), Enikolopov and
Petrova (2015), and DellaVigna and Ferrara (2015) and in particular Zhuravskaya et al. (2020) for the case of
social media.
3
2008-2016 presidential elections using variation in Twitter adoption resulting from transfers
of NBA players with Twitter accounts.9
Existing research thus provides an incomplete picture. On one hand, social media has
been painted as a key force behind political change and experimental studies indeed suggest
that social media affects individuals’ self-reported political beliefs. On the other hand, it
remains unclear whether social media can indeed persuade voters and affect elections results
at a larger scale. Our paper sheds light on this question by focusing on how Twitter affects
federal elections in the United States.
4
Much of the recent public discussion about the role of social media platforms has been
shaped by controversies, including the consulting firm Cambridge Analytica’s involvement
in multiple political campaigns (e.g., The Guardian, 2018); the Russian Internet Research
Agency’s efforts to support Trump’s election campaign (e.g., New York Times, 2017); and the
role of widely shared false information (“fake news”) in both the 2016 U.S. elections (e.g.,
Allcott and Gentzkow, 2017) and the Brexit referendum in the United Kingdom (e.g., UK
Parliament, 2019). Both Hillary Clinton and Donald Trump have argued that these factors
were instrumental in the 2016 election outcome, as has former president Barack Obama
(The New Yorker, 2016). As Brad Parscale, Trump’s digital media director in 2016, put
it: “Facebook and Twitter were the reason we won this thing. Twitter for Mr. Trump.
And Facebook for fundraising” (Wired, 2016). In Appendix Figure A.3, we document that
discussions of social media have become increasingly frequent in major American news outlets.
Mentions of Twitter in particular spiked with the 2016 presidential election when compared
to 2012 levels.
Despite its prominence in the political discourse, the empirical relevance of social media
for electoral outcomes is largely unknown, and some have suggested that concerns about its
effects may be overblown. As one example, in the 2016 presidential election, Trump received
fewer votes from demographic groups with higher propensities to use social media or the
internet more broadly (The Hill, 2016; Boxell et al., 2017, 2018). Indeed, Trump’s broadest
support came from older white voters without college education in rural areas, who are among
the least likely people to use social media actively (Hargittai, 2015; Pew Research Center,
2015, 2018). These patterns seem difficult to square with the idea that online channels were
an important driver of the 2016 presidential election result, although such observations also
do not rule this out.
Further, most social media users—particularly on Twitter—appear to be disproportion-
ately left-leaning. This is not suprising given that most Twitter users are relatively younger,
educated, and urban. While there appears to be a cluster of right-wing networks, Pew
Research Center (2019d) estimates that, in 2018, 60% of Twitter users identified as Democrat
and only 35% as Republican. Among Democrats, those on Twitter are considerably more
liberal and focus less on finding common ground with Republicans (Pew Research Center,
2020). In 2019, 26% of American Twitter users followed Obama and 19% followed Trump (Pew
Research Center, 2019a). Survey evidence suggests that 80% of Twitter content is produced by
people who strongly disapprove of Trump (Pew Research Center, 2019b). “Liberals” are also
more likely to get political news on Twitter or Facebook and follow more media and political
accounts compared to “conservatives” (Pew Research Center, 2014; Eady et al., 2019). Twitter
and Reddit, which are often said to be pro-Trump factors, were considerably more popular
5
among Clinton supporters before the 2016 election (Hargittai, 2015). Although social media
allows users to partially select which content they see, Twitter content disproportionately
leans toward the Democratic party.
We provide additional evidence for the composition of political content on Twitter by
analyzing the Twitter reach of Democratic and Republican politicians. We collected data on
the Twitter accounts of all Senators and House Representatives from the 110th to the 115th
Congresses (2007-2019). In Figure 1, we plot the average number of tweets and followers that
members of each party have on Twitter, as well as the average number of retweets and “likes”
their tweets receive. The patterns here again clearly indicate that Democratic politicians are
more active on Twitter and have larger follower bases than their Republican counterparts.
Tweets by Democrats also receive, on average, three times the number of “likes.”11
3 Data
The main analysis is based on a county-level dataset on election outcomes, political opinions,
and Twitter use. It covers 3,065 counties in 48 states (we exclude Alaska and Hawaii) and
the District of Columbia (except in congressional elections). County-level election results
are from Dave Leip’s Atlas of U.S. Presidential Elections and the MIT Election Lab. We
complement our analysis with individual-level survey data on approval ratings from the
Gallup Daily Tracker and voting data from the Cooperative Congressional Election Study
(CCES). Our measure of Twitter usage is derived from an archive of 475 million geo-located
tweets compiled by Kinder-Kurlanda et al. (2017). We combine this with newly collected
data on Twitter’s early adopters at the 2007 SXSW festival; data on the Twitter activity of
U.S. Congress members; and a large corpus of tweets related to the 2012, 2016, and 2020
presidential elections. Additional county characteristics were obtained from the U.S. Census,
the U.S. Religious Census, the American Community Survey (ACS), and the Bureau of Labor
Statistics (BLS). We describe the individual data sources in more detail below. Appendix
Table A.2 provides additional details and summary statistics.
6
vote shares. Figure 2 visualizes the Republican party’s vote share in the 2016 presidential
elections.12
7
decisions by users (e.g. tagging their holiday location) and prevents users from tagging wrong
locations (either intentionally or by accident).
The individual tweets from this dataset are already assigned to counties. Additionally,
we collected the underlying user profiles for each tweet in the database. This allows us to
construct a user-based measure by assigning users to the county from which they tweet most
frequently. The resulting measure, which we use throughout the paper, is a proxy for the
number of Twitter users per county, based on 3.7 million individual users (around 7% of the
Twitter population in 2015). Figure 3a plots the number of Twitter users per capita across
counties. Each user profile further provides us with a short biography and the date that each
user joined Twitter. We use the join dates to construct a time-varying proxy of Twitter usage
based on how many of the Twitter users had opened an account at each point in time.
The great advantage of this dataset is that it allows us to provide individual-level
evidence for the adoption of Twitter in a county following the SXSW festival and further
compare user profiles in different counties. The drawback of using geo-located Twitter data is
that only a sub-sample of tweets is geo-located. To overcome concerns of measurement error
in our Twitter measure, we validate the data in two ways. First, Appendix Figure A.1a shows
that our Twitter usage measure’s evolution closely tracks the number of daily Twitter users
from Statista (2019), which were directly obtained from the platform. Secondly, our measure
of county-level Twitter usage also strongly correlates with the number of Twitter users in a
county based on the GfK Media Survey (see Figure A.1b). Lastly, note that measurement
error is less of a concern in our setting because we largely rely on 2SLS estimation throughout
the paper. We return to the discussion of measurement error in Section 5.
Twitter Data for the South by Southwest Festival. We collected data for our instru-
ment for Twitter usage, based on early adoption during the SXSW festival, through the
Twitter API. More specifically, we scraped the account data for 658,240 users who followed
the Twitter account of SXSW Conference & Festivals (@SXSW) at the time of collection
(January 2019). We assign these users to counties based on the location people report in their
user profile.16
A user profile contains the month and year that they joined Twitter, which allows us to
determine the number of SXSW followers in each county that joined Twitter in a particular
month. The two key variables in our analysis are: i) the number of SXSW followers that
joined Twitter in the month of March 2007 and ii) the number of SXSW followers that joined
Twitter during 2006 (the year the platform was launched). We refer to (ii) as the number of
SXSW followers who joined before the March 2007 festival. We also scraped the follower lists
16
Of the 44,625 SXSW followers who joined between 2006 and 2008, we are able to geo-code 25,830 (58%).
8
of SXSW followers who joined in March 2007, which allows us to investigate the connections
of Twitter users to the SXSW festival. Further, we additionally collected tweets mentioning
the festival, based on the term “SXSW,” as well as a proxy for overall Twitter activity based
on the 100 common English words.17 We use these measures to document the SXSW festival’s
impact on local Twitter adoption.18
Data on Political Twitter Activity. We scraped the tweets and user profiles and followers
of the 901 Senators and House Representatives from the 110th to 115th (2007-2019) Congress
who have Twitter accounts. This includes 424 Democrats and 465 Republicans.19 In total,
the data contain 4,300,579 tweets, which we use to analyze the Twitter reach of Democratic
and Republican Congress members. Appendix Table A.1 lists the 20 Congress members with
the most Twitter followers.
We complement this dataset with election-related tweets to shed light on the overall
partisan slant of Twitter activity during the 2012, 2016, and 2020 elections. For each election,
we obtained the universe of tweets mentioning the last name of a Democrat and Republican
presidential candidates.20
To determine the likely political affiliation of Twitter users, we create two measures of
political slant. The first measure is based on the political accounts a user is following. In
particular, we check whether a user follows more Democrat or Republican Congress members
on Twitter. If they follow more Republican than Democrats, all their tweets would be
classified as Republican. In case a user either does not follow any Congress members or an
equal number of Congress members from either party, their tweets are classified as neutral.21
The second measure of political slant is based on the similarity of the text of tweets
to those sent by Republican or Democratic Congress members. We train a L2 regularized
logistic regression model separately for each election based on 901 Congress members’ Twitter
accounts to classify whether a tweet contains language frequently used by either Republican or
17
We report the full list of words in Table A.7.
18
Data on SXSW 2007 attendants (e.g., their county of residence) is not available, despite our efforts to
obtain it from the SXSW organizers on multiple occasions.
19
The remaining 12 politicians are either Independents or switched their party affiliation.
20
For the 2012 election, we use data collected by Diaz et al. (2016), comprising 24 million tweets containing
either “Obama” or “Romney” for the period from July 1, 2012 through November 7, 2012. For 2016, we use
the archive from Littman et al. (2016), which contains 280 million tweets, collected between July 13, 2016, and
November 10, 2016. The 2020 election tweets are based on the archive from Chen et al. (2020), which covers
the period from March 2020 to November 2020. To make these datasets comparable, we restrict the 2016
election sample to tweets mentioning either “Clinton” or “Trump” (112 million tweets). Similarly, we restrict
the 2020 data set to the time period from July 1, 2020 through November 3, 2020 and tweets mentioning
either “Biden” or “Trump” (339 million tweets).
21
The idea of using the Twitter network to determine a user’s ideology is inspired by Barberá (2015).
9
Democratic politicians.22 We then use this classifier to predict a partisan score between 0 and
1 for each of our election-related tweets. These scores can be interpreted as the probability of
a tweet with the same content being sent by a Republican. As such, our approach is similar
to how Gentzkow and Shapiro (2010) measure newspaper slant. Both approaches lead to
similar overall slant classifications for the elections tweets in our data.
10
“We didn’t actually launch Twitter at SXSW – SXSW just chose to blow it up.
We launched it nine months before – to a whimper. By the time SXSW 2007 rolled
around, we were starting to grow finally and it seemed like all of our users (which
were probably in the thousands) were going to Austin that year ... I don’t know
what was the most important factor, but networks are all about critical mass,
so doubling down on the momentum seemed like a good idea. And something
clicked.”24
SXSW’s immediate impact on Twitter’s popularity in early 2007 can be seen in Figure 4a,
which plots our proxy for the daily number of tweets as well as the number of tweets explicitly
mentioning SXSW. The figure shows that Twitter’s growth rate accelerated during the festival,
visible as the spike in SXSW-related tweets. The month-to-month growth rate of Twitter
quadrupled with the start of the SXSW festival.25 After SXSW 2007, Twitter experienced
further rapid growth (Venture Beat, 2008). The platform went from an average of 5,000
tweets a day in 2007 to 300,000 in 2008, and 2.5 million in 2009 (Twitter, 2010). In 2019,
users sent roughly 500 million tweets a day.
We exploit that the SXSW festival had persistent effects on Twitter’s spatial diffusion.
This is likely the result of network effects that are key to the social media experience, as
a larger number of users makes it more interesting for potential new users to join. Such a
mechanism also applies at the local level. For example, a boost in the number of neighbors,
personal connections, local businesses and/or people who play a prominent role in an area
should also boost the value of joining the platform for those living there. As Evan Williams’
quote above notes, “networks are all about critical mass,” and initial differences in adoption
can lead to persistent differences in network adoption. Enikolopov et al. (2020) document a
similar mechanism for spatial dispersion in adoption of the Russian VK social network: the
home towns of the first university students invited to be users in 2006 exhibit substantially
higher number of users in 2011.
We provide further support for this hypothesis by investigating whether the inflow
of early-stage adopters put these counties on a differential growth path of Twitter usage.
Figure 4b plots the estimates of βτ from the following panel event study regression at the
county (c) and week (t) level:
X X
tweetsct = βτ SXSWcM arch2007 × 1(t = τ ) + δτ SXSWcP re × 1(t = τ ) + θc + γt + εct .
τ τ
24
Appendix Figure B.1 provides Williams’ full post describing the role of SXSW 2007.
25
Our proxy for Twitter usage is created by scraping tweets that contain any of the 100 most common
English words listed in Table A.7. Our data contain any tweet that contains at least one of these words. We
should therefore obtain a large fraction of the English speaking tweets at that point in time.
11
where tweetsct is the log of (one plus) the number of tweets in county c on week t, SXSWcM arch2007
is the logarithm of (one plus) the number of SXSW followers in county c that joined Twitter
on March 2007 and SXSWcP re is a similarly defined variable for followers that joined Twitter
before March 2007. βτ thus illustrates, conditional on county and week fixed effects, the
difference in the number of tweets sent from counties with relatively larger numbers of SXSW
followers that joined on March 2007. The variables are standardized to have a mean of zero
and standard deviation of one. The whiskers represent 95% confidence intervals based on
standard errors clustered at the state level. The sample includes the period between the third
and fourteenth week of 2007.
Figure 4b illustrates that home counties of SXSW followers who joined during the
festival in March 2007 saw a rapid, disproportionate increase in Twitter usage around the
time of SXSW. Importantly, however, this increase came only after the SXSW festival, and
we find no evidence for pre-existing trends. In Appendix Figure B.2 we show further that, in
line with our expectations, March 2007 is also a clear outlier when it comes to the number of
people who started following the SXSW festival. This is consistent with the idea that SXSW
was a catalyst for the spread of Twitter in the United States.
Appendix Figure B.3a presents additional evidence on the long-term adoption effect of
the 2007 SXSW festival. It plots estimates from a similar regression as the one in Figure 4b
but in a county-quarter panel covering the period from Twitter’s launch in 2006 to 2016.
The dependent variable is substituted by the number of Twitter users per capita in a county
based on our baseline measure. The resulting S-shaped pattern in the figure is consistent with
models of technology adoption in the presence of network effects. More importantly, we find
that the amount of early adopters in a county still matters for the amount of Twitter usage
today.26
5 Empirical Framework
Our identification strategy leverages the 2007 SXSW festival as a shock to early Twitter
adoption. We show that, conditional on a set of controls (described in further detail below),
a county’s number of SXSW followers that joined Twitter in March 2007 is uncorrelated with
levels and trends in election outcomes before Twitter’s launch and during its early years. It is
26
Additionally, Figure B.3b shows just how dominant Twitter users connected to the SXSW festival were
among early adopters. In 2007, we estimate that around 60% of Twitter users either followed the SXSW
festival or followed someone who followed SXSW and joined in March 2007. As the number of Twitter users
increased over time, the importance of SXSW followers in the platform declined. But as Figure B.3a shows,
the festival created persistent differences at the county level. The next section outlines how we use the SXSW
festival in our 2SLS estimates.
12
also uncorrelated with a host of observable county characteristics. This feature of the data can
be interpreted as idiosyncratic factors (e.g., who attended the 2007 SXSW, who decided to
join Twitter at the time), giving us a “natural experiment” or “exogenous shock” in Twitter
adoption that allows to estimate its effect on election outcomes. This interpretation is, of
course, not self-evident, and we provide several pieces of evidence to support it.
An important concern is that counties whose population are more interested in the
SXSW festival (and its Twitter account) may be systematically different from other counties.
To address this issue, our empirical strategy exploits variation in the exact timing of when
Twitter users interested in SXSW joined the platform across counties. In particular, our
regressions control for the number of SXSW followers who joined in the months before the
festival. Intuitively, our empirical strategy compares a “treatment” group of counties with
SXSW followers that joined in March 2007 (during the festival) against a “control” group of
counties with followers that joined before. While both groups of followers were interested in
SXSW, we show that only the number of followers that joined on March 2007 are predictive of
later Twitter adoption, consistent with the evidence that users that joined during the festival
were key in the platform’s diffusion. In contrast, counties with more users that joined before
the festival do not have more additional Twitter users in subsequent years.27
The “treatment” and “control” counties are similar along several characteristics. Ta-
ble A.5 compares the average characteristics of three types of counties relevant for our
identification strategy: 1) the 47 counties with SXSW followers that joined Twitter both in
March 2007 and the “pre-period;” 2) the 108 counties with SXSW followers that joined in
March 2007 (but none in the “pre-period”’); and 3) the 20 counties with SXSW that joined in
the “pre-period” (but none in March 2007). Differences in vote shares in the 1996 presidential
election, demographics (e.g., race, age, education), and media consumption (e.g., share that
watches Fox News) are quantitatively small or zero. This is particularly true for groups
(2) and (3) — which are key to the identification — with t-tests indicating that differences
between the two groups are not statistically different from zero.28 The geographical variation
in the three groups of counties is shown in Figure 3b. As the results in Table A.5 suggest, the
counties do not differ systematically in size and how distant they are from major American
cities.
Moreover, observable individual characteristics of SXSW followers who joined Twitter
in March 2007 and the “pre-period” are also similar. We validate this using data on Twitter
27
An alternative approach is to compare the counties of users who signed up for Twitter during SXSW
2007 with those of users who signed up during other festivals in the same year. We discuss the results from
such an exercise in the robustness section below.
28
Given the large number of county characteristics, we report Šidàk-corrected t-statistics, which are smaller
than those generated by applying the Bonferroni correction.
13
user profiles we obtained from the platform. Table B.1 shows that followers who joined in
March 2007 have similar first names and profile descriptions compared to those that joined
before: users in both groups tend to have common names (e.g., “Michael” or “Chris”) and use
words such as “founder” or “tech” to describe themselves in their profiles. The correlations of
the frequency of first names and terms used in their bios between the two groups are 0.63 and
0.89, respectively. We also investigate differences in the political leanings of the two groups
using the network-based methods we outline in Section 3. In particular, we test whether the
users in March 2007 follow more Democrats or Republicans than the users in the “pre-period”.
We find that the political leanings of the two groups are nearly identical. A t-test rejects
differences in the average political slant with a p-value of 0.93.
Specification. Motivated by the evidence above, our main results are based on estimating
the following two equations:
where c indexes counties, SXSWcM arch2007 is the logarithm of (one plus) the number of
SXSW followers in county c that joined Twitter on March 2007, and SXSWcP re is a similarly
defined variable for followers who joined Twitter before March 2007. Xc is a vector of control
variables that hold constant geographical factors (e.g., population density, distance from
Austin, TX), demographic factors (e.g., the share of the population in different age and
ethnic groups), socioeconomic factors (e.g., the share of adults with a high school degree
or employed in IT), factors related to the “China shock” (e.g., exposure to Chinese import
competition), and previous election results. Note that the right-hand side of both equations is
similar. T witter usersc is the logarithm of the number of Twitter users in the county (during
2014-2015). yc are election outcomes (e.g., vote shares), which we estimate in both levels and
changes (e.g., yc can be the vote share in 2016 or the change in vote shares between 2000 and
2016).
In a 2SLS framework, equations (1) and (2) are the first-stage and reduced form, while
the second stage is
yc = φ + θ · T witter
\ usersc + π · SXSWcP re + Xc ρ + εc , (3)
14
where T witter
\ usersc is predicted from the first stage regression in equation (1). We weigh
observations by turnout (total number of votes cast) in the 2000 presidential election.29 We
cluster standard errors at the state level.30
Identification. Formally, the identification condition for the effect of Twitter users (θ)
is that E(SXSWcM arch2007 · εc ) = 0 holds. Intuitively, this states that, conditional on the
SXSWcP re and other controls (Xc ), the number of SXSW followers who joined in March 2007
is uncorrelated with other determinants of political outcomes yc , implying that it only affects
political outcomes via Twitter usage (the “exclusion restriction”).
We provide six pieces of evidence in support of this condition. First, as discussed above,
including the SXSWcP re control implies that the identifying variation comes from comparing
counties with similar observable characteristics.
Second, the coefficient of SXSWcP re is small and statistically insignificant in our first
stage regressions. This provides us with a “placebo” test based on checking if it is also
unrelated to political outcomes in the reduced form and 2SLS regressions. Intuitively, we
have two variables that are correlated with interest in the SXSW festival among early Twitter
adopters, but only one predicts Twitter users in later years, allowing us to disentangle interest
in the festival from its effect via more Twitter users.
Third, we provide additional placebo tests for other festivals in 2007 and show that
none of these other festivals is either correlated with higher Twitter adoption or changes in
the Republican vote share today. This allows us to rule out that our effects are simply driven
by the selection of users into attending festivals that are highly similar to SXSW.
Fourth, estimating equations (2) and (3) for different time periods shows that SXSWcM arch2007
does not correlate with both levels and trends in election outcomes before Twitter’s launch in
2006 and in its early years, when the platform had few users and was unlikely to affect election
outcomes. Intuitively, outcomes in “treatment” and “control” counties behaved similarly
before Twitter could plausibly affect elections.
Fifth, we find an effect of SXSWcM arch2007 on Trump’s vote share in 2016 and 2020 but
not on House and Senate elections (neither in 2016 nor 2020 or other periods between 2000
and 2018). This pattern is consistent with an effect of Twitter, since there is more content on
presidential candidates than on congressional elections in the platform.
29
We weigh observations to make our sample representative of national election results. For example, due
to many less populated counties tilting Republican, the unweighted average Republican vote share across
counties in the 2016 elections is 64%. Weighing makes our sample mean match the national average of 46%.
Moreover, we obtain similar results without using weights or using election-year turnout as weights (Appendix
Table C.3), suggesting that effect heterogeneity along dimensions correlated with weights do not play an
important role in our context (Solon et al., 2015).
30
We consider spatial standard errors using the methods described in Colella et al. (2019) for robustness.
15
Sixth, results based on survey data suggest the effects are concentrated among moderate
or independent voters, which is also the expected pattern from Twitter having a causal effect
due to voter persuasion.
Stated differently, a violation of the identification condition would require an omitted
variable that correlates with SXSWcM arch2007 , T witter usersc , and yc but is uncorrelated
with: i) SXSWcP re as well as followers of other festivals, ii) levels and trends in election
results before Twitter’s launch and rise to popularity, iii) the observable variables presented in
Table A.5, and iv) election results in congressional elections both during the Trump elections
and before, while also v) being correlated with vote choices of moderate voters. Our argument
is that the existence of such an omitted variable is implausible to an extent that allows us to
interpret θ as the effect of Twitter users on election outcomes.
Measurement error in county-level Twitter usage and SXSWcM arch2007 is also unlikely to
explain an effect in 2016 and 2020 presidential elections, but no effect in previous presidential
elections or congressional elections. Moreover, SXSWcP re and the measures for other festivals
are constructed similarly as SXSWcM arch2007 and should thus have similar measurement error.
However, SXSWcP re and the other festivals are uncorrelated with Twitter usage and election
outcomes.
Lastly, another possible concern is that the SXSW adoption shock led to differences
in the composition of Twitter users when compared to other U.S. counties. In particular,
one might be concerned that the SXSW festival lead to a more liberal Twitter population
in the treated counties. While this would not influence the causal interpretation of our
findings, it could make the local average treatment effect harder to interpret. Three pieces of
evidence suggest that this appears to be an unlikely concern. First, as we show in Appendix
Figure B.3b, Twitter’s user base became less connected to the SXSW festival over time and,
in this process, likely reached people from more diverse backgrounds. Second, the findings of
Müller and Schwarz (2019) indicate that the SXSW adoption shock was associated with an
increase in hate crime with Trump’s presidential run. This suggests that the shock eventually
reached even the right-wing fringes of the political spectrum. Third, we can directly address
this concern by comparing the profiles of Twitter users in SXSW home counties with those
in the rest of the country. The results are presented in Appendix Table B.2. We find that
the user profiles in SXSW counties are highly similar to the general Twitter population. If
the Twitter population in SXSW counties was significantly more liberal, their Twitter and
names and biographies should also be different, as Oliver et al. (2016) document that names
predict political ideology. We find similar results when we look at which politicians users in
the different counties follow. If anything, Twitter users in the “pre-period” counties appear
to have a slightly more liberal Twitter network.
16
To be transparent we want to stress what our findings do not imply. First, they cannot
speak about social media platforms other than Twitter, such as Facebook. Our empirical
strategy exploits a “shock” specific to early Twitter adoption and we do not have a credible
research design to estimate the effects of other platforms. While many other platforms share
similarities with Twitter, such as being popular among younger and more educated people
in urban areas (Pew Research Center, 2019c), other platforms may have different effects
on political outcomes. Second, our research design cannot separate the effect of particular
types of social media content on Twitter (e.g foreign governments or misinformation), but
rather speaks to the overall effect of Twitter exposure. Third, like other papers in media
economics, we estimate a “partial equilibrium” effect. In our case, we estimate the effect of
adding Twitter users to a county while keeping other counties’ Twitter use constant. We
thus cannot address whether Twitter had a national-level effect on the election (e.g., Trump’s
tweets driving traditional media content).
6 Results
6.1 Main results
First-stage. Table 1 reports results from estimating equation (1) with different sets of
control variables (described in Appendix Table A.6). The results indicate that counties with
more SXSW followers who joined Twitter in March 2007 have higher numbers of Twitter
users during 2014-2015. Since the variables are in logs, the coefficients can be interpreted
as elasticities. A 10% increase in SXSW followers in March 2007 is associated with 5.2%
more Twitter users. The results do not seem to be sensitive to the set of included covariates.
For example, the distance from Austin, Texas (the location of SXSW) has no significant
explanatory power. Importantly, the coefficients on SXSW followers before the 2007 festival
are statistically insignificant and small in size: Twitter usage in 2014-2015 is not higher in
areas with more SXSW followers who joined Twitter before March 2007.
Figure 5 presents the graphical representation of the estimates in column (5) of Table 1.
Specifically, we show a binned scatter plot of T witter usersc against SXSWcM arch2007 after
both variables are “residualized” by partialling out the control variables. The figure is
constructed by dividing the x-axis variable into 40 equal-sized bins and plotting the average
values of both variables in each bin.31
31
The fitted line is based on the unbinned data. Observations are weighted by turnout in the 2000 presidential
election. This procedure guarantees the slope of the fitted line matches the estimate on column (5) of Table 1.
17
Reduced Form and 2SLS Estimates. Table 2 shows the reduced form estimates from
equation (2) and both OLS and 2SLS estimates of equation (3), focusing on the Republican
vote share in the 2016 and 2020 presidential elections. The specifications across columns
match those in Table 1. Panel B indicates that the number of SXSW followers who joined
Twitter in March 2007 is associated with a lower Republican vote share. Panel C presents
the 2SLS effects of Twitter usage on vote shares. The 2SLS estimate in column (5) indicates
that a 10% increase in the number of Twitter users in a county lowers Trump’s vote share
by 0.21 p.p. (e.g., a reduction from a 46.1% vote share to 45.8%).32 We discuss potential
differences between LATE and ATE as well as heterogeneous treatment effects at the end of
this section. The results for the 2020 presidential election shown in Table 2 column (6)-(10)
are nearly identical.
Figure 6 plots the reduced form and OLS estimates from Table 2 graphically, specifically
the models in columns (5) and (10). These figures are constructed similarly to Figure 5 but
show the Republican vote share on the y-axis. The estimated slopes are negative.
A first draft of this paper, using only data up to the 2016 election, was posted online on
October 2020. When updating it to include the November 2020 election results, we made no
revisions to the research design and regression specifications (e.g., sample selection, choice of
controls, and variable definitions). In the case of Table 2, columns (1)-(5) are exactly the same
in this version and in the October 2020 version. Since columns (6)-(10) replicate the same
specifications (“right hand sides”) using 2020 election outcomes, this can be interpreted as a
“pre-registered” research design. We followed the same approach throughout this draft: every
figure or table presenting results for the 2020 election exactly replicates the specifications
used for previous elections in our October 2020 draft.33
Although the estimated effects are always negative and significant at the 1% level,
comparing columns shows that the effect sizes vary somewhat with the inclusion of controls,
especially demographic and socioeconomic ones. This sensitivity to controls is perhaps
expected given their explanatory power over vote shares. The results in other papers that
explore effects on vote shares in similar frameworks, such as DellaVigna and Kaplan (2007),
Martin and Yurukoglu (2017), and Autor et al. (2020), show a similar sensitivity to controls.
To further probe the sensitivity of our results to different controls, we provide three
separate pieces of evidence. First, we apply the Oster (2019) approach of using coefficient
stability to gauge the potential importance of unobservable variables in driving the results.
We compare our reduced form (Panel B) specifications with all controls (columns 5 and 10) to
32
The F-statistic of our estimated first-stage range from 70 to 120. This suggests that estimation and
inference concerns related to weak instruments are unlikely to apply in our case.
33
The October 2020 draft is available at Social Science Research Network (SSRN):
https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3719998.
18
those with the fewest controls (columns 1 and 6) and obtain an “Oster-δ” of approximately 7
for both 2016 and 2020. This suggests, in order for the true effect to be zero, unobservable
variables would have to be seven times as “important” as the (numerous) controls added in
columns (5) and (10) in driving selection into “treatment.”34
Second, Appendix Table C.1 provides further evidence for the robustness of our findings
to the inclusion of different controls. It allows for more flexible interactions between observ-
able characteristics by using partialling-out LASSO variable selection (Chernozhukov et al.,
2015a,b).35 More specifically, we estimate regressions in which we allow for state fixed effects
and all possible interactions of the control variables (i.e., up to 654 controls). The LASSO
then selects controls that either predict the vote shares, SXSW participation, or Twitter usage.
All estimates are statistically significant and of similar magnitude as our baseline estimates.
Third, Figure D.3 plots the estimated θ of our 2SLS equation (3) while flexibly allowing
the included control variables to vary. The resulting “specification curves” suggest that our
results (for both the 2016 and 2020 elections) are robust to how our regressions are specified.
The estimated coefficients are always negative, almost always statistically significant at the
5% level, and in the overwhelming number of specifications considerably more negative than
our “baseline estimates,” which is marked by the vertical line.
19
This rate is smaller than the estimated pro-Republican persuasion rate of Fox News,
which Martin and Yurukoglu (2017) estimate to range between 27% and 58% (depending
on the year).37 It is also smaller than the 19.5% pro-Democrat persuasion rate that Gerber
et al. (2009) estimate for the Washington Post. As a further comparison, Gentzkow et al.
(2011) estimate a persuasion rate of 12.9% for the effect of reading a local newspaper in the
1869-1928 period on voter turnout. DellaVigna and Gentzkow (2010) report that randomized
get-out-the-vote canvassing experiments have estimated persuasion rates in the 10-15% range.
Placebo Test: Interest in SXSW Before March 2007. The coefficients on SXSWCP re
in Table 2 are statistically insignificant and substantially smaller than those on SXSWcM arch2007 .
As discussed in Section 5, this provides support for our identification condition (exclusion
restriction). Suppose that our instrument merely captured that counties with an interest
in SXSW’s Twitter account during the platform’s early years also differ in (unobservable)
characteristics that predict the 2016 election outcome. If this was the case, the coefficients
on SXSWCP re should be similar in size to those on SXSWcM arch2007 . Intuitively, we have two
variables that are correlated with interest in the SXSW festival, but only one predicts Twitter
users in later years, allowing us to disentangle interest in the festival (and its correlates) from
its effect via more Twitter users.
Placebo Test: Other Festivals in 2007. We also provide an additional placebo check by
investigating five other festivals that are similar in nature to SXSW (Burning Man, Coachella,
Pitchfork, Austin City Limits, and the Electric Daisy Carnival).38 One of these placebo
festivals, Austin City Limits, takes place in the same city as SXSW. For this exercise, we
construct analogous measures to our instrument for other festivals using the same exact
procedure: the Twitter followers of the respective festival that joined in the month the festival
took place in 2007. In Appendix Table C.2, we then show that none of these festival variables,
including that for Austin City Limits, has predictive power for future Twitter usage. More
importantly, the variables are also essentially uncorrelated with 2016 and 2020 presidential
vote shares. These results provide further support for our identifying assumption and suggest
that a spurious relationship between counties’ Twitter users interest in festivals during 2007
and future election outcomes cannot explain our findings.
population. Note also that we estimate county-level effects, which may also capture local spillovers and social
interaction effects.
37
DellaVigna and Kaplan (2007) estimate a smaller persuasion rate of 11.6% for Fox News.
38
These festivals take place in the Black Rock Desert (NV), Indio (CA), Chicago (IL), Austin (TX), and
Las Vegas (NV), respectively.
20
Further Robustness and Additional Tests. The Online Appendix presents a number
of additional sensitivity checks. In Table C.3, we consider changes to the baseline regression
specification. In particular, we allow for unweighted regressions; weighting by the relevant
election-year turnout (2016 or 2020); alternative functional forms of the pre-SXSW user
variable; restrict the sample to the sub-sample of counties where we observe either SXSW
followers who joined in March 2007 or the pre-period; allow for flexible spatial correlation in
the standard errors; and per capita measures of Twitter usage. Further, Table C.5 replaces
our baseline time-invariant measure of Twitter usage (making use of all available user data)
with a time-varying measure based on how many users were using the platform in a given
year. None of these adjustments make a substantial difference in the magnitudes or statistical
significance of the estimates.
Online Appendix Table D.1 reports results for additional outcome variables, all of which
support the idea that Twitter exerts a pro-Democrat effect. In column (1), we use a probit
IV model to investigate the likelihood of a county switching from Obama in 2008 to Trump
in 2016. The coefficient suggests that a one standard deviation increase in Twitter usage is
associated with a -24% lower probability of a county electing Obama in 2008 and Trump in
2016. Columns 2 and 3 look at changes in campaign donations to Democrats and Republicans
between 2000 and 2016. We find a positive and statistically significant effect for donations to
Democrats, and no effect for Republicans. Lastly, columns 4 and 5 look at approval ratings
for President Trump in 2017 based on data from the Gallup Daily Poll. We find that exposure
to social media is associated with a decrease in Trump’s popularity, and more so among
Republicans.
The 2007 SXSW Shock and Previous Election Outcomes Table 3 repeats the analysis
from column (5) of Table 2 using Republican vote share in the previous presidential elections
of 2000, 2004, 2008, and 2012 as the dependent variable.39 All estimates for those years are
substantially smaller than the ones for 2016 and 2020 and statistically insignificant. For 2000,
2004 and 2008, this can be interpreted as a placebo or pre-trends test: conditional on the
covariates, our instrument is uncorrelated with outcomes before Twitter’s launch (2000 and
2004) and when the platform had few users (2008). In 2012, Twitter already had a substantial
user base, so our estimates can be interpreted as a genuine “zero effect”; we return to a
comparison of the 2020, 2016, and 2012 results in Section 6.3. The estimates for 2000, 2004,
2008, and 2012 are also jointly statistically insignificant (p-value=0.329). The average effect
39
Note that the right-hand side of the regressions remain the same as in the previously reported regressions.
That is, the instrument is the same and the “endogenous variable” in the 2SLS is Twitter users measured in
2014-2015. We explore the role of time-varying Twitter user measures on Appendix Table C.5.
21
in 2016 and 2020 is statistically different from the average effect in 2000, 2004, 2008, and
2012 (p-value=0.002).
In Appendix Figure C.1, we further show the reduced form estimates for presidential
election going back as far as 1924. We find that our instrument is also uncorrelated with
any of the earlier presidential election results. As discussed in Section 5, this result lends
additional support for our exclusion restriction. If our instrument merely captured uncontrolled
differences across counties, these should also correlate with vote shares in previous elections.
While these findings make it unlikely that our instrument is correlated with pro-
Democratic attitudes at the county level, a possible concern could be that we are picking up
“anti-populist” attitudes, which could have harmed Trump’s electoral results. To address this
concern, we turn to the historical case study of Ross Perot’s political campaign in 1992 and
1996. Perot, a billionaire businessman, also ran as a “third-party candidate” on a populist
platform. However, when we replace the dependent variable with the third party vote share
in the 1992 and 1996 presidential election (see Appendix Table C.6), we find no evidence that
our instrument is associated with lower vote shares for Ross Perot. This makes it unlikely we
are capturing differences in “demand for populism” across counties.40
Effects on Vote Share Changes. We also consider specifications of equations (2) and (3)
using vote share changes instead of levels as the dependent variable. All our estimates based
on changes take differences relative to the base year 2000 (akin to the approach in Autor et al.
(2020)) and use the full set of controls (as in columns 5 and 10 of Table 2). Figure 7a plots
the reduced form estimates for changes in the Republican vote share in presidential elections.
The results corroborate the previously presented evidence based on specifications in
levels. Our instrument is essentially uncorrelated not only with levels but also with changes
(or trends) in election outcomes during the 2000-2012 period. Given our arguments above,
this also lends support for our identification strategy. Again the estimates for 2000, 2004,
2008, and 2012 are also jointly indistinguishable from 0 (p-value=0.398). The reduced form
effects for 2016 and 2020 are similar to the one estimated using levels. For example, the
estimated effect (θ) for 2016 using changes is -0.017, similar to the one estimated in levels
(-0.021).41 Unsurprisingly, the average effect in 2016 and 2020 is statistically different from
the average effect in 2000, 2004, 2008, and 2012 (p-value=0.001).
40
In Appendix Table C.7, we also investigate the vote shares in the 2020 democratic primaries. Here we
find a positive association between Twitter exposure and the vote share of Bernie Sanders, often described as
a left-wing populist. This further speaks against the hypothesis of “anti-populist” sentiment.
41
Appendix Table C.4 present the OLS and 2SLS estimates.
22
Effects on Turnout and Congressional Elections. Figure 7b, Figure 8a, and Figure 8b
replicate the exercise in Figure 7a using voter turnout and vote shares in House and Senate
elections as the outcomes. We do not find a statistically significant association between our
instrument and election turnout except for 2020. Before 2020, the estimated point effects
are usually small. For example, the upper bound on the 95% confidence interval of the 2SLS
estimate for the effect of turnout in the 2016 election implies that a 10% increase in Twitter
users raises turnout by 0.036 p.p. (Appendix Table D.2).
In the 2020 election, which saw the highest turnout rate in more than a century (NPR,
2020c), we find that Twitter is associated with a larger fraction of the voting age population
casting their ballot. Why did Twitter have an effect on 2020 turnout but not in the previous
election? One possible explanation could be that calls to turn out and vote were widespread
on the platform, partially because of an initiative by Twitter itself that was not present in
2016 (Twitter, 2020). Further, the 2020 election was unique in its prevalence of mail and
early voting because of the Covid-19 pandemic. Another possibility is that Twitter served as
platform to convey information on how to vote by mail or before election day.
The coefficients for House and Senate races are more noisily estimated, particularly for
the smaller sample of Senate races (where only a third of seats is renewed every two years).
Overall, there is little evidence suggesting an effect of Twitter on congressional elections,
including in 2016, the 2018 midterm election, and 2020. Finding an effect on presidential vote
shares but not in these “down-ballot” races is perhaps expected since content on presidential
candidates (and in particular on Trump in 2016 and 2020) is more common on Twitter than
content on congressional races.42
23
Table A.5, and iv) election results in congressional elections. At the same time, such omitted
variable would also v) be more strongly correlated with the vote choices of moderates and
independents than “partisans.” Our argument that is the existence of such omitted variable
is unlikely.
Towards an Average Treatment Effect As with any instrument, our 2SLS results
identify a local average treatment effect (LATE). In our setting, the “compliers” are counties
with higher Twitter usage as a result of the inflow of SXSW attendees. While the negative
treatment effect of Twitter usage for these counties is in itself an interesting finding, the
ATE for the US overall—and therefore the overall impact of Twitter on elections—may differ.
However, two pieces of evidence suggest that our estimates, despite this concern, allow us to
infer information about the ATE.
A first indication comes from comparing the OLS and 2SLS results in Table 2. Both
estimates are always negative and relatively similar in magnitude. Second, we build on
the approach suggested by Andrews and Oster (2019) and more formally investigate the
external validity bias of our estimates. For experimental settings, Andrews and Oster
(2019) suggest using the observable heterogeneity in estimated treatment effects within the
experimental sample to learn about the ATE in the overall population. Similarly, we can
use the heterogeneity of treatment effects within counties that provide the variation that
identify our results to approximate the ATE for the United States as a whole. Using all
included control variables from our main specification for the prediction of heterogeneity of
the treatment effect, the Andrews and Oster (2019) approach suggests that the ATE should, if
anything, be larger than the LATE we estimate in our baseline results. This seems plausible as
the more urban counties for which we have variation in our instrument tend to be Democratic
strongholds, and thus likely have fewer independents, for which we find the largest persuasion
effects (in survey data). We provide additional details on our approach in Appendix E43
24
particular, we estimate the following instrumental variable Probit regression:
yic = φ + θ · T witter
\ usersc + π · SXSWcP re + Xic ρ + εic , (4)
where yic is an indicator variable equal to 1 if an individual i living in county c voted for
Trump in the 2016 or 2020 election and 0 for Clinton.44 The definition of the county-level
variables T witter usagec and SXSWcP re remains unchanged. Xic is now a vector of individual-
level control variables including age, gender, race, family income, and education. We again
instrument for county-level Twitter usage based on the SXSW followers who joined in March
2007.
Table 4 presents results from estimating equation (4). In Panel A and B, Column
(1) suggests county-level Twitter usage has a statistically significant negative effect on the
likelihood to vote for Trump. The marginal effect implies that a 10% increase in the number
of Twitter users in a county would lower Trump’s vote share by 0.49 p.p. in the 2016 and
0.46 p.p. in the 2020 presidential election.45
Columns (2)-(6) report results estimated separately by voters’ reported party affiliation.
The effect is strongest for voters who identify as independents, and thus likely to not hold
strong priors. The results suggest weaker or zero effects for those with stronger political views,
whether Republican or Democrat.
These results on party affiliation further vary with age in a way that may be suggestive
of a role for social media. In the Online Appendix, Table D.4 repeats the 2016 exercise of
splitting CCES respondents by their party affiliation, but further divide voters into those
below and above 50. Since Twitter users tend to be younger than the general population
(Section 2), one may expect larger effects among voters below 50.
Indeed, we find larger marginal effects for Log(T witter users) among younger voters.
For independents, the estimate for young voters is 5% larger than for older voters (−0.079
compared to −0.071). Among moderate Republicans and Democrats, the estimated coefficients
are close to zero for voters aged 50+ but sizeable (at least twice as large) for younger voters
(although they are not statistically significant at conventional levels). Because young voters
are less likely to vote for Trump, this implies larger elasticities of vote outcomes with respect
to Twitter for those below 50 relative to the baseline probabilities.
44
Note that we use data on all CCES respondents, not only those in the 2016 wave (i.e., respondents from
the 2018 and 2020 waves were asked how they voted in 2016). The results are similar if we only use the 2016
wave. In unreported regressions, we do not find an effect on votes for Jill Stein.
45
This effect size is within the range of the county-level estimates presented in Table 2. Appendix Table C.8
shows this baseline result is robust to using only individuals with validated turnout and/or who stated that
they originally intended to vote for Trump in the pre-election wave of the CCES.
25
A potential concern with this exercise is that party affiliation may itself be affected
by Twitter usage. We thus present further support for Twitter having persuasion effects on
moderates using county-level data that is not subject to such concerns. In particular, we
estimate our county-level specification (equation 3) for the 2016 elections, but split counties
based on how consistently they voted for either the Republican or the Democratic party.
Specifically, we define “swing counties” as counties that were not consistently won by one
party in the presidential elections from 2000 to 2012. Because we find no effect of Twitter on
vote shares before 2016 (see Table 3), splitting the sample by swing counties is not subject to
potential concerns that this variable might itself be affected by social media.
Appendix Table D.5 shows the results. For the 2016 presidential elections, we find that
Twitter usage only negatively impacts the Republican vote share in “swing” counties. We find
no evidence for Republican or Democratic strongholds, where people likely have the strongest
priors. The patterns are similar in 2020. But here, we also find a small effect on counties that
usually vote Republican, which could suggest that the effect on moderate Republicans we
find in the CCES also apply to the county level.
26
Two pieces of evidence presented above are consistent with the political shock channel.
First, if Twitter had little effect before 2016 because it was not widely used, its effect on vote
shares should systematically increase over time. We do not find evidence supporting this idea
in effects for presidential, House, and Senate elections (Figures 7 and 8). Instead, we find a
discontinuous negative effect in the 2016 presidential election that persists in 2020. Second,
we do not find substantial effects for the 2016, 2018, and 2020 House and Senate elections.
This implies that Twitter usage lowered Trump’s vote share without substantially affecting
other Republican candidates on the same election day.
Results from the 2016 Republican Primaries. We provide additional evidence for a
Trump-specific effect of Twitter exposure by investigating the 2016 county-level Republican
primaries results. The primaries allow us to focus on the favorability of different candidates
among Republican voters. The results from this exercise are presented in Table 5. We find that
Twitter usage is associated with a lower vote share for Trump. We also find a positive effect
on the vote share of John Kasich, the most moderate of the major Republican candidates.46
Results from Gallup Approval Ratings. A similar pattern emerges when we use data
from the Gallup Daily Tracker, which contains approval ratings for three other Republican
presidential candidates who ran alongside Trump during the primaries (Ted Cruz, Marco
Rubio, and Kasich). Table 6 shows the results of running individual-level IV probit regressions
as in equation (4), where the dependent variable is an indicator variable equal to one if
the respondent approved of a specific candidate. As in Table 4, we differentiate between
respondents’ political affiliation.47
Table 6 confirms our main county-level result from general elections and primaries:
Twitter usage is associated lower approval of Trump, especially among independents (who are
presumably more likely to be persuaded by social media content). We also find lower approval
of Cruz, who is substantially more right-wing than other presidential primary candidates in
recent years (FiveThirtyEight, 2015). We find no link between Twitter use and approval of
the more moderate Republican candidates, Rubio and Kasich. For the Democrat candidates,
we find an effect for Clinton’s but not Sanders’ approval.48
46
In Appendix Table C.7, we also investigate the voting behavior in the 2016 and 2020 Democratic primaries.
Twitter usage appears to be associated with a higher support for Bernie Sanders in 2020.
47
We pool people who identify as leaning Republicans or Democrats with independents, because—in contrast
to the CCES data—only a few individuals in the survey are classified as “leaners.”
48
Note that Appendix Table C.7 reports a significant positive effect for Sanders county-level primary vote
share in 2020, but not in 2016. As mentioned earlier, Gallup Daily Tracker data is not available for 2020 at
the time of writing. Note also that Gallup collects data on candidate approval (as opposed to vote choice)
for the entire population (as opposed to Democratic primaries votes), which may explain some differences
between these results and the ones based on county-level primaries results.
27
Taken together, these results are consistent with Twitter turning voters against voting
for Trump in particular and not against the Republican party more generally. Our results
may also explain the absence of an effect in the 2008 and 2012 elections: Obama’s opponents
John McCain and Mitt Romney were widely considered to be moderate Republicans (e.g.,
more similar to Kasich than Trump or Cruz).49
Slant of Election-Related Tweets. We provide further support for the hypothesis that
Trump’s 2016 campaign and presidency triggered opposition on Twitter by analyzing the
content of more than 460 million tweets mentioning the last name of presidential candidates
during the 2012, 2016, and 2020 presidential campaigns.
First, we classify the tweets’ slant as Republican, Democrat, or neutral using two
approaches described in Section 3. In the first case, we classify the political affiliation of
Twitter users by counting the number of Democrat and Republican Congress members they
follow. If a user follows more Democrats than Republicans, they are classified as Democrat,
and vice-versa. Tweets sent by a user classified as Democrat are classified as Democrat,
and so forth.50 In the second case, we classify individual tweets (not users) following an
approach in the spirit of Gentzkow and Shapiro (2010) and train a L2 regularized logistic
regression classifier to predict whether a tweet is more likely to have a Democrat or Republican
slant, depending on its content’s similarity to tweets sent by Congress members. If a tweet’s
content has higher similarity with those of Democratic Congress members, it is classified as
Democratic, and as Republican otherwise.51
Figure 9 plots the amount of Twitter attention directed at the Republican and Democratic
presidential candidates in the 2012, 2016, and 2020 elections, as well as the tweets’ estimated
slant based on users’ following of Congress members. To account for the attention and
popularity of tweets, we base the graphs on the number of “likes” the tweets mentioning the
last name of candidates received. In Appendix Figure D.1, we confirm that the results are
similar using the number of tweets and when we base the slant classification on the text of
the tweets (Figure D.2).52
49
As previously discussed, the number of Twitter users in 2008 was relatively small, but by 2012, it was
relatively close to its 2016 level (Figure A.1a).
50
Users who follow an equal number of Democrats and Republican or no Congress members are classified
as neutral. This approach is similar in spirit to Barberá (2015), who uses the network of Twitter users to
create a measure of ideology. Because we are only interested in a binary measure of partisan slant and not
the ideological distance of users, we do not estimate the full Bayesian ideal point mode. The advantage of our
simplified approach is that it is faster to compute and the resulting measure is easier to interpret.
51
See Section 3 and Appendix A.1. in the Appendix for more details. In unreported robustness checks, we
confirm that these findings are robust using different slant cutoffs for classifying Republican and Democratic
tweets, such as only classifying tweets for which the class probabilities are above 75% or 90%.
52
Twitter users can choose to “like” each tweet they see in their “timeline.” A user can only “like” a
particular tweet once. “Likes” thus provide an useful metric since they capture the popularity and attention
28
Panel A shows the number of “likes” for tweets mentioning the Republican presidential
candidate (Romney and Trump), while Panel B provides similar evidence for the Democrats
(Obama, Clinton, and Biden). There are three noteworthy facts presented in the figure. First,
there was a sizable growth in the overall volume of Twitter content mentioning presidential
candidates. Second, the number of “likes” for tweets mentioning Trump is larger than those
mentioning his opponents (the difference is fourfold in 2016 and almost threefold in 2020).
Note that a “like” for a tweet mentioning a candidate can occur for tweets that are positive
or negative about the candidate, so the overall size of the bars are not informative about slant
or sentiments of Twitter content.
Third, Figure 9 also breaks down the share of tweets mentioning the candidates by
slant. The content sent by users classified as Democrats are more sizable than that from
those classified as Republicans. In particular, the amount of attention (proxied via “likes”)
on Twitter content mentioning Trump posted by Democrats is almost twice as large as the
amount posted from Republicans (for both 2016 and 2020). On the other hand, content on
Biden was more likely to have a Democrat slant and content on Clinton was almost equally
likely to have a Democrat or Republican slant.
Overall, a potential interpretation of these results is as follows. Twitter users, and users
of other social media platforms, are more likely to be young, well-educated, live in dense urban
areas, and support the Democratic party (see discussion in Section 2). Perhaps as a result,
Democratic politicians are more popular on Twitter than Republicans (Figure 1). In 2016 and
2020, Twitter became a vehicle for spreading opinions, particularly from Democratic-slanted
users, on Trump. This may, in turn, have persuaded voters with weaker priors—independents
and perhaps more moderate Republicans—to vote against Trump in the presidential election.53
7 Conclusion
Election officials around the globe are concerned about social media’s increasing influence
on voting decisions (e.g. NPR, 2020a). At the time of writing, there is a heated debate
about whether platform providers should “moderate” election-related content in the U.S. (e.g.
Politico, 2020). Exploiting variation based on a shock to Twitter’s initial rise to popularity,
that the content received. For example, if an account sent millions of Republican-slanted tweets about Clinton,
but such account had few followers and thus few users who can “like” the message, it would not meaningfully
affect measures based on “likes,” but could potentially do so for measures based on number of tweets.
53
Note that the 2012 election differed from 2016 and 2020 both in terms of having less twitter content
referring to it (as Figure D.2 indicates) and also for having a more moderate Republican candidate (Romney)
that may not have attracted as much Democrat-slanted content.
29
our paper provides some of the first empirical evidence that social media can affect election
outcomes.
We find that Twitter lowered the Republican party’s vote share in the 2016 and 2020
presidential elections. While this finding runs counter to a popular narrative that places social
media at the heart of Trump’s election win, it is consistent with a growing body of evidence
showing that social media users were less, not more likely to vote for Trump in 2016 or hold
polarized views (Boxell et al., 2017, 2018).
We also provide support for the idea that the demographics of Twitter users may account
for the platform’s partisan effects. People who use Twitter are 25 percentage points more
likely to identify as Democrats rather than Republican, and Democratic politicians are more
popular on Twitter than Republican ones. Our work suggests that this environment not only
reflects selection of like-minded individuals, but also affects voting decisions, particularly for
people with more moderate views.
30
References
Acemoglu, D., T. A. Hassan, and A. Tahoun (2017, 08). The Power of the Street: Evidence
from Egypt’s Arab Spring. The Review of Financial Studies 31 (1), 1–42.
Allcott, H., L. Braghieri, S. Eichmeyer, and M. Gentzkow (2020, March). The Welfare Effects
of Social Media. American Economic Review 110 (3), 629–76.
Allcott, H. and M. Gentzkow (2017, May). Social Media and Fake News in the 2016 Election.
Journal of Economic Perspectives 31 (2), 211–36.
Andrews, I. and E. Oster (2019). A Simple Approximation for Evaluating External Validity
Bias. Economics Letters 178 (C), 58–62.
Autor, D., D. Dorn, G. Hanson, K. Majlesi, et al. (2020). Importing Political Polarization?
The Electoral Consequences of Rising Trade Exposure. American Economic Review 110 (10),
3139–83.
Autor, D. H., D. Dorn, and G. H. Hanson (2013, October). The China Syndrome: Local
Labor Market Effects of Import Competition in the United States. American Economic
Review 103 (6), 2121–2168.
Bakshy, E., S. Messing, and L. A. Adamic (2015). Exposure to Ideologically Diverse News
and Opinion on Facebook. Science 348 (6239), 1130–1132.
Barberá, P. (2014). How Social Media Reduces Mass Political Polarization. Evidence from
Germany, Spain, and the US.
Barberá, P. (2015). Birds of the Same Feather Tweet Together: Bayesian Ideal Point
Estimation Using Twitter Data. Political analysis 23 (1), 76–91.
Bessi, A., F. Zollo, M. D. Vicario, A. Scala, G. Caldarelli, and W. Quattrociocchi (2015, 08).
Trend of Narratives in the Age of Misinformation. PLOS ONE 10 (8), 1–16.
31
Bloom, N., J. Liang, J. Roberts, and Z. J. Ying (2015). Does Working from Home Work?
Evidence from a Chinese Experiment. The Quarterly Journal of Economics 130 (1), 165–218.
Boxell, L., M. Gentzkow, and J. M. Shapiro (2017). Greater Internet Use Is Not Associated
with Faster Growth in Political Polarization Among US Demographic Groups. Proceedings
of the National Academy of Sciences of the United States of America, 201706588.
Boxell, L., M. Gentzkow, and J. M. Shapiro (2018, 07). A Note on Internet Use and the 2016
U.S. Presidential Election Outcome. PLOS ONE 13 (7), 1–7.
Braghieri, L., R. Levy, and A. Makarin (2021). Social Media and Mental Health. Available at
SSRN: https://ssrn.com/abstract=3919760 or http://dx.doi.org/10.2139/ssrn.3919760 .
Bursztyn, L., G. Egorov, R. Enikolopov, and M. Petrova (2019, December). Social Media and
Xenophobia: Evidence from Russia. Working Paper 26567, National Bureau of Economic
Research.
Bursztyn, L., G. Egorov, and S. Fiorin (2020). From Extreme to Mainstream: How Social
Norms Unravel. American Economic Review 110 (11), 3522–3548.
Campante, F., R. Durante, and F. Sobbrio (2017, 12). Politics 2.0: The Multifaceted Effect
of Broadband Internet on Political Participation. Journal of the European Economic
Association 16 (4), 1094–1136.
Campante, F. and D. Hojman (2013). Media and Polarization: Evidence from the Introduction
of Broadcast TV in the US. Journal of Public Economics 100, 79–92.
Chen, E., A. Deb, and E. Ferrara (2020). # election2020: The first public twitter dataset on
the 2020 us presidential election. arXiv preprint arXiv:2010.00600 .
Chen, Y. and D. Y. Yang (2019, June). The Impact of Media Censorship: 1984 or Brave New
World? American Economic Review 109 (6), 2294–2332.
32
Chernozhukov, V., C. Hansen, and M. Spindler (2015b). Valid post-selection and post-
regularization inference: An elementary, general approach. Annu. Rev. Econ. 7 (1), 649–688.
Colella, F., R. Lalive, S. O. Sakalli, and M. Thoenig (2019). Inference with Arbitrary
Clustering. IZA Discussion Papers 12584, Institute of Labor Economics (IZA).
Del Vicario, M., A. Bessi, F. Zollo, F. Petroni, A. Scala, G. Caldarelli, H. E. Stanley, and
W. Quattrociocchi (2016). The Spreading of Misinformation Online. Proceedings of the
National Academy of Sciences 113 (3), 554–559.
DellaVigna, S. and E. L. Ferrara (2015). Chapter 19 - Economic and Social Impacts of the
Media. In S. P. Anderson, J. Waldfogel, and D. Strömberg (Eds.), Handbook of Media
Economics, Volume 1 of Handbook of Media Economics, pp. 723 – 768. North-Holland.
DellaVigna, S. and E. Kaplan (2007). The Fox News Effect: Media Bias and Voting. The
Quarterly Journal of Economics 122 (3), 1187–1234.
Diaz, F., M. Gamon, J. M. Hofman, E. Kıcıman, and D. Rothschild (2016). Online and Social
Media Data as an Imperfect Continuous Panel Survey. PloS one 11 (1), e0145406.
Eady, G., J. Nagler, A. Guess, J. Zilinsky, and J. A. Tucker (2019). How Many People Live
in Political Bubbles on Social Media? Evidence From Linked Survey and Twitter Data.
SAGE Open 9 (1), 2158244019832705.
eMarketer (2019). US Time Spent with Social Media 2019, By Debra Aho Williamson.
Enikolopov, R., A. Makarin, and M. Petrova (2020). Social Media and Protest Participation:
Evidence from Russia. Econometrica 88 (4), 1479–1514.
33
Enikolopov, R., M. Petrova, and E. Zhuravskaya (2011). Media and Political Persuasion:
Evidence from Russia. The American Economic Review 101 (7), 3253–3285.
Falck, O., R. Gold, and S. Heblich (2014, July). E-lections: Voting Behavior and the Internet.
American Economic Review 104 (7), 2238–65.
Fergusson, L. and C. Molina (2021, April). Facebook Causes Protests. Documentos CEDE
018002, Universidad de los Andes - CEDE.
FiveThirtyEight (2015). Let’s Be Serious About Ted Cruz From The Start: He’s Too Extreme
And Too Disliked To Win.
Flaxman, S., S. Goel, and J. M. Rao (2016, 03). Filter Bubbles, Echo Chambers, and Online
News Consumption. Public Opinion Quarterly 80 (S1), 298–320.
Gavazza, A., M. Nardotto, and T. Valletti (2019). Internet and Politics: Evidence From UK
Local Elections and Local Government Policies. The Review of Economic Studies 86 (5),
2092–2135.
Gentzkow, M. (2006). Television and Voter Turnout. The Quarterly Journal of Eco-
nomics 121 (3), 931–972.
Gentzkow, M. and J. M. Shapiro (2010). What Drives Media Slant? Evidence from US Daily
Newspapers. Econometrica 78 (1), 35–71.
Gentzkow, M. and J. M. Shapiro (2011). Ideological Segregation Online and Offline. The
Quarterly Journal of Economics 126 (4), 1799–1839.
Gentzkow, M., J. M. Shapiro, and M. Sinkinson (2011, December). The Effect of Newspaper
Entry and Exit on Electoral Politics. American Economic Review 101 (7), 2980–3018.
Gerber, A. S., J. G. Gimpel, D. P. Green, and D. R. Shaw (2011). How Large and Long-lasting
Are the Persuasive Effects of Televised Campaign Ads? Results from a Randomized Field
Experiment. The American Political Science Review 105 (1), 135–150.
Gerber, A. S. and D. P. Green (2000). The Effects of Canvassing, Telephone Calls, and Direct
Mail on Voter Turnout: A Field Experiment. The American Political Science Review 94 (3),
653–663.
Gerber, A. S., D. Karlan, and D. Bergan (2009). Does the Media Matter? A Field Experiment
Measuring the Effect of Newspapers on Voting Behavior and Political Opinions. American
Economic Journal: Applied Economics 1 (2), 35–52.
34
Guess, A. (2018). (Almost) Everything in Moderation: New Evidence in Americans’ Online
Media Diets.
Guriev, S., N. Melnikov, and E. Zhuravskaya (2020, 12). 3G Internet and Confidence in
Government. The Quarterly Journal of Economics.
Halberstam, Y. and B. Knight (2016). Homophily, Group Size, and the Diffusion of Political
Information in Social Networks: Evidence from Twitter. Journal of Public Economics 143,
73 – 88.
Hargittai, E. (2015). Is Bigger Always Better? Potential Biases of Big Data Derived from Social
Network Sites. The Annals of the American Academy of Political and Social Science 659,
63–76.
Howard, P. N., A. Duffy, D. Freelon, M. Hussain, W. Mari, and M. Maziad (2011). Opening
Closed Regimes: What Was the Role of Social Media During the Arab Spring? Working
Paper .
Jones, J. J., R. M. Bond, E. Bakshy, D. Eckles, and J. H. Fowler (2017, 04). Social Influence
and Political Mobilization: Further Evidence From a Randomized Experiment in the 2012
U.S. Presidential Election. PLOS ONE 12 (4), 1–9.
Larcinese, V. and L. Miner (2017, June). The Political Impact of the Internet on US
Presidential Elections. STICERD - Economic Organisation and Public Policy Discussion
Papers Series 63, Suntory and Toyota International Centres for Economics and Related
Disciplines, LSE.
Lelkes, Y., G. Sood, and S. Iyengar (2017). The Hostile Audience: The Effect of Access to
Broadband Internet on Partisan Affect. American Journal of Political Science 61 (1), 5–20.
Levy, R. (2021, March). Social Media, News Consumption, and Polarization: Evidence from
a Field Experiment. American Economic Review 111 (3), 831–70.
Littman, J., L. Wrubel, and D. Kerchner (2016). 2016 United States Presidential Election
Tweet IDs.
35
Manacorda, M. and A. Tesei (2016, May). Liberation Technology: Mobile Phones and Political
Mobilization in Africa. CEPR Discussion Papers 11278, C.E.P.R. Discussion Papers.
Martin, G. J. and A. Yurukoglu (2017, September). Bias in Cable News: Persuasion and
Polarization. American Economic Review 107 (9), 2565–2599.
Mosquera, R., M. Odunowo, T. McNamara, X. Guo, and R. Petrie (2020). The Economic
Effects of Facebook. Experimental Economics 23 (2), 575–602.
Müller, K. and C. Schwarz (2019). From Hashtag to Hate Crime: Twitter and Anti-Minority
Sentiment. Working Paper .
Müller, K. and C. Schwarz (2020). Fanning the Flames of Hate: Social Media and Hate Crime.
Journal of the European Economic Association.
Nelson, J. L. and J. G. Webster (2017). The Myth of Partisan Selective Exposure: A Portrait
of the Online Political News Audience. Social Media + Society 3 (3), 2056305117729314.
New York Times (2017). The Fake Americans Russia Created to Influence the Election, By
Scott Shane.
NPR (2020a). Facebook Clamps Down On Posts, Ads That Could Undermine U.S. Presidential
Election.
NPR (2020b). Facebook Keeps Data Secret, Letting Conservative Bias Claims Persist.
NPR (2020c). President-Elect Joe Biden Hits 80 Million Votes In Year Of Record Turnout.
Oliver, J. E., T. Wood, and A. Bass (2016). Liberellas versus Konservatives: Social Status,
Ideology, and Birth Names in the United States. Political Behavior 38 (1), 55–81.
Oster, E. (2019). Unobservable Selection and Coefficient Stability: Theory and Evidence.
Journal of Business & Economic Statistics 37 (2), 187–204.
Pariser, E. (2011). The Filter Bubble: What The Internet Is Hiding From You. Penguin
Books Limited.
36
Pew Research Center (2009). The Internet’s Role in Campaign 2008. Technical report.
Pew Research Center (2011). Social Networking Sites and Our Lives. Technical report.
Pew Research Center (2014). Political Polarization & Media Habits. Technical report.
Pew Research Center (2015). Americans’ Internet Access: 2000-2015. Technical report.
Pew Research Center (2018). For Most Trump Voters, ‘Very Warm’ Feelings for Him Endured.
Technical report.
Pew Research Center (2019a). About One-In-Five Adult Twitter Users in the U.S. Follow
Trump. Technical report.
Pew Research Center (2019b). National Politics on Twitter: Small Share of U.S. Adults
Produce Majority of Tweets. Technical report.
Pew Research Center (2019c). Share of U.S. Adults Using Social Media, Including Facebook,
Is Mostly Unchanged Since 2018. Technical report.
Pew Research Center (2020). Democrats On Twitter More Liberal, Less Focused On Compro-
mise Than Those Not On The Platform. Technical report.
Schmidt, A. L., F. Zollo, M. Del Vicario, A. Bessi, A. Scala, G. Caldarelli, H. E. Stanley, and
W. Quattrociocchi (2017). Anatomy of News Consumption on Facebook. Proceedings of
the National Academy of Sciences 114 (12), 3035–3039.
Solon, G., S. J. Haider, and J. M. Wooldridge (2015). What Are We Weighting For? Journal
of Human resources 50 (2), 301–316.
Spenkuch, J. L. and D. Toniatti (2018, 05). Political Advertising and Election Results. The
Quarterly Journal of Economics 133 (4), 1981–2036.
Statista (2020). Leading countries based on number of Twitter users as of October 2020.
Strömberg, D. (2015). Media and Politics. Annual Review of Economics 7 (1), 173–205.
37
Sunstein, C. R. (2017). # Republic: Divided Democracy in the Age of Social Media. Princeton
University Press.
The Guardian (2018). Revealed: 50 Million Facebook Profiles Harvested for Cambridge
Analytica in Major Data Breach, by Carole Cadwalladr and Emma Graham-Harrison.
The Hill (2016). Stop Blaming Facebook for Trump’s Election Win, By Keith N. Hampton
and Eszter Hargittai.
The New Yorker (2016). Obama Reckons with a Trump Presidency, By David Remnick.
UK Parliament (2019). Disinformation and ‘fake news’: Final Report, By Digital, Culture,
Media and Sport Committee.
Venture Beat (2008). Hitwise: Twitter Traffic Is, in Fact, Going up but Still Not Big, by Eric
Eldon.
Weber, I., V. R. K. Garimella, and A. Batayneh (2013). Secular vs. Islamist Polarization in
Egypt on Twitter. In Proceedings of the 2013 IEEE/ACM International Conference on
Advances in Social Networks Analysis and Mining, ASONAM ’13, New York, NY, USA, pp.
290–297. Association for Computing Machinery.
Wired (2016). Here’s How Facebook Actually Won Trump the Presidency, By Issie Lapowsky.
Zhuravskaya, E., M. Petrova, and R. Enikolopov (2020). Political Effects of the Internet and
Social Media. Technical report.
38
Figure 1: Twitter Reach by Party
(a) Average # of tweets (per account) (b) Average # of followers (per account)
5,000 600,000
4,000
400,000
3,000
2,000
200,000
1,000
0 0
Democrat Republican Democrat Republican
(c) Average # of likes (per tweet) (d) Average # of retweets (per tweet)
800
400
600
300
400
200
200 100
0 0
Democrat Republican Democrat Republican
Notes: This figure plots data on the Twitter reach of Congress members. The sample includes all 901
senators and House representatives who were in office between 2007 and 2019 for whom we could identify
a Twitter account. For each account, we plot the average number of tweets and followers, and the
average number of “likes” and retweets of their tweets. Appendix Figure A.2 replicates the figure using
medians instead of averages. The data were collected from Twitter in November 2019.
39
Figure 2: Republican Vote Share in the 2016 Presidential Election
1st Quintile
2nd Quintile
3rd Quintile
4th Quintile
5th Quintile
Notes: This map plots the geographical distribution of the Republican two-party vote share in the 2016
presidential election.
40
Figure 3: Twitter Usage and Identifying Variation
(a) Twitter Usage per Capita
1st Quintile
2nd Quintile
3rd Quintile
4th Quintile
5th Quintile
Notes: This map plots the proxy for social media usage based on data from Twitter and the
identifying variation of our instrument. Panel (a) plots quintiles of the number of Twitter users
per capita. Panel (b) plots the three types of counties relevant for our identification strategy: 1)
the 47 counties with SXSW followers that joined Twitter both in March 2007 and the “pre-period”
(light red); 2) the 108 counties with SXSW followers that joined in March 2007, but none in the
“pre-period” (dark red); and 3) the 20 counties with SXSW followers that joined in the “pre-period,”
but none in March 2007 (blue).
41
Figure 4: South by Southwest (SXSW) 2007 and the Spread of Twitter
(a) Twitter Activity Around SXSW 2007
400 40,000
SXSW tweets (left axis) All tweets (right axis)
300 30,000
200 20,000
100 10,000
0 0
January February March April May
0.10
0.05
0.00
-0.05
3
9
10
11
12
13
14
ek
ek
ek
ek
ek
ek
ek
ek
ek
ek
ek
ek
we
we
we
we
we
we
we
we
we
we
we
we
07
07
07
07
07
07
07
07
07
07
07
07
20
20
20
20
20
20
20
20
20
20
20
20
Notes: Panel (a) plots the total number of tweets and the number of tweets containing the term
“SXSW” over time, smoothed using a 7-day moving average. Panel (b) plots the estimates of βτ from
the panel event study regression tweetsct = τ βτ SXSWcM arch2007 ×1(t = τ )+ τ δτ SXSWcP re ×
P P
1(t = τ ) + θc + γt + εct where tweetsct is the log of (one plus) the number of tweets in county c on
week t, SXSWcM arch2007 is the logarithm of (one plus) the number of SXSW followers in county
c that joined Twitter on March 2007 and SXSWcP re is a similarly defined variable for followers
that joined Twitter before March 2007. We standardize the variables to have a mean of zero and
standard deviation of one. The whiskers represent 95% confidence intervals based on standard
errors clustered by state.
42
Figure 5: First Stage – South by Southwest (SXSW) and Twitter Usage
0.50
Log(Twitter users), residualized
0.25
0.00
-0.25
-0.50
43
Figure 6: South by Southwest, Twitter, and the Republican Vote Share
2016 Election
(a) Reduced Form (b) OLS
0.02
0.02
Republican vote share, 2016 (residualized)
0.00
0.00
-0.01
-0.01
-0.02
-0.02
-1.00 -0.50 0.00 0.50 1.00 -1.50 -1.00 -0.50 0.00 0.50 1.00
2020 Election
(c) Reduced Form (d) OLS
0.04
Republican vote share, 2020 (residualized)
0.02
0.01
0.02
0.00
0.00
-0.01
-0.02
-0.02
-1.00 -0.50 0.00 0.50 1.00 -1.50 -1.00 -0.50 0.00 0.50 1.00
Notes: Panel (a) presents a binned scatter plot of the relationship between the Republican vote share in
the 2016 presidential election and the number of SXSW followers who joined Twitter in March 2007.
Variables are residualized with respect to SXSW followers who joined before March 2007, population
deciles, Census region fixed effects, as well as geographical, demographic, socioeconomic, China shock,
and 1996 election control variables. The figure is constructed by dividing the x-axis variable into 40
equal-sized bins and plotting the average values of both variables in each bin. The fitted line is estimated
using the unbinned data. Panels (b) and (d) replicate the exercise using Twitter users in 2014-2015 as
the x-axis variable. Panel (c) and (d) show results for the 2020 election.
44
Figure 7: Twitter and Presidential Elections – Reduced Form
(a) Changes in Republican Vote Share
0.005
-0.005
-0.010
-0.015
0.015
Point estimate on Log(SXSW followers, March 2007)
0.010
0.005
0.000
-0.005
-0.010
2000-04 2000-08 2000-12 2000-16 2000-20
Notes: These figures plot reduced form estimates β̂ 0 from county-level regressions as in equation (2).
They measure the effect of Log(1 + SXSW followers, March 2007), while controlling for Log(1 + SXSW
followers, Pre), on changes in the Republican vote share in presidential elections relative to the year
2000 in Panel (a), and changes in the ratio of voter turnout to voting-age population relative to 2000 in
Panel (b). All regressions control for population deciles and Census region fixed effects, and the full
set of controls (as in columns 5 and 10 of Table 2). On Panel (a), the estimates for 2004, 2008, and
2012 are jointly statistically insignificant (p-value=0.398). Further, the average effect in 2016 and 2020
is statistically distinct from the average effect for 2004, 2008, and 2012 (p-value=0.001). Regressions
are weighted by turnout in the 2000 presidential election. Whiskers represent 95% confidence intervals
based on standard errors clustered by state.
45
Figure 8: Twitter and Congressional Election Results – Reduced Form
(a) House Elections
0.06
Point estimate on Log(SXSW followers, March 2007)
0.04
0.02
0.00
-0.02
2000-02 2000-04 2000-06 2000-08 2000-10 2000-12 2000-14 2000-16 2000-18 2000-20
0.04
Point estimate on Log(SXSW followers, March 2007)
0.02
0.00
-0.02
-0.04
1996-02 1998-04 2000-06 1996-08 1998-10 2000-12 1996-14 1998-16 2000-18 1996-20
Notes: These figures plot reduced form estimates β̂ 0 from county-level regressions as in equation (2).
They measure the reduced form effect of Log(1 + SXSW followers, March 2007), while controlling for
Log(1 + SXSW followers, Pre), on the Republican vote share in House and Senate elections since 2000.
For House elections in Panel (a), the dependent variable is the change in the Republican vote share since
2000. For Senate elections in Panel (b), the dependent variable is the change in the Republican vote
share from six, twelve, or eighteen years ago (to accommodate senators’ 6-year terms). All regressions
control for population deciles and Census region fixed effects and the full set of controls (as in columns 5
and 10 of Table 2). Regressions are weighted by turnout in the 2000 presidential election. The whiskers
represent 95% confidence intervals based on standard errors clustered by state.
46
Figure 9: Twitter’s Partisan Slant Across Presidential Elections
(a) Tweets about Republican Presidential Candidates
24%
600
26%
Number of Likes (in millions) 50%
500
400
300
200 15%
26%
100 27% 59%
11%
0
62%
Election 2012 Election 2016 Election 2020
600
Number of Likes (in millions)
500
400
300 18%
37%
200 45%
14%
100 29%
44%
23%
42%
0
48%
Election 2012 Election 2016 Election 2020
Notes: These figures present the number of “likes” received by tweets that contain the last name of
the candidates in the 2012, 2016 and 2020 presidential elections, depending on whether the tweet was
classified as having a Republican, Democratic, or neutral slant. We classify the slant of a tweet based on
the Twitter network of the user who sent the tweet. If the user follows more Democratic than Republican
Congress members, they will be classified as a Democrat, and vice versa. Users who follow an equal
number of Democrats and Republican or no Congress members are classified as neutral.
47
Table 1: South by Southwest 2007 and Twitter Usage
48
Table 2: Twitter and the 2016/2020 Republican Vote Share
Dep. var.: Republican vote share in 2016 Dep. var.: Republican vote share in 2020
(1) (2) (3) (4) (5) (6) (7) (8) (9) (10)
Panel A: OLS
Log(Twitter users) -0.065*** -0.067*** -0.013*** -0.011*** -0.007** -0.058*** -0.064*** -0.012*** -0.011*** -0.008***
(0.009) (0.008) (0.004) (0.003) (0.003) (0.009) (0.009) (0.004) (0.003) (0.003)
Panel B: Reduced form
Log(SXSW followers, March 2007) -0.053*** -0.058*** -0.019*** -0.014*** -0.011*** -0.046*** -0.055*** -0.017*** -0.013*** -0.011**
(0.011) (0.012) (0.005) (0.004) (0.004) (0.009) (0.010) (0.006) (0.005) (0.005)
Log(SXSW followers, Pre) -0.021 -0.003 -0.000 -0.002 0.001 -0.022 -0.005 -0.002 -0.004 -0.001
(0.016) (0.013) (0.006) (0.006) (0.004) (0.016) (0.013) (0.007) (0.007) (0.005)
Panel C: 2SLS
Log(Twitter users) -0.072*** -0.085*** -0.034*** -0.027*** -0.021*** -0.064*** -0.080*** -0.031** -0.025*** -0.020**
(0.016) (0.018) (0.010) (0.008) (0.008) (0.015) (0.017) (0.011) (0.009) (0.009)
Log(SXSW followers, Pre) -0.014 0.007 0.002 -0.001 0.002 -0.015 0.004 -0.000 -0.002 0.000
49
(0.020) (0.016) (0.007) (0.006) (0.005) (0.020) (0.015) (0.008) (0.007) (0.006)
Population deciles Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes
Census region FE Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes
Geographical controls Yes Yes Yes Yes Yes Yes Yes Yes
Demographic controls Yes Yes Yes Yes Yes Yes
Socioeconomic controls Yes Yes Yes Yes Yes Yes
China shock controls Yes Yes Yes Yes
1996 election control Yes Yes
Observations 3,065 3,065 3,064 3,064 3,064 3,065 3,065 3,064 3,064 3,064
Mean of DV 0.46 0.46 0.46 0.46 0.46 0.47 0.47 0.47 0.47 0.47
Robust F-stat. 69.57 74.65 106.08 118.21 121.18 69.57 74.65 106.08 118.21 121.18
Notes: This table presents county-level regressions where the dependent variable is the Republican vote share in the 2016 or 2020 presidential election. Log(SXSW followers,
March 2007) is the number of Twitter users (in logs, with 1 added inside) who joined in March 2007 and follow South by Southwest (SXSW). SXSW followers, Pre is the
number of SXSW followers who registered at some point in 2006, defined similarly. Twitter users are the number of users in 2014-2015. Regressions include the indicated
control variables (see the Online Appendix for their descriptions). The first-stage regressions for 2SLS results (Panel B) are presented in Table 1, with the F-stat for the
excluded instrument in the bottom row. The “Oster-δ” for gauging the possible role of unobservables in driving coefficient stability is 6.93 (comparing column 5 to 1) and
7.02 (comparing column 10 to 6); see text for further details. Observations are weighted by turnout in the 2000 presidential election. Standard errors in parentheses are
clustered by state. *** p < 0.01, ** p < 0.05, * p < 0.1.
Table 3: Twitter and the Republican Vote Share, 2000-2020
50
Table 4: Twitter and Individuals’ Vote Decisions in 2016/2020
51
Table 5: Twitter and Vote Shares in the 2016 Republican Primaries
52
Table 6: Twitter and Candidate Approval during the 2016 Primaries
53
Online Appendix
This appendix presents further details on data construction, the SXSW festival, additional
robustness exercises, further results, and the LATE extrapolation:
Table A.1: List of Congress Members with the Most Twitter Followers
1
Figure A.1: Validation of Twitter usage measure
(a) Twitter Usage over Time
80
Statista (right axis)
5
Number of Twitter users (in millions)
60
4
3
40
20
1 Gesis (left axis)
0 0
2007q1 2009q1 2011q1 2013q1 2015q1 2017q1 2019q1
12
Log(# of households using Twitter)
10
2 4 6 8 10
Log(Twitter users)
Notes: This graph shows two validation exercises for the Twitter usage measure in the Gesis data
(Kinder-Kurlanda et al., 2017). Panel (a) plots the number of Twitter users in the Gesis data
and the number of active monthly users reported by Statista based on Twitter’s own reporting.
Panel (b) plots the percentiles of the number of Twitter users in the Gesis data at the county-level
against the number of users based on the GfK Media Survey.
2
Figure A.2: Twitter Reach by Party (Median)
(a) Median # of tweets (per account) (b) Median # of followers (per account)
4,000
25,000
3,000 20,000
15,000
2,000
10,000
1,000
5,000
0 0
Democrat Republican Democrat Republican
(c) Median # of likes (per tweet) (d) Median # of retweets (per tweet)
30 150
20 100
10 50
0 0
Democrat Republican Democrat Republican
Notes: This figure plots data on the Twitter reach of Congress members. The sample includes all 901
senators and House representatives who were in office between 2007 and 2019 for whom we could identify
a Twitter account. For each account, we plot the median number of tweets and followers, and the median
number of “likes” and retweets of their tweets. The data were collected from Twitter in November 2019.
3
Figure A.3: News Reports About Social Media
4000
Twitter
Number of news reports on social media
1000
0
1
1
m
m
04
06
08
10
12
14
16
18
20
20
20
20
20
20
20
20
Notes: This graph plots the number of times the terms “Twitter” and “Facebook” are
mentioned in USA Today, The Washington Post, The New York Post, and The New
York Times based on data from Nexis.
4
Table A.2: Summary Statistics (County-Level)
5
Table A.3: Summary Statistics (2016 CCES Individual-Level)
6
Table A.5: Instrument Balancedness
7
Table A.6: Description of Main Variables
8
SXSW followers, Pre The total number of Twitter users following the @SXSW account in each county that joined Twitter Search API
Twitter at any point in 2006.
County-level control variables
Geographical controls Include the distance (in miles) from the country centroid from Austin (TX), Chicago, NYC, U.S. Census Tigerline File
San Francisco, and Washington, DC, population density, and the logarithm of the land area
for each county.
Demographic controls Include the share of people in the age buckets 20-24, 25-29, 30-34, 40-44, 45-49 and 50+, as
well as the percentage change in county population between 2000 and 2016.
Socioeconomic controls Include a county’s poverty rate; unemployment rate; the share of the population employed U.S. Census/BLS/SimplyAnalytics
in construction/real estate, manufacturing, and information technology; the share of adults
over 25 with at least a high school degree or some college education; as well as prime time
TV viewership and the share of Fox News viewership.
China shock controls Include the measure of import competition from China, routine tasks, and offshorability from
Autor et al. (2013).
Election control Contains the vote share of the Republican party in the 1996 presidential election. Dave Leip Election Atlas
Table A.7: Search Terms Used to Create a
Proxy for Total Tweets
9
A.1. Additional Details on the Logistic Regression Classifier
We train a seperate machine learning classifier for each of the three election years in our data
using the Python sci-kit package (Pedregosa et al., 2011). These classifiers help us to determine
whether tweets are more likely to be sent by Democratic-leaning or Republican-leaning users.
The classification process starts with the preparation of the underlying Twitter data. The
inputs are the text of each of the 4,300,579 tweets from U.S. Congress members. To focus on
election-related tweets, we restrict the sample to tweets that were sent either in the election
year or in the year leading up to the election (e.g. 2011 and 2012 for the 2012 election) and
mention at least one of the presidential candidates.
The target variable y for the classifier is an indicator variable equal to one for tweets
sent by Republicans and zero otherwise. The feature matrix X for the classifier are created by
count-vectorizing the texts of the tweets. In other words, we transform the text of the tweets
into n × v matrix, where n is the number of tweets and v is the number of unique 1,2-grams
that occur in the tweets. In preparation for this step, we removed common words (stopwords),
links, and special characters from the tweets. Additionally, we reduced the words in the
tweets to their morphological roots using a lemmatizer, which improves the performance of
the classifier. As an example, the lemmatizer changes words like “walking” and “walked” to
“walk”. Lastly, we reweight the n-grams v of tweet i using term frequency–inverse document
frequency (tf-idf):
1+T
tf idf (fi,v ) = (1 + ln(fi,v ) · (ln( ) + 1) (A.1)
1 + dv
where dv is the number of tweets n-gram v appears in. This reweighting reduces the importance
of words that appear frequently in many tweets, which help little to discriminate between
tweets. The tf-idf vectorizer also normalizes the feature matrix by its L2-norm.
The vector y and the matrix X then serve as the input for a L2 regularized logistic
regression classifier. The optimization of the classifier involves the minimization of the
following cost function54 :
n
X 1
min C log(exp(−yi (XiT w + c) + 1) + wT w (A.2)
w,c
i=1
2
where w are the weights (coefficients) of the logistic regression, c is a constant (intercept), and
C is the inverse of the regularization strength. Larger values of C imply weaker regularization.
For C −→ ∞, the classifier converges towards a normal logistic regression. As is standard in
most machine learning applications, we choose the optimal regularization strength C using
54
Note that this formulation of the cost function assumes that yi takes values −1; 1. We use this formulation
in line with the sci-kit documentation.
10
10-fold cross-validation. This involves randomly splitting the training data into ten equal
slices. Nine of the ten slices are then used to train the classifier, while the out-of-sample
performance is evaluated against the remaining slice using F1-scores.
The final classifiers achieves an out-of-sample F1-score of 0.916 in 2012, of 0.843 in
2016 and of 0.904 in 2020. The classifiers, therefore, accurately predict the party affiliation
of Congress members. We then take these classifiers and apply it to the universe of tweets
sent during the 2012, 2016 and 2020 presidential election. For each tweet in the election
data, the classifiers provides us with a predicted class (either Democrat or Republican) and
a probability for this class label. To avoid that our results are driven by tweets for which
the classifier is “uncertain,” we code tweets with a predicted class probability below 60% as
neutral. This adjustment has no bearing on our findings. In spirit, this approach is similar
to the work of Gentzkow and Shapiro (2010). While they identify expressions that are more
frequently used by Democrats and Republicans by hand, we use a machine learning classifier
to identify n-grams in the tweets of Congress members that help us to differentiate between
the two parties.
We visualize the most predictive n-grams identified by the classifiers for each election
cycle in Figure A.4. Overall, the classifiers identifies words, hashtags, and Twitter handles
that are intuitively associated with a Republican slant for each election year. Among the
most predictive term are the hashtags “tcot” (Top Conservatives on Twitter) and “maga”
(Make America great again) and particularly in 2020 many references to Donald Trump’s
Twitter account (“realdonaldtrump”).
11
Figure A.4: Most Predictive Terms Of “Republican” Tweets by Election
(a) 2012 Election
Notes: This word cloud plots the n-grams most predictive of tweets sounding like those of Republican
Congress members, as identified by the logistic regression classifier for each election cycle. The size of
the word represents the magnitude of the coefficients.
12
B Additional Details on the SXSW Festival
Notes: This screenshot shows the full post of Twitter co-founder Evan Williams posted on Quora
on January 4, 2011 describing the role of the SXSW festival in the platform’s rise to popularity
(Quora, 2011).
13
Figure B.2: Number of SXSW Follower Joining each Month
1,000
500
0
06
06
06
em 06
cto 06
em 6
em 06
nu 6
M 07
07
0
y0
ov er 0
r0
Fe ry 0
ch
ril
ay
ne
D ber
ry
ch
s
be
be
l
Ju
gu
a
ua
Ap
M
ar
Ju
ar
Au
br
M
Ja
pt
ec
O
Se
Notes: The figures shows the number of SXSW Follower joining in each month.
14
Figure B.3: Additional Evidence for the Impact of the SXSW Festival
(a) Long-term Effects of the 2007 SXSW on Twitter
Adoption
0.004
0.002
0.000
2006q2 2009q1 2012q1 2015q1
0.6
Share of Twitter users with SXSW connection
0.4
0.2
0.0
2007q1 2009q1 2011q1 2013q1 2015q1
Notes: The figures provide evidence on the long-term impact of the SXSW festival on Twitter
usage across
P the United States. Panel (a) plots P the βτ from the panel event study regression
usersct = τ βτ SXSWcM arch2007 × 1(t = τ ) + τ δτ SXSWcP re × 1(t = τ ) + θc + γt + εct where
usersct is the number of Twitter users per capita in county c on quarter t, SXSWcM arch2007 is the
logarithm of (one plus) the number of SXSW followers in county c who joined Twitter in March
2007 and SXSWcP re is a similarly defined variable for followers who joined Twitter before March
2007. We standardize the variables to have a mean of zero and standard deviation of one. The
whiskers represent 95% confidence intervals based on standard errors clustered by state, where
2006q4 serves as excluded period. While the confidence intervals for 2006q2 and 2006q3 cannot be
seen, they include zero. Panel (b) plots the share of Twitter who either follow SXSW or follow an
user that follows SXSW.
15
Table B.2: Are Twitter Users in Counties With SXSW Followers Different?
16
C Additional Robustness Checks
17
Figure C.1: Twitter and the Republican Vote Share, 1924-1996 (Reduced Form)
1924
1928
1932
1936
1940
1944
1948
1952
1956
1960
1964
1968
1972
1976
1980
1984
1988
1992
1996
Notes: This figure plots reduced form estimates β̂ 0 from county-level regressions as in equation
(2). These estimates reflect the correlation of Log(1 + SXSW followers, March 2007) with the
Republican vote share in presidential elections while controlling for Log(1 + SXSW followers, Pre).
All regressions control for population deciles and Census region fixed effects, and the full set of
controls except 1996 Election controls (same as columns 4 and 9 of Table 2). Regressions are
weighted by turnout in the 2000 presidential election. Whiskers represent 95% confidence intervals
based on standard errors clustered by state.
18
Table C.2: Placebo Tests for Other Festivals
19
Table C.3: Twitter and the Republican Vote Share – Robustness
20
Table C.4: Twitter and Changes in Republican Vote Share, 2004-2020
21
Table C.5: Time-Varying Twitter Usage and Changes in Vote Shares
22
Table C.6: Twitter and the Ross Perot Vote
23
Table C.7: Twitter and Vote Shares in Democratic Primaries
Table C.8: Twitter and Vote Decisions in the 2016 CCES – Robustness
24
D Further Results
25
Table D.2: Twitter and Changes in Voter Turnout, 2004-2020
26
Table D.3: Twitter and Congressional Elections – 2SLS Estimates
27
1996-02 1998-04 2000-06 1996-08 1998-10 2000-12 1996-14 1998-16 2000-18 2000-20
(1) (2) (3) (4) (5) (6) (7) (8) (9) (10)
Log(Twitter users) 0.019 0.015 0.007 0.008 -0.011 0.021 -0.035 -0.013 -0.008 -0.019
(0.023) (0.024) (0.021) (0.021) (0.018) (0.031) (0.026) (0.026) (0.022) (0.018)
Observations 2,183 1,985 1,832 2,183 1,985 1,832 2,183 1,985 1,832 2,183
Mean of DV 0.01 -0.03 -0.06 -0.05 0.01 -0.06 0.03 -0.07 -0.10 -0.00
Robust F-stat. 79.24 85.91 72.85 79.24 85.91 72.85 79.24 85.91 72.85 79.24
Notes: This table presents results estimated using 2SLS, as in equation (3). For House elections in Panel A, the dependent
variable is the change in the Republican vote share since 2000. For Senate elections in Panel B, the dependent variable is
the change in the Republican vote share from six, twelve, or eighteen years ago (to accommodate senators’ 6-year terms).
Log(Twitter users) is instrumented using the number of users who started following SXSW in March 2007 (in logs with 1 added
inside). All regressions control for the (log) number of SXSW followers that joined Twitter at some point in 2006, population
deciles and Census region fixed effects and the full set of controls (as in columns 5 and 10 of Table 2). Regressions are weighted
by turnout in the 2000 presidential election. Standard errors in parentheses are clustered by state. *** p < 0.01, ** p < 0.05, *
p < 0.1.
Table D.4: Twitter and Individuals’ Vote Decisions in 2016 (Age Splits)
28
Table D.5: Twitter and the Republican Vote Share in Swing and Safe Counties
29
Figure D.1: Twitter’s Partisan Slant (Tweet Measure)
(a) Tweets about Republican Presiden- (b) Tweets about Democratic Presiden-
tial Candidates tial Candidates
34%
28%
250 37% 250
Number of Tweets (in millions)
150 150
24% 31%
100 27% 100 27%
49% 42%
22%
50 35% 50 38% 39%
14% 36% 39%
51% 36%
0 0
Election 2012 Election 2016 Election 2020 Election 2012 Election 2016 Election 2020
Democratic Slant Republican Slant Neutral Slant Democratic Slant Republican Slant Neutral Slant
Notes: These figures present the number of tweets (as opposed to the number of “likes” of such tweets in
Figure 9) that contain the last name of the candidates in the 2012, 2016 and 2020 presidential elections,
depending on whether the tweet was classified as having a Republican (instead of Democratic) slant.
We classify the slant of a tweet based on the Twitter network of the user who sent the tweet. If the user
follows more Democratic than Republican Congress members, they will be classified as a Democrat, and
vice versa. Users who follow an equal number of Democrats and Republican or no Congress members
are classified as neutral.
30
Figure D.2: Twitter’s Partisan Slant (Text-Based Classifier)
(a) Likes for Tweets about Republican (b) Likes for Tweets about Democratic
Presidential Candidates Presidential Candidates
11%
600
38% 600
51%
Number of Likes (in millions)
400 400
Democratic Slant Republican Slant Neutral Slant Democratic Slant Republican Slant Neutral Slant
(c) Tweets about Republican Presiden- (d) Tweets about Democratic Presiden-
tial Candidates tial Candidates
12%
53%
250 35% 250
Number of Tweets (in millions)
200 200
150 150
15% 7%
100 30% 100 47%
55% 46%
11%
50 9% 50 14% 51%
12% 40% 38%
78% 46%
0 0
Election 2012 Election 2016 Election 2020 Election 2012 Election 2016 Election 2020
Democratic Slant Republican Slant Neutral Slant Democratic Slant Republican Slant Neutral Slant
Notes: These figures present the number of “likes” received by tweets, or the number of tweets, that
contain the last name of the candidates in the 2012, 2016 and 2020 presidential elections, depending on
whether the tweet was classified as having a Republican (instead of Democratic) slant. We classify the
slant of a tweet based on similarity in the language to that of a congressional Republican or Democrat,
using a L2 regularized logistic regression classifier using the tweets sent by Congress members. Optimal
normalization strength is chosen using 10-fold cross-validation. Tweets with a predicted class probability
below 60% are coded as neutral. See Appendix A.1. for details.
31
Figure D.3: Specification Curve
(a) 2016 Presidential Election Results
0
-.02
Coefficient Twitter Usage (IV)
-.04
-.06
-.08
-.1
Census_Region
Socioeconomic
Demographic
China_Shock
Geographic
Population
Election
Ethnic
TV
-.12
Population
Census_Region
Geographic
pad
Demographic
Ethnic
Socioeconomic
TV
China_Shock
Census_Region
Socioeconomic
Demographic
China_Shock
Geographic
Population
Election
Ethnic
TV
Notes: These figures plot the 2SLS estimates and 95% confidence intervals from a regression of the
Republican vote share in 2016 on Log(T witter users), instrumented with SXSW M arch 2007 . All
regressions include population deciles, census region fixed effects, and SXSW P re . The combination
of the other included control variables is shown at the bottom; filled circles mean a set of controls
was included. The baseline specification with all controls is marked by the vertical line.
32
Figure D.3: Specification Curve
(b) 2020 Presidential Election Results
0
-.02
Coefficient Twitter Usage (IV)
-.04
-.06
-.08
-.1
Census_Region
Socioeconomic
Demographic
China_Shock
Geographic
Population
Election
Ethnic
TV
Population
Census_Region
Geographic
pad
Demographic
Ethnic
Socioeconomic
TV
China_Shock
Census_Region
Socioeconomic
Demographic
China_Shock
Geographic
Population
Election
Ethnic
TV
Notes: These figures plot the 2SLS estimates and 95% confidence intervals from a regression of the
Republican vote share in 2020 on Log(T witter users), instrumented with SXSW M arch 2007 . All
regressions include population deciles, census region fixed effects, and SXSW P re . The combination
of the other included control variables is shown at the bottom; filled circles mean a set of controls
was included. The baseline specification with all controls is marked by the vertical line.
33
E Additional Details on the Extrapolation for the Av-
erage Treatment Effect
Andrews and Oster (2019) show how selection into participating in an experiment can be
used to make extrapolations regarding the external validity of an experiment. They illustrate
their approach using the experiment from Bloom et al. (2015) within a Chinese call centre
in which workers were asked to volunteer for a work-from-home program. 50% of workers
volunteered and were then randomly assigned to either treatment and control group. Given
this, the measured treatment effect from the experimental sample might be different from the
average treatment effect for the population as a whole, since the volunteers likely receive a
higher utility from the work-from-home program.
When a set of covariates X is observed for both the “experimental sample” and “popu-
lation,” Andrews and Oster (2019) provide a procedure that uses effect heterogeneity based
on X estimated within the experimental sample to extrapolate to the average treatment effect
for the “population.”
We build on their procedure and argue that we can similarly use heterogeneity in the
treatment effect within the counties that “identify” our results to extrapolate the treatment
effect to all other counties in the US. Column (5) of Table C.3 show that we obtain similar
estimates to our baseline when we only compare counties with SXSW followers that joined
Twitter in March 2007 to counties with followers that joined in the pre-period, while excluding
those counties in neither group. We can use this subsample of counties as the “experimental
sample,” and extrapolate effects to the “population” of all other counties.
Since Andrews and Oster (2019) approach is designed for a binary treatment, we adjust
our regression framework by defining a treatment indicator variable equal to 1 for counties
with SXSW followers who joined in March 2007 and 0 for the counties with followers that
joined in the pre-period. We estimate the treatment effect for the subsample of counties
that do not have zero SXSW followers in both periods using the regression specification
yc = α + β · 1[SXSWcM arch2007 > 0] + c . The resulting treatment effect estimate is −0.075,
which is similar Table 2 Panel B column (1).55 We then perform a linear prediction of this
treatment effect based all observable variables in Table A.5 within this subsample. The
resulting predicted treatment effect is −0.085. Last, we extrapolate the treatment effect for
the rest of US counties. Based on the variation in observable characteristics we would predict
an ATE of −0.218 for the US overall.
55
Note that our regression specification does not include controls, as in the Andrews and Oster (2019)
approach. Moreover, partialling out the controls and applying the Frisch-Waugh theorem was not feasible
since the residualized treatment would no longer be binary.
34
Note that this extrapolation is based on adjusting our reduced-form estimates to use a
binary indicator variable for treatment thus the coefficients are not directly comparable to
our baseline estimates. The approach further assumes quasi-random treatment assignment
with in the counties with SXSW variation. Taken this into account, the extrapolation should
therefore be viewed as suggestive, but confirming the notion that the effect for all US counties
would be larger since the more urban counties for which we have variation in our instrument
tend to be Democratic strongholds, and thus likely have fewer independents and moderate
Republicans, for which we find the largest persuasion effects (in survey data).
35