Does Twitter language reliably predict heart disease? A commentary on Eichstaedt et al. (2015a)

View article
PeerJ
An earlier version of the present article, available in preprint form at https://psyarxiv.com/dursw, stated that we had not been able to obtain Eichstaedt et al.’s (2015a) code because it had not been made available in the same OSF repository as the data. We are happy to acknowledge here that Eichstaedt and colleagues had in fact made their code available on the website of their Differential Language Analysis ToolKit (DLATK) software project, a fact that they have now documented in their recent preprint (Eichstaedt et al., 2018). We followed the installation instructions for DLATK and were able to reproduce the analyses described by Eichstaedt et al. (2015a, p. 161) under the heading of “Predictive models”.
We report 95% CIs here for consistency with Eichstaedt et al. (2015a); however, given their use of a Bonferroni-corrected significance threshold, it could be argued that Eichstaedt et al. should have reported 99.9975% CIs.
In their recent preprint, Eichstaedt et al. (2018) noted that the relation between self-harm and positive Twitter language disappeared when they added measures of county-level rurality and elevation as covariates. We do not find this surprising. Our point is not that Twitter language actually predicts county-level mortality from suicide (or AHD); rather, it is that with a sufficient number of predictors, combined with unreliability in the measurement of these, one can easily find spurious relations between variables (cf. Westfall & Yarkoni, 2016).
Examination of Eichstaedt et al.’s (2015a) data set shows that for 267 counties (19.8%), less than 100,000 words were included in the database, which corresponds to around 5,000 tweets. However, no indication is available of the number of unique Twitter users per county.
The spelling “nigga” is often used by African Americans in a neutral or positive sense; the form “nigger” is the one typically used by members of other groups as a racial slur (Goudet, 2013).
We assume that the apparent omission of “Jew[s]” and “Muslim[s]” was motivated by concerns that at least some of the tweets mentioning these words might be expressing hostility towards these groups.
It appears that the relative size of the words in Eichstaedt et al.’s (2015a, Fig. 1) word clouds is determined by the relative frequency of these words in the Facebook data from which the topics were derived, and does not represent the prevalence of these words in the Twitter data.
Eichstaedt et al. (2018) have recently explained that the exclusion of love from their positive relationships dictionary in their earlier article (Eichstaedt et al., 2015a) was the result of discussions with a reviewer of that article.
More precisely, 200 counties (32.9%) had a discrepancy of three or more color intervals, while 123 counties (20.2%) had a discrepancy of six or more. Assuming for simplicity that rounding error is uniformly distributed, a difference of three intervals corresponds to a mean difference of 21.4 percentage points, and a difference of six intervals to a mean difference of 42.8 percentage points. Thus, even with the extremely conservative simplifying assumption that “three (six) or more color intervals” actually means “exactly three (six) intervals,” the mean discrepancy across these counties is ((200 × 21.4) + (123 × 42.8)) / 323 = 29.5 percentage points. Note also that the possible extent of the discrepancy is bounded at between seven and 13 color intervals, depending on the relative positions of the two counties along the 1–14 scale.
Eichstaedt et al. (2015a) included a disclaimer about causality on p. 166 of their article. However, we feel that this did not adequately compensate for some of their language elsewhere in the article, such as “Local communities create [emphasis added] physical and social environments that influence [emphasis added] the behaviors, stress experiences, and health of their residents” (p. 166; both of the italicized words here seem to us to imply causation at least as strongly as our word “exerts”), and “Our approach . . . could bring researchers closer to understanding the community-level psychological factors that are important for the cardiovascular health of communities and should become the focus of intervention” (p. 166, seemingly implying that an intervention to change these psychological factors would be expected to lead to a change in cardiovascular health).
For example, using data from the CDC for the 2009–2010 period, county-level mortality from assault is strongly correlated with county-level mortality from cancer (r = .55), but completely uncorrelated with county-level mortality from AHD (r = .00). There seems to be no obvious theoretical explanation for these results.

Main article text

 

Introduction

Issues related to the idea of psychological causes of AHD

Issues related to the etiology of AHD

Issues related to the secondary analysis of data collected online

Issues associated with considering counties as communities

[I]n New York County, New York, . . . neighborhoods range from the Upper East Side and SoHo to Harlem and Washington Heights. . . . [I]n San Mateo County, California, . . . neighborhoods range from the Woodside estates of Silicon Valley billionaires to the Redwood City bungalows of Mexican immigrants. (Abrams & Fiorina, 2012, p. 206)

Method

Results

Variability in ICD-10 coding of cause of death

Use of mortality from AHD as the outcome variable

Bias caused by selection of counties

Problems associated with county-level aggregation of data

Apparent censorship of the Twitter data

Potential sources of bias in the “Topics” database

Flexibility in interpretation of dictionary data

Comparison of Twitter-based and “traditional” prediction models

How similar are the comparative maps?

Discussion

Conclusions

Additional Information and Declarations

Competing Interests

James C. Coyne is an Academic Editor for PeerJ.

Author Contributions

Nicholas J.L. Brown conceived and designed the experiments, performed the experiments, analyzed the data, contributed reagents/materials/analysis tools, prepared figures and/or tables, authored or reviewed drafts of the paper, approved the final draft.

James C. Coyne authored or reviewed drafts of the paper, approved the final draft.

Data Availability

The following information was supplied regarding data availability:

Brown, Nicholas JL. 2018. “Reanalysis of Eichstaedt et al. (2015: Twitter/heart disease).” Open Science Framework. February 12. https://osf.io/g42dw/.

Funding

The authors received no funding for this work.

6 Citations 5,241 Views 545 Downloads

Your institution may have Open Access funds available for qualifying authors. See if you qualify

Publish for free

Comment on Articles or Preprints and we'll waive your author fee
Learn more

Five new journals in Chemistry

Free to publish • Peer-reviewed • From PeerJ
Find out more