English verb regularization in books and tweets

The English language has evolved dramatically throughout its lifespan, to the extent that a modern speaker of Old English would be incomprehensible without translation. One concrete indicator of this process is the movement from irregular to regular (-ed) forms for the past tense of verbs. In this study we quantify the extent of verb regularization using two vastly disparate datasets: (1) Six years of published books scanned by Google (2003–2008), and (2) A decade of social media messages posted to Twitter (2008–2017). We find that the extent of verb regularization is greater on Twitter, taken as a whole, than in English Fiction books. Regularization is also greater for tweets geotagged in the United States relative to American English books, but the opposite is true for tweets geotagged in the United Kingdom relative to British English books. We also find interesting regional variations in regularization across counties in the United States. However, once differences in population are accounted for, we do not identify strong correlations with socio-demographic variables such as education or income.


I. INTRODUCTION
Human language reflects cultural, political, and social evolution. Words are the atoms of language. Their meanings and usage patterns reveal insight into the dynamical process by which society changes. Indeed, the increasing frequency with which electronic text is used as a means of communicating, e.g., through email, text messaging, and social media, offers us the opportunity to quantify previously unobserved mechanisms of linguistic development.
While there are many aspects of language being investigated towards an increased understanding of social and linguistic evolution [1][2][3][4][5][6], one particular area of focus has been on changes in past tense forms for English verbs [1][2][3]. These investigations have collectively demonstrated that English verbs are going through a process of regularization, where the original irregular past tense of a verb is replaced with the regular past tense, formed using the suffix -ed.
For example, the irregular past tense of the verb 'burn' is 'burnt' and the regular past tense is 'burned'. Over time, the regular past tense has become more popular in general, and for some verbs has overtaken the irregular form. For example, in Fig. 1, we use the Google Ngram Online Viewer to compare the relative frequency of 'burnt' with that of 'burned' over the past 200 years. (As shown in an earlier paper involving two of the present authors [7], and expanded on below, the Google Ngram dataset is highly problematic but can serve as a useful barometer of lexical change.) In the first half of the 19th century, the irregular past tense 'burnt' was more popu-lar. However, the regular past tense 'burned' gained in popularity and in the late 1800s became the more popular form, which has persisted through to today.
Looking at several examples like this, in a 2011 paper Michel et al. studied the regularization of verbs, along with other cultural and language trends, as an accompaniment to their introduction of the Google Books Ngram corpus (hereafter Ngrams) and the proto-field 'Culturomics' [2]. They found that most of the verb regularization over the last two centuries came from verbs using the suffix -t for the irregular form, and that British English texts were less likely than American English ones to move away from this irregular form.
In a 2007 study, Lieberman et al. explored the regularization of English verbs using the CELEX corpus, which gives word frequencies from several textual sources [1]. Focusing on a set of 177 verbs that were all irregular in Old English, they examined how the rate of verb regularization relates to frequency of usage, finding that more common verbs regularized at a slower rate. They calculated half-lives for irregular verbs binned by frequency, finding that irregular verbs regularize with a half-life proportional to the square root of frequency of usage.
In a more recent study, Newberry et al. proposed a method for determining the underlying mechanisms driving language change, including the regularization of verbs [3]. Using the Corpus of Historical American English and inspired by ideas from evolution, the authors described a method to determine if language change is due to selection or drift, and applied this method to three areas of language change. They used a null hypothesis of stochastic drift and checked if selection would be strong enough to reject this null hypothesis. Of the 36 verbs Newberry et al. studied, only six demonstrated statistical support for selection. They also claimed that rhyming patterns might be a driver of selection.
Unfortunately, the corpora used in these studies have considerable limitations and corruptions. Relative word frequencies for the irregular and regular past verb forms for 'burn' during the 19th and 20th centuries, using the Google Ngram Online Viewer with the English Fiction 2012 corpus. Google Ngram trends can be misleading but capture basic shifts in a language's lexicon [7,8]. The irregular form 'burnt' was once more popular, but the regular form 'burned' overtook it in the late 19th century and its popularity has steadily increased ever since while that of 'burnt' has decreased. The dynamics of verb tense changes are rich, reflecting many processes at play in the Google Books Ngram data. An interactive version of this graphic can be found at https://books.google.com/ngrams/graph?content=burned%2Cburnt& year start=1800&year end=2000&corpus=16&smoothing=3. early versions of the Ngrams data includes scientific literature, whose explosive growth through the 20th century is responsible for the decreasing trend in relative word usage frequency observed in many common search terms [7]. Moreover, the library-like nature of the corpus admits no accounting for popularity: Lord of the Rings and an unknown work contribute with equal weight to token counts.
Another general concern with large corpora of a global language like English is that language use varies tremendously with culture and geography. Ngrams allows only for the regional exploration of the English language with the British English corpus and the American English corpus. Twitter data enables us to focus on much smaller spatial regions (e.g., county or state).
Prior studies of verb regularization have also focused on data reflecting a formal editorial process, such as the one undergone by any published book. This editorial process will tend to normalize the language, reflecting the linguistic opinions of a small minority of canon gatekeepers, rather than portray the language used by everyday people. For example, maybe the irregular from of a particular verb is considered proper by scholars, but a vast majority of the English speaking population uses the regular form. While it is not a verb form, one illustrative example is 'whom'. Although 'whom' is the correct word to use in the objective case, it is common for everyday speakers to use 'who'.
In the present study we take tweets to be a closer representation of everyday language. For the vast majority of accounts, tweets are authored by individuals without undergoing a formal editing process. As such, the lan-guage therein should more accurately represent average speakers than what is found in books.
The demographic groups contributing to Twitter are by no means a carefully selected cross-section of society, but do offer natural language use by the roughly 20% of adult English speakers who use Twitter [9]. When exploring temporal changes in language use, the Ngrams and CELEX datasets evidently cover a much longer period than the decade for which social media is available. As a result, we are unable to infer anything about the temporal dimension of regularization looking at Twitter.
In this paper we use the Ngrams and Twitter datasets to establish estimates of the current state of English verb regularization. We structure our paper as follows: In Sec. II, we describe the datasets we use. In Sec. III, we present our results. We study verb regularization in English in general in Sec. III A. We compare verb regularization in American English (AE) and British English (BE) using both Ngrams and geotagged Twitter data in Sec. III B. In Sec. III C, we employ methods to study regional variation in verb usage, leveraging county level user location data in the United States. We also explore correlations between verb regularization and a number of socio-demographic and economic variables. Finally, in Sec. IV, we provide concluding remarks.

II. DESCRIPTION OF DATA SETS
To be consistent with prior work, we chose the verb list for our project to match that of Michel et al. [2]. When comparing BE with AE, we use the subset of verbs that form the irregular past tense with the suffix -t. When calculating frequencies or token counts for the 'past tense' we use both the preterite and past participle of the verb. See Table A1 in Appendix A for a complete tabulation of all verb forms.
The Ngrams data reflects relative frequency, providing, for a verb and a given year, the percentage of corpus tokens that are the given verb, where a token is an individual occurrence of a word. The Google Ngram Online Viewer also has a smoothing parameter, s, which averages the relative frequency for the given year with that of each of the s years before and after the given year, if they exist. For example, Fig. 1 uses a smoothing of 3 years and shows that, averaged across the years 1997-2000 (the value displayed for the year 2000), the word 'burned' appeared with relative frequency 0.004321% (roughly once every 23,000 tokens), while 'burnt' appeared with relative frequency 0.000954% (roughly once every 105,000 tokens).
We downloaded the Ngrams verb data for the most recent 6-year period available (2003-2008) [10]. Specifically, we chose the 2008 values of relative frequency with a smoothing of 5 years, resulting in an average case insensitive 1 word frequency for the years 2003-2008. For general English, as suggested by [7], we queried the English Fiction 2012 corpus, which uses "books predominantly in the English language that a library or publisher identified as fiction." For AE we used the American English 2012 corpus, which uses "books predominantly in the English language that were published in the United States." For BE we used the British English 2012 corpus, which uses "books predominantly in the English language that were published in Great Britain" [11].
The Twitter messages for our project consist of a random sample of roughly 10% of all tweets posted between 9 September 2008 and 22 October 2017. This 'decahose' dataset comprises a total of more than 106 billion messages, sent by about 750 million unique accounts. From this larger set, we performed a case-insensitive search for verb forms of interest, also extracting geographic location when available in the meta-data associated with each tweet. Tweets geotagged by mobile phone GPS with a U.S. location comprise about a 0.27% subset of the decahose dataset; United Kingdom locations comprise about a 0.05% subset. Many individuals provide location information, entered as free text, along with their biographical profile. We matched user specified locations of the form 'city, state' to a U.S. county when possible, comprising a 2.26% subset of the decahose dataset. Details on this matching process can be found in Appendix B.
For general English, we counted the number of tokens in the decahose dataset for each verb. For AE, we used the tweets whose geotagged coordinates are located in the 1 When Ngrams computes a case insensitive word frequency it uses "the yearwise sum of the most common case-insensitive variants of the input query" [11].
United States, and for BE we used the tweets whose geotagged coordinates are located in the United Kingdom. For the analysis of verbs by county, we used the tweets with the user entered location information. The demographic data for U.S. counties comes from the 2015 American Community Survey 5-year estimates, tables DP02-Selected Social Characteristics, DP03-Selected Economic Characteristics, DP04-Selected Housing Characteristics, and DP05-Demographic and Housing Estimates, which can be found by searching online at https://factfinder.census.gov/. These tables comprise a total of 513 usable socio-demographic and economic variables.
We compute the regularization fraction for a verb as the proportion of instances in which the regular form was used for the past tense of the verb. More specifically, for Ngrams we divide the relative frequency for the regular past tense by the sum of the relative frequencies for the regular and irregular past tenses. Similarly, for Twitter we divide the token count for the regular past tense by the sum of the token counts for both the regular and irregular past tenses. If the resulting regularization fraction is greater than 0.5, the regular past tense is more popular and we call the verb regular. Otherwise we call the verb irregular.
When calculating an average regularization across all verbs, we first compute the regularization fraction for each verb individually. Then we compute the average of the regularization fractions, with each verb contributing the same weight in the average, irrespective of frequency. We perform this 'average of averages' to avoid swamping the contribution of less frequent verbs.

A. Verb regularization using Ngrams and Twitter
Using the datasets in row (I) of Table I, we begin by comparing Ngrams and Twitter with respect to regularization of English verbs in Fig. 2, where we find that 21 verbs are more regular in Ngrams, and 85 are more regular on Twitter. A Wilcoxon signed rank test of the data 10 −6 10 −5 10 −4 10 −3 10 −2 10 −1  Table I. Verbs are centered at their regularization fraction in Ngrams (horizontal) and Twitter (vertical). Both axes are on a logit scale, which spreads out both extremes of the interval (0, 1). Verbs to the right of the vertical dashed line are regular in Ngrams; verbs above the horizontal dashed line are regular on Twitter. The diagonal dashed line separates verbs that are more regular on Twitter (those above and to the left of the line) from those that are more regular in Ngrams (those below and to the right of the line). For example, compared with 'knew', the word 'knowed' appears roughly 3 times in 1000 in Ngrams, and 2 times in 10,000 on Twitter, making 'know' irregular in both cases, but more than an order of magnitude more regular in Ngrams than on Twitter. has a p-value of 7.9 × 10 −6 , demonstrating strong evidence that verbs on Twitter are more regular than verbs in Ngrams.

Ngrams Regularization Fraction
What mechanisms could be responsible for the observed increase in regularity on Twitter? One possibili-ty is that authors of fiction published in the 2000s, along with their editors, being professional users of English, have a larger vocabulary than the typical user of Twitter. If so, their commitment to proper English would contribute to the appearance of relatively more irregu- Another driver may be that non-native English speakers writing English tweets may be more likely to use the default regular form. We will find quantitative support for this mechanism below. As a preview, we note that Fig. 2 shows that 'burn' is predominantly regular on Twitter globally, but we see later (Fig. 3B) that 'burn' is irregular on Twitter for both American English and British English. Thus, it is likely that non-native speakers are contributing to this difference.

B. American and British English
We next study how verb regularization varies with geographic region. In this subsection we use the datasets in row (II) of Table I for AE and row (III) for BE and the subset of verbs that form the irregular past tense with the suffix -t.
In Fig. 3A, we compare American and British English in Ngrams. The average regularization fraction is 0.49 in AE and 0.42 in BE. For 17 out of 22 verbs, AE shows more regularization, with a Wilcoxon signed rank test pvalue of 9.8 × 10 −4 , giving statistical support that AE verbs are more regular on average in Ngrams than BE verbs.
As we show in the inset scatter plot of Fig. 3A, regularization in AE and BE are also strongly positively correlated with a Spearman correlation coefficient of 0.97 (p = 2.3 × 10 −14 ). Verbs that are more regular in AE are also more regular in BE, just not to the same extent.
In Fig. 3B, we compare regularization in AE and BE on Twitter. For Twitter, the average regularization fraction is 0.54 for AE, higher than Ngrams, and 0.33 for BE, much lower than Ngrams. As with Ngrams, 17 verbs out of 22 show more regularization in AE than in BE. The Wilcoxon signed rank test gives a weaker but still significant p-value of 1.9 × 10 −3 .
The inset in Fig. 3B also shows a positive correlation, although not as strong as Ngrams, with a Spearman correlation coefficient of 0.87 (p = 1.1 × 10 −7 ). Generally on Twitter, regular AE verbs are also regular in BE, but the difference in regularization fraction is much greater than for Ngrams.
In Fig. 4A, we demonstrate the difference in regularization between AE and BE for both Ngrams and Twitter. The values in this figure for Ngrams can be thought of as, for each verb in Fig. 3A, subtracting the value of the bottom bar from the top bar, and likewise for Twitter and Fig. 3B. Positive numbers imply greater regularization in AE, the more common scenario. When the difference is near zero for one corpus, it is usually close to zero for the other corpus as well. However, when Ngrams shows that AE is notably more regular than BE, Twitter tends to show a much larger difference. The average difference in regularization fraction between AE and BE for Twitter is 0.21, whereas it is only 0.08 for Ngrams. Again, we find that these averages are significantly different with a Wilcoxon signed rank p-value of 1.9 × 10 −2 .
The inset scatter plot tells a similar story, with a cluster of points near the origin. As the difference in regularization fraction between regions increases in Ngrams, it also tends to increase in Twitter, with Spearman correlation coefficient 0.65 and p-value 1.0 × 10 −3 . The steep rise shows that the difference increases faster on Twitter than in Ngrams. Fig. 4B returns to comparing Ngrams and Twitter, but now between AE and BE. For each verb, the bar chart shows the difference between the regularization fraction for Twitter and Ngrams in both AE and BE, with positive values showing that regularization for Twitter is greater. In this case, the values can be thought of as subtracting the values for the bars in Fig. 3A from the corresponding bars in Fig. 3B. As we find for English in general, regularization is greater on Twitter than in Ngrams for AE, with an average difference of 0.04. However, for BE, regularization is greater in Ngrams than on Twitter, with an average difference in regularization fraction of −0.09.
We summarize our findings in Table II. We found again that verbs on Twitter are more regular than in Ngrams for American English, likely for many of the same rea-  5. (A) The average verb regularization fraction by county for the lower 48 states, along with (B) residuals and (C) Gi * z-score. A higher Gi * z-score means a county has a greater regularization fraction than expected. Counties colored black did not have enough data. We used the dataset in row (IV) of Table I. more resistant to the change. Indeed, from the figures in the study by Gonçalves et al., one can see that the 'Americanization' of British English is more pronounced in Ngrams than on Twitter [12], agreeing with what we have found here.

C. Regularization by US county
In Sec. III B, we demonstrated regional differences in verb regularization by comparing BE and AE. Here, we consider differences on a smaller spatial scale by quantifying regularization by county in the United States using the dataset in row (IV) of Table I. We use methods inspired by Grieve et al. to study regional variation in language [14].
We only include counties that had at least 40 total tokens for the verbs under consideration. We plot the average regularization fraction for each county in the continental U.S. in Fig. 5A, where counties with not enough data are colored black. To control for the skewed distribution of samples associated with county population (see below for more details), we use residuals for this portion of the analysis. After regressing with the log 10 of data volume (total number of tokens) for each county, we compute the average regularization fraction residual, which is plotted in Fig. 5B.
That is, if we let d i be the total number of tokens for verbs in tweets from county i; α and β be the slope and intercept parameters computed from regression; and R i be the average regularization fraction for county i, then we compute the average regularization fraction residual for county i, r reg i , as Using the average regularization residual at the county level as input, we measure local spatial autocorrelation using the Getis-Ord Gi * z-score [15], where r reg = 1 n i r reg i , n is the number of counties, and w ij is a weight matrix. To obtain the weight matrix used in this calculation, we first create a distance matrix, s ij , where the distance between each pair of counties is the larger of the great circle distance, s GC ij , in miles between the centers of the bounding box for each county and 10 miles. That is, FIG. 6. The Gi * z-score for verb regularization by county for the verb 'dream' for the lower 48 states. Counties colored black did not have enough data. People tweet 'dreamed' rather than 'dreamt' more often than expected in the southeastern U.S.
We make the minimum value for s ij 10 miles to prevent a county from having too large of a weight. We then compute the weight matrix as Fig . 5C shows the results for the lower 48 states, where black represents counties left out because there was not enough data. For each county, the Gi * z-score computes a local weighted sum of the residuals, r reg j , for the surrounding counties and compares that to the expected value of that weighted sum if all the counties had exactly the average residual, r reg , as their value, where the weighting is such that closer counties have a higher weight. Areas that are darker blue (positive z-score) belong to a cluster of counties that has higher regularization than average, and those that are darker red (negative z-score) belong to a cluster that has lower regularization than average. So, Fig. 5C shows that, in general, western counties show less regularization than average and eastern counties show more, except that the New England area is fairly neutral.
As usual, the z-score gives the number of standard deviations away from the mean. For this we would do a two tail test for significance because we are looking for both high value and low value clusters. For example, a z-score greater in magnitude than 1.96 is significant at the .05 level. If we do a Bonferroni correction based on 3161 counties (the number included for this part of the analysis), then a z-score greater in magnitude than 4.32 is  Table I). The point's horizontal position is given by the total sample token count in a synthetic county; the vertical position is given by its average regularization fraction.
significant for a two tail test at the .05/3161 ≈ 1.58×10 −5 level. We do this same process looking at individual verbs as well. However, when looking at individual verbs, we use the regularization fraction rather than residuals, because the data skew is not as problematic. This is because the main problem with data volume comes when averaging across verbs that have different frequencies of usage, as explained below. Also, here we include counties that have at least 10 tokens. Fig. 6 gives an example map showing the Gi * z-scores for the verb 'dream'. The maps showing local spatial autocorrelation for the complete list of verbs can be found in the Online Appendix A at https://www. uvm.edu/storylab/share/papers/gray2018a/.
For many of the counties in the US, there is a small sample of Twitter data. We restrict our analysis to counties with a total token count of at least 40 for the verbs we consider. Even for the counties meeting this criteria, the volume of data varies, leading to drastically different sample sizes across counties.
More common verbs tend to have popular irregular forms (e.g., 'found' and 'won'), and less common verbs tend to be regular (e.g., 'blessed' and 'climbed') [1]. As a result, samples taken from populous counties are more likely to contain less common verbs. Our 'average regularization' is an average of averages, resulting in an underlying trend toward higher rates for more populous counties due to the increased presence of rarer regular verbs. Fig. 7 demonstrates the relationship between data volume and regularization. To explore the connection further, we perform a synthetic experiment as follows.
To simulate sampling from counties with varying population sizes, we first combine all verb token counts (using the Twitter dataset from row (I) of Table I) into a single collection. We then randomly sample a synthetic county worth of tokens from this collection. For a set of 1000 logarithmically spaced county sizes, we randomly draw five synthetic collections of verbs (each is a blue circle in Fig. 7). For each sample, we compute the average regularization fraction, as we did for U.S. counties. The goal is to infer the existences of any spurious trend introduced by the sampling of sparsely observed counties.
The resulting simulated curve is comparable to the trend observed for actual U.S. counties. As the data volume increases, the simulated version converges on roughly 0.17, which is the average regularization fraction for all of Twitter.
We also explored correlations between verb regularization and various demographic variables. Fig. 7 showed a strong relationship between data volume and verb regularization. It has been shown elsewhere that tweet density positively correlates with population density [16], and population size is correlated with many demographic variables. As a result, we use partial correlations as an attempt to control for the likely confounding effect of data volume.
For each demographic variable, we compute the regression line between the log 10 of data volume, d i , and regularization, and compute the residuals as in Eq. 1. Then, if the demographic variable is an 'Estimate' variable, where the unit is number of people, we similarly compute the regression line between the log 10 of data volume and the log 10 of the demographic variable 2 and compute 2 We do not include any county that has a value of zero for the For each county, the horizontal coordinate is given by the residual left after regressing the demographic variable with the log 10 of data volume and the vertical coordinate is given by the residual left after regressing that county's average regularization fraction with the log 10 of data volume. Data volume, for a county, is the total token count of all verbs found in tweets from that county. the residuals, r dem i , as where D i is the value of the demographic variable for county i, and γ and δ are the slope and intercept parameters calculated during regression.
demographic variable here to prevent errors when taking the log 10 . Otherwise, the demographic variable is a 'Percent' variable, with units of percentage, and we compute the regression line between the log 10 of data volume and the demographic variable, and compute residuals as The correlation between residuals r reg i and r dem i gives the partial correlation between average regularization and the demographic variable.
Our findings suggest that data volume is a confounding variable in at least some of the cases because, after controlling for data volume, there is generally a large decrease in the correlation between verb regularization and the demographic variables. The largest in magnitude Pearson correlation between verb regularization and a demographic variable is 0.68, for the variable 'Estimate; SCHOOL ENROLLMENT -Population 3 years and over enrolled in school', whereas the largest in magnitude partial correlation is only −0.18, for the variable 'Percent; OCCUPATION -Civilian employed population 16 years and over -Management, business, science, and arts occupations'. Table III lists the 10 demographic variables with largest in magnitude partial correlation. Fig. 8 shows an example for one of the demographic variables, the 'Percent' variable with largest simple correlation. Fig. 8A is the scatter plot of the demographic variable with average regularization, which corresponds to simple correlation. Fig. 8B is the scatter plot of the residuals, r dem i and r reg i , after regressing with the log 10 of data volume, and corresponds with partial correlation. We can see that there is a strong simple correlation (−0.52), but after accounting for data volume that correlation largely vanishes (−0.15). Similar plots for all of the demographic variables can be found in the Online Appendix B at https://www.uvm.edu/storylab/ share/papers/gray2018a/.

IV. CONCLUDING REMARKS
Our findings suggest that, by and large, verb regularization patterns are similar when computed with Ngrams and Twitter. However, for some verbs, the extent of regularization can be quite different. If social media is an indicator of changing patterns in language use, Ngrams data ought to lag with a timescale not yet observable due to the recency of Twitter data. Very reasonably, Ngrams data may not yet be showing some of the regularization that is happening in everyday English.
We also found differences in verb regularization between American and British English, but found that this difference is much larger on Twitter than Ngrams. Overall, and in American English specifically, verbs are more regular on Twitter than in Ngrams, but the opposite is true for British English. In the U.S., we also find variation in average verb regularization across counties. Lastly, we showed that there are significant partial correlations between verb regularization and various demographic variables, but they tend to be weak.
Our findings do not account for the possible effects of spell checkers. Some people, when tweeting, may be using a spell checker to edit their tweet. If anything, this will likely skew the language on Twitter towards the 'correct' form used in edited textual sources. For example, in Fig. 2 we see that 'stand' is irregular for both Ngrams and Twitter, and likely most spell checkers would consider the regular 'standed' a mistake, but we see that 'stand' is still over 100 times more regular on Twitter than in Ngrams. So, the differences between edited language and everyday language may be even larger than what we find here suggests. Future work should look into the effects of spell checkers.
Our study explored the idea that edited written language may not fully represent the language spoken by average speakers. However, tweets do not, of course, fully represent the English speaking population. Even amongst users, our sampling is not uniform as it reflects the frequency with which different users tweet (see Fig. D1 in Appendix D). Furthermore, the language used on Twitter is not an unbiased sample of language even for people who use it frequently. The way someone spells a word and the way someone pronounces a word may be different, especially, for example, the verbs with an irregular form ending in -t, because -t and -ed are close phonetically. However, the fact that we found differences between the language of Ngrams and the language of Twitter suggests that the true language of everyday people is not fully represented by edited written language. We recommend that future studies should investigate speech data.    Table I in Sec. II are considered 'high frequency', those in the interval [10 4 , 10 6 ) are 'mid frequency', and those in the interval [10 2 , 10 4 ) are low frequency. The bins contain 37, 55, and 14 verbs, respectively. For each county (with at least 40 total tokens), the average regularization fraction of the verbs in each of the three bins is calculated (if it is not empty) and plotted against the total token count for all verbs for that county.

Appendix D
FIG. D1. The frequency counts of tweets by unique users in our Twitter decahose dataset (row (I) of Table I in Sec. II). Users are ranked by their total number of tweets along the horizontal axis and the vertical axis gives the total number of tweets we have associated with each user's account.