Skip to main content
Browse Subject Areas

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Mapping online hate: A scientometric analysis on research trends and hotspots in research on online hate

  • Ahmed Waqas,

    Roles Data curation, Formal analysis, Methodology, Writing – original draft

    Affiliations University of Liverpool, Liverpool, United Kingdom, CMH Lahore Medical College & Institute of Dentistry, Lahore, Pakistan

  • Joni Salminen ,

    Roles Conceptualization, Supervision, Writing – original draft, Writing – review & editing

    Affiliations Qatar Computing Research Institute, Hamad Bin Khalifa University, Doha, Qatar, Turku School of Economics at the University of Turku, Turku, Finland

  • Soon-gyo Jung,

    Roles Conceptualization

    Affiliation Qatar Computing Research Institute, Hamad Bin Khalifa University, Doha, Qatar

  • Hind Almerekhi,

    Roles Conceptualization, Writing – review & editing

    Affiliation Hamad Bin Khalifa University, Doha, Qatar

  • Bernard J. Jansen

    Roles Conceptualization, Project administration, Supervision, Writing – review & editing

    Affiliation Qatar Computing Research Institute, Hamad Bin Khalifa University, Doha, Qatar


Internet and social media participation open doors to a plethora of positive opportunities for the general public. However, in addition to these positive aspects, digital technology also provides an effective medium for spreading hateful content in the form of cyberbullying, bigotry, hateful ideologies, and harassment of individuals and groups. This research aims to investigate the growing body of online hate research (OHR) by mapping general research indices, prevalent themes of research, research hotspots, and influential stakeholders such as organizations and contributing regions. For this, we use scientometric techniques and collect research papers from the Web of Science core database published through March 2019. We apply a predefined search strategy to retrieve peer-reviewed OHR and analyze the data using CiteSpace software by identifying influential papers, themes of research, and collaborating institutions. Our results show that higher-income countries contribute most to OHR, with Western countries accounting for most of the publications, funded by North American and European funding agencies. We also observed increased research activity post-2005, starting from more than 50 publications to more than 550 in 2018. This applies to a number of publications as well as citations. The hotbeds of OHR focus on cyberbullying, social media platforms, co-morbid mental disorders, and profiling of aggressors and victims. Moreover, we identified four main clusters of OHR: (1) Cyberbullying, (2) Sexual solicitation and intimate partner violence, (3) Deep learning and automation, and (4) Extremist and online hate groups, which highlight the cross-disciplinary and multifaceted nature of OHR as a field of research. The research has implications for researchers and policymakers engaged in OHR and its associated problems for individuals and society.


The advent of the modern Internet opens doors to a plethora of positive opportunities for the general public. These opportunities span across equity in education and general access to knowledge, modes of entertainment, consumerism, and e-participation. However, in addition to these positive aspects, digital technology also provides an effective medium for spreading hateful content in the form of bigotry and hateful ideologies, as well as cyberbullying and harassment of individuals and groups on social media platforms [1,2]. Online hate, albeit conducted in the virtual world, may have dire real-life consequences at both individual and population levels. For example, the cyberbullying among youth and student populations and subsequent links with poor mental health, depression, trauma, substance misuse, and a higher risk of suicide are well-documented [36]. Recent estimates have placed exposure to online hate ranging from 31% to 67% across different study samples [7]. Among New Zealanders, for example, 11% of adults have been personally targeted by online hate [1], whereas, in the US, 41% of adults have experienced online hate speech and harassment [8]. Online hate has been shown to predominantly target and influence minorities, young age groups, people with disabilities, and the LGBTQ (Lesbian, Gay, Bisexual, Transgender, Queer) community [1].

Online hate spreading has also emerged as a tool for politically motivated bigotry, xenophobia, homophobia, and excessive nationalism [912]. An example can be seen in the 2016 US elections; the narrative of “Make America Great Again” has empirically been shown to have amplified the online presence of white supremacists [9]. Social media platforms have granted a new spirit to radical nationalist groups including Klansmen and Neo-Nazis by ensuring anonymity or pseudonymity (i.e., disguised identity), ease of discussions, and spread of radical ideologies [1]. Moreover, social media and online forums have provided hate-driven terrorist groups a medium for launching propaganda to radicalize youth globally [13]. These groups use images and Internet videos to communicate their hateful intent, to trigger panic, and to cause psychological harm to the general public [14]. As a prime example of cyberterrorism, the Islamic State of Iraq & Syria (ISIS) effectively used social media to recruit youngsters from Europe to participate in the Syrian conflict [12]. Their social media campaigns led to at least 750 British youngsters joining Jihadi groups in Syria [13]. Overall, these real-world phenomena highlight the very real negative impact of spreading online hate and suggest that online hate can be considered as a major public concern.

However, online hate is a complex phenomenon—with its definition depending on theoretical paradigms, disciplines, and forms of victimization [1,15]. Due to this complexity, online hate research (OHR) is a fragmented field with a growing number of research papers across disciplines, as the adverse effects of online hate are more widely recognized in society and as new disciplines (e.g., computer science, psychology) are introducing their own approaches to study and solve the associated problems. Due to this increasing body of research, there is a need for literature analyses that map the current state of OHR. While several evidence-synthesis approaches have attempted to summarize and critically review the literature on online hate, these tend to be based on heterogeneous methodologies and restricted to a particular discipline or field of study [9,10,23,13,1622]. For example, an elaborate effort by the British Institute of Human Rights sought to systematically map studies about initiatives against cyberbullying and inform legislative efforts by the European Union [21]. A qualitative approach by Awan sought to provide evidence regarding the use of social media platforms by ISIS by examining 100 Facebook pages and 50 Twitter users [13]. Country-specific efforts included Gagliardone et al.’s efforts to map politically driven online hate in Ethiopia by reviewing relevant Facebook profiles, pages, and groups with more than 100 followers [23], which provided a framework for analyzing online hate speech and explored the continuum between freedom of expression [23]. Cyber-bullying has also attracted attention from public health and mental health professionals. Most influential and cited work in this domain is attributed to Tokunaga, who critically reviewed and synthesized evidence on cyberbullying victimization [20].

However, none of the previous work, to the knowledge of authors, has focused on the mapping of general research indices, prevalent themes of research, research hotspots, and influential stakeholders such as organizations and contributing regions regarding OHR. This undertaking is essential as such analyses help to evaluate the field-specific impact of scholarly research, as well as the impact of scientists, collaborative networks, and institutes. Therefore, we set out to map OHR using scientometric analysis, defined as the “quantitative study of science, communication in science, and science policy” [24]. Most importantly, scientometrics helps identify influential research studies resulting in the progress and evolution of a specific field of science [24]. By using reproducible statistical techniques, stakeholders can quantize the research output, citation rates, influential funding agencies, journals, scientists, institutes, and regions involved in the progress of the scientific discipline [24]. By mapping these trends, researchers, policymakers, and funding agencies can determine areas where an increase or restriction in research work and funding is required [2527]. Therefore, this investigation aims to address this paucity of data using advanced scientometric techniques.


Search strategy

We defined the focal topic of study as online hate. We identified several definitions from the prior literature that helped us understand the nature of the phenomenon and to collect a list of concepts that reflect the multifaceted nature of OHR. Definitions of online hate vary, but a unifying factor is the use of technology for expressions that are harmful to individuals, groups, or society as a whole. An example of a definition that encompasses this duality is that of Kaakinen et al., according to whom online hate has two defining characteristics: it is technology-mediated and intends to offend, discriminate and abuse a person or a group based on group defining characteristics such as gender, race, nationality, ethnicity, disability, or sexual orientation [7].

In the course of exploring the definitions, we compiled a list of keywords for the electronic search carried out to identify the body of research about OHR (see Table 1).

Table 1. Key concepts in online hate research, operationalized as search terms.

In addition to operationalizing the concepts in Table 1 as search terms, we defined a list of popular social media platforms that were also used as search terms, as several studies focus on hate taking place in a specific social media platform. Using the Web of Science core database, an electronic search was conducted to retrieve peer-reviewed research studies (published through March 2019) pertaining to online hate. Overall, this search strategy encompassed important concepts pertaining to online hate and popular platforms: “TS = (Hate OR toxicity OR cyberbullying OR bullying OR harass* OR firestorm* OR abuse OR abusive OR ‘abusive language’ OR maltreat* OR oppress* OR persecut* OR taunt* OR bully* OR bullies OR victim* OR ‘hate speech’) AND TI = (Online OR ‘social media’ OR web OR virtual OR cyber OR Orkut OR Twitter OR facebook OR Reddit OR Instagram OR snapchat OR youtube OR whatsapp OR wechat OR QQ OR Tumblr OR linkedin OR pinterest)”. As mentioned, this search strategy was formulated based on an initial reading of the literature and identifying commonly emerging terms in the studies about online hate. No restrictions were applied for year of publication or language.

The search process resulted in a total of 3,371 research articles for a scientometric analysis. The data curated from the Web of Science (core database) included the citation characteristics, citation counts, and cited references. The Web of Science core database is one of the most frequently used databases for scientometric analyses. It was chosen primarily because it indexes detailed citations and full records of cited references that help in elucidating co-citation relationships between related documents [28].

Operational definitions and inclusion criteria

The present mapping study is a broad overview of OHR. In line with our objectives, a broader interpretation of online hate was preferred, covering all forms of expressions that spread, incite, promote, or justify hate against groups or individuals [21]. This interpretation was adapted from the framework for online hate proposed by the British Institute of Human Rights [21]. All forms of expressions on a macro-level including racial hatred, xenophobia, anti-Semitism, aggressive nationalism, and hatred against minorities and migrants were included. On an individual level, various forms of expression, for instance, partner abuse as well as cyber-bullying against school children owing to their racial, ethnic, sexual background, and disabilities were included [21]. We acknowledge that there are alternative definitions for online hate and online toxicity, the latter of which can be defined as rude, disrespectful, or unreasonable commenting that is likely to make one leave a discussion [29,30]. Most of these definitions perceive online hate as a conceptually broad phenomenon that touches many stakeholder groups. For that reason, we consider broad inclusion criteria to be relevant for this research.

Co-citation analysis and knowledge mapping

In the first phase, data curated from the Web of Science core database (WOS) was utilized for knowledge mapping based on the theory of document co-citation. According to this theory, when two documents are co-cited by one document, they are connected in a co-citation relationship [31].

Co-citation analyses were performed using CiteSpace software (n = v4.0, Drexel University, Pennsylvania, US). The bibliographic records retrieved from WOS were fed into the CiteSpace software, and “sliced” into three-year slices, where each slice was represented by 50 documents with the highest cited frequency. Titles, abstracts, and keywords were used as terms sources while cited references were used as nodes.

After that, network analysis was run using pathfinder network scaling while allowing for the pruning of sliced networks [2527]. All bibliographic data were then visualized as merged and static networks/clusters. Articles were represented as nodes, while the relationship between nodes was visualized as lines or edges. Two important matrices were used to demonstrate the overall structural properties of the network: modularity and silhouette value. Note that a high value of modularity (close to 1) corresponds to a good network structure that is reasonably divided into loosely coupled clusters, and a high silhouette score represents an appropriately homogenized cluster. This technique allowed for the visualization of important publications in a collaborative network based on their centrality values, also identified as a tree ring representing their history of citations and year-wise patterns [2527]. New theories and landmark studies with high between-ness centrality were identified as purple rings while citation bursts were visualized as red tree rings [2527].

Citation bursts were defined as articles attracting significant research activity in a given period. Clusters and themes of research in this field were identified by running a cluster analysis that identified the publication record cited in a specific set of publications, and the clusters were named using naming algorithms including TF*IDF; Mutual Information (MI) and Log Likelihood Ratio (LLR) [2527]. Each cluster was also depicted by a year representing the mean year of publications of all included research studies. Out of these methods, LLR has been shown to be the most accurate [2527]. The first method, TF*IDF, utilizes terms that are weighted by term frequencies (TF) multiplying inverted document frequencies (IDF) [2527]. Log-likelihood ratio tests choose the most appropriate clustering label by assessing the strength of the bond between a term and the cluster [2527]. Generally, the higher the LLR, the better the evidence. Lastly, the mutual information method is used for feature selection in machine learning; however, it works better with larger datasets [2527].


Research activity

The search process yielded a total of 3,371 publications that were included in the scientometric analysis. These publications boasted an h-index of 82, 11.23 citations per item, cited for a total of 37,848 times overall (n = 33,721 excluding self-citations). Increased publication and citation activities were observed post- 2005 starting from >50 publication to > 550 in the year 2018 (Figs 1 and 2).

Top organizations, funders, and regions

The United States of America (US) was the most frequent publisher in this domain with 1,205 publications, followed by England, Australia, China, Canada, India, Germany, Spain, the Netherlands, and Italy. Among universities, the University of London, UK was the most frequent contributor, followed by university systems in the US: the University of California System, the Pennsylvania Commonwealth System of Higher Education, State University of Florida, the University of North Carolina, the University of Texas System, the University of Georgia, the University of Washington, Columbia University, and the University of Washington in Seattle. Top funders included United States Department of Health and Human Services (HHS)/National Institutes of Health (NIH), National Natural Science Foundation of China, National Science Foundation, Economic and Social Research Council, National Institute of Drug Abuse, European Union, and Catalan Institution for Research and Advanced Studies (ICREA). Collaborative networks of countries and institutes are presented as Figs 3 and 4, while frequencies of publications by top countries are presented in Table 2.

Table 2. Top countries, institutes, and sources according to the number of publications.

Top sources

Top sources included Computer in Human Behavior, Lecture Notes in Computer Science, Cyberpsychology, Behavior & Social Networking, Journal of Medical Internet Research, Journal of Adolescent Health, Journal of Youth and Adolescence, Procedia Social and Behavioral Sciences, PLOS One, New Media Society, and Child Abuse & Neglect. While most frequent conference proceedings were published by IEEE ACM International Conference on Advances in Social Network Analysis and Mining, Annual International Conference on Education Research and Innovation, International Conference on World Wide Web, ACM Conference on Computer Supported Cooperative Work and Social Computing, Saudi Computer Society National Computer Conference, IEEE International Conference on Trust Security and Privacy in Computing and Communication Trustcom, ACM SIGSAC Conference on Computer and Communications Security and International Conference on Intelligence and Security Informatics Cybersecurity and Big Data. Frequencies of publications by top sources are presented in Table 2.

Fields of publication

Top ten fields of publication included computer science information systems (n = 325), computer science theory methods (n = 282), criminology (n = 263), communication (n = 221), multidisciplinary psychology (n = 193), electrical/electronic engineering (n = 187), computer science interdisciplinary publications (n = 183), psychiatry (n = 168), educational research (n = 180) and clinical psychology (n = 154).

Top papers based on centrality in respective clusters

Top papers were judged based on their values of centrality, where a value of 0.1 indicates a central publication. In a collaborative and co-cited network of publications, a high centrality value reflects highly significant research studies. However, in this analysis, none of the studies reached a centrality value of 0.1, indicating no central publication in the respective cluster. However, top centrality value (> 0.01) was achieved by 14 studies (Table 3 and Fig 5). The majority of these papers focused on cyberbullying among adolescents. Tokunaga RS (2010) and Kowalski RM (2007) were found to be most central to entities with centrality values of 0.04.

Six publications, including Raskauskas and Stoltz [32]; Kowalski and Limber’s as well as Smith et al.’s work from 2007 to 2008 [5] were one of the earliest studies that noted the prevalence and nature of electronic bullying, victimization, and perpetration among American pupils [5,32,33]. Dehue et al. [34] focused on youngsters’ experience of cyberbullying as well as their parents’ perception about it. They found that parents do set rules for the use of the Internet for their children but are not conscious of their perpetrating behavior and also underestimate victimization experiences [34]. Slonje and Smith reported four types of cyberbullying—by text message, email, phone call, and video clip—and emphasized that bullying by video clips is perceived as most negative in the society, and most of the pupils tell their school friends about their experiences and not their parents [35]. Erdur-Baker explained the risky use of the Internet and its association with cyberbullying in Turkey and was one of the rarer studies conducted outside the US [36].

Tokunaga provided synthesized critical review evidence of cyberbullying and provided an integrative definition of cyberbullying, differentiated it from traditional bullying, and linked it with serious psychosocial and affective problems [20]. His work also outlined the areas of concern in research on cyberbullying and provided a framework for future research [20]. In a similar vein, Junon and Gross [37] reported patterns of cyberbullying and their association with social anxiety among school going children [37]. Hinduja and Patchin provided the earliest link of cyber-aggression and increased risk of suicide [4]. Ybarra et al. [38] associated cyberbullying to rule-breaking behavior and aggression in real life in a dose-dependent manner [38].

Two studies focused on the development of the most widely used psychometric questionnaires in cyberbullying. Calvete et al.’s [39] work was the earliest work that led to the development and validation of the Cyberbullying questionnaire for profiling aggressors and cyberbullies [39]. They also reported that the use of proactive aggression, justification of violence, exposure to violence, and less perceived social support of friends was prevalent among cyberbullies [39]. A cyber-dating abuse questionnaire assessed two latent constructs: direct aggression among romantic partners and monitoring control, such as the use of personal passwords [40]. Another of the two studies reported teen dating abuse using an online medium and online sexual solicitations in chat rooms and its risk factors including using chat rooms, using the Internet with a cell phone, talking with people met online, sending personal information to people met online, talking about sex online, and experiencing offline physical or sexual abuse [41,42].

Domains of research: Cluster analysis

A total of 101 clusters of research emerged in the cluster analysis (Fig 6). These clusters were given names according to four methods: Latent Semantic Indexing (LSI), Term Frequency * Inverted Document Frequency (TF*IDF), loglikelihood ratio (LLR), and Mutual Information (MI). We report in parentheses which method was used to derive the name for a given cluster; generally, it is not important to report all of them, as the outputs of each method were not always sensical. Detailed information regarding the top 10 clusters and their timelines have been presented as Figs 6 and 7. This analysis was based on 499 nodes and 906 lines or edges and yielded modularity of 0.86.

Fig 7. Timeline view depicting clusters of research arranged on a horizontal timeline from 2000 to 2018.

Clusters on cyberbullying

Five clusters focused on the theme of cyber-bullying. The first meaningful cluster (n = 48, silhouette value = 0.91) emerged as a social networking site as per TF*IDF, cyberbullying, internet harassment and sexual harassment and cyberbullying experience (MI) in 2006 (mean year of publication of included studies). In other words, there were 48 research articles with a similar theme that could be presented with the cluster title of “social networking site” by the TF * IDF method. These 48 articles were placed in this cluster because all of them were cited by a similar group of publications, thus, representing a co-citation relationship. The most cited of this group was Mishna [43] who investigated cyberbullying behaviors among Canadian adolescents. They reported that bullying perpetrators perceived themselves as funny, popular, and powerful, albeit feeling guilty as well [43]. The second meaningful cluster included 48 studies with a silhouette value of 0.88 in 2011. It was named as general strain theory (TF*IDF), cyberaggression (LLR), and Australian youth (MI). The most active citer was Kowlaski et al. [44], who reported cyberbullying behavior among college students across multiple domains of life [44].

Cyberbullying and utilization of routine activity theory were discussed in the seventh cluster with 15 members, a silhouette value of 0.99 and the mean year 2004. It was termed as social networking site by TF*IDF method, internet user, utilizing routine activity theory, potential factor by LLR method, and case study by MI method. The most active citer of this cluster was Marcum et al. [45], who provided causal reasoning for cyber-victimization utilizing the framework of routine activity theory [45]. This theory posits that victimization requires three factors: the presence of a likely offender, a suitable target, and the absence of a capable guardian [45].

The 12th cluster focused on the association of spending time in online communities (TF*IDF) with the mental health of adolescents and caregiver-child relationships (LLR and MI). This cluster included seven papers with a silhouette value of 1.00 in 2000. The most active citer of this group was Ybarra et al. in 2004, who focused on Internet harassment and its association with quality of child-caregiver relationship [46]. The 16th cluster reported papers on an educational and artistic intervention to prevent cyberbullying. It was termed as virtual drama, the emergent narrative approach, and anti-bullying education (TF*IDF, MI, LLR), and emerged in 2005 [47]. The most active citer, Aylett et al. [47] presented evidence for virtual educational software to prevent cyber-bullying.

Clusters of sexual solicitation and intimate partner violence

A total of three important clusters focused on the theme of sexual solicitation, dating abuse, and intimate partner violence. The third cluster focused on social support (TF*IDF) sexual solicitation via electronic mail; seeking human service; social support (LLR and MI) and included 44 papers. The most active citer was Finn (2000), who described the dangers involved when women seek human services on the internet [48]. This cluster emerged in the year 1998, highlighting early years of research.

The sexual solicitation was the focus of another cluster with 17 papers and a silhouette value of 0.94, emerging in the year 2012. It was termed as extent, situational factor (LSI); hate speech, network site, and online sexual solicitation (LLR, MI). It focused on the abuse of minors as well as online exposure among the youth as evident by its most active citers [49].

The tenth cluster focused on intimate partner violence by utilizing routines activity theory, comprising ten papers in the year 2011 and a mean silhouette value of 0.99. It was labeled as information security; the extent of cyberbullying behavior (TF*IDF), cyber partner abuse, systematic review, routine activities theory, and empirical study (LLR, MI). The most active citer for this cluster was Arntfield (2015), who proposed a new framework for understanding cyber victimology using the Routines Activity Theory Framework [50]. The author stressed the role of victims as both a facilitator and factor for predation [50]. The terms “systematic review” and “empirical study” refer to the study designs utilized by studies in these clusters.

Clusters on deep learning & automation

Deep learning and automation were studied in two important clusters. The fourth cluster focused on cyber defense (TF*IDF) and adaptive use and network-centric mechanism (LLR) and emerged in 2000. The most active citer was Atighetchi in 2000, whose work focused on defending against network-based attacks, and development of technologies augmenting an application’s resilience against hackers [51]. The 20th cluster revealed deep learning models and text classification as a viable source for identification of hate speech on Facebook groups in 2016 with a silhouette value of 1.0. The papers by Agrawal et al. [52] and Pitsilis et al. [53] were the most common citers of these clusters. Pitsilis et al. [53] proposed recurrent neural network models to discern hateful content on social media utilizing user-related information such as their tendency toward racism and sexism [53], while Agrawal et al. [52] showed that previous algorithms aiding in detection of cyberbullying have bottlenecks: specific platform, a specific topic of bullying, and thirdly, reliance on handcrafted features of the data. They proposed that deep learning models are viable in all of these situations [52].

Clusters on extremist & online hate groups

This cluster (#5) emerged in the year 2002 and included 18 research items. It was named as extremist groups and mining communities (TF*IDF); online hate group, mining communities, attack tolerance (LLR, MI). The most active citing paper of this cluster was published by Chau et al. [10], who emphasized the importance of analyzing the trends of online hate communities and terrorist groups who share their ideologies to recruit new members. They proposed network analysis and mining techniques as important weapons in this arena [12]. The 14th cluster revealed the use of discourse theory and critical theory as a framework for studying online Islamophobia (TF*IDF, MI). This cluster also had studies focusing on feminism and compensatory manhood (LLR). The most active citer reported harassment and misogyny in online sexual market places and dating websites such as Tinder [54]. The cluster also includes papers on automatic identification and classification of misogynistic languages on social media using NLP and machine learning methods [55]. Moreover, a paper on Islamophobia revealed 11 fake Facebook pages run by Danish citizens posing as Muslims threatening to kill and rape Danish citizens, termed as platformed antagonism [56].

Keyword analysis

Furthermore, we used keywords from titles, abstracts, and keywords sections of the research papers to construct keyword co-occurrence networks (see Fig 6). Co-occurrence and frequency of occurrence of keywords provide a snapshot and a reasonable description of trends of research in a specific area [26]. Also, analysis of burst items provides short periods of significant activity in a particular domain or an emerging topic and research frontier [26]. Fig 6 presents the most frequently cited keywords, with larger rings presenting significant keywords. According to it, Internet, adolescents, victimization, social media, Facebook, Twitter, experience, gender, children, victim, victimization, youth, school, toxicity, abuse, and risk most frequently occurring items cited at least 90 times in the literature. Table 4 lists the top 25 cited keywords, and Fig 8 presents co-citation relationship between keywords.

When burst items analysis was conducted, a total of 53 burst items were identified (see Fig 9). The time interval of the scientometric analysis (2000–2018) has been depicted as a blue line and the period that represents the burst activity, as a red line [26,57]. It presented four main themes of research hotspots in this field, including:

  1. Cyberbullying: this hotspot focuses on the pattern of cyberbullying such as cyber-victimization; cyber-bullying, harassment; privacy intrusion; sexual solicitation and involvement.
  2. Social media platforms: focused on online communities and specific social media platforms for detection and prevention of hate speech using deep learning and automation.
  3. Co-morbid disorders: this hotspot is characterized by keywords such as addiction; substance use; post-traumatic stress disorder; and Internet addiction, citing the importance of co-morbid mental health symptoms among aggressors and victims of cyberhate.
  4. Profiling of aggressors and victims: It was characterized by keywords such as identity; school student; personality; gender differences; and identification and risk assessment. These citation bursts exhibit increased research focused on psychological characteristics of both the aggressor and victims. This group also stratifies the population based on their demographic characteristics and increased risk of bullying behaviors.


Summary of results

The present study highlights the trends of research in the field of OHR. It revealed several clusters of OHR, innovative techniques to detect hate speech, sexual solicitation, exposure to pornography, Islamophobia, misogyny, and cyber-bullying along with its effects among the youth. The US was the lead contributor to this field of research, and our analysis also revealed a clear dominance of Western universities as well as funders from North America, Europe, and China. This global dominance and a higher share of Western institutions have been noted in several empirical investigations [5860]. Moreover, our analysis revealed a major contribution from psychology-related fields, spanning across the study of human behavior, psychological profiling of aggressors and victims, and co-morbid disorders such as depression and Internet addiction or pathological Internet use, as well as the association between offline and online bullying behaviors. These studies are highlighting the negative consequences of online hate, such as the increased risk of suicide among the victims of cyberbullying [3537,43]. Overall, there has been a significant increase in publication and citation trend in OHR after the year 2005, which coincides with the proliferation of social media platforms and the Internet becoming a central arena for public and private discourse.

Strengths, limitations, and future work

There are several strengths and limitations to this study. This is a first concerted effort to map the research activity on online hate. In contrast to previous studies designed as qualitative content analyses or literature reviews on a restricted topic, this study provides a broader analysis of publications of online hate. However, there are a few limitations to this study. Co-citation analyses is a quantitative technique to map research output in a field, and there are several other indicators such as the number of citations accrued or quality of a research article [61]. The role of citation frequency alone to map most influential studies has been long debated [61].

Moreover, while our analysis revealed a major contribution from psychology-related fields, this high representation of psychology-related contributions may be due to several reasons; for instance, the choice of WOS core as the database. Its coverage may be geared towards health and social science disciplines rather than engineering or computer sciences [62], thereby excluding some relevant research from these fields from the analysis. It may also be because there has been a mushroom growth and development in psychology-related publications, interdisciplinary and collaborative networks, as well as higher citation rates, took place in this domain. While we defend the choice of the WOS core database because it is one of the few databases yielding records for cited references [25,28] and embodying a curated collection of over 20,000 peer-reviewed publications pertaining to 250 disciplines in science, social sciences, and humanities [25,28], thereby being accessible for scientometric analyses, we acknowledge that there is a body of OHR literature that is not included in our analysis due to sampling limitations. Future research should aim at replicating or extending this study by accessing literature from other databases, such as ACM Digital Library.

Implications for research and practice

The main lessons learned from this scientometric analysis are as follows:

  1. Most of the publications originate from the discipline of psychology and psychiatry with recurring themes of the prevalence of cyber bullying, psychiatric morbidity, and psychological profiles of bullies and victims, particularly among the youth. In later years, there was some focus on dating violence and harassment of women. The main implication is that policy makers, and funders need to shift their focus on other fields, such as intervention and implementation sciences to design both technological and non-technological solutions to identify and curb online hate.
  2. Almost all the influential studies have been conducted in the context of high-income countries. Research is needed in low and middle-income countries to justify the generalizability of OHR findings as well as to produce culturally applicable interpretations.
  3. As far as we are aware of, this is the first concerted effort to map global research output regarding OHR, spanning across scientific disciplines such as psychology, computer sciences, and the social sciences. However, the dominance of psychology related publications may have skewed the overall results. For this reason, we also encourage discipline-specific scientometric studies because most of the studies published to date were i) discipline or population-specific, ii) simplistic literature reviews, and iii) lacked systematic search process and iv) reproducible data science techniques.

In conclusion, the increase in OHR is a reaction to the increased occurrences of hate speech, in all of its various forms, on the many social media and other online platforms. Online hate speech is, obviously, a complex societal problem that intersects many aspects of everyday life. The cross-disciplinary and multifaceted nature of OHR as a field of research is a witness to the complex issue of online hate. The findings from research so far hint at the need for both technology and non- technology approaches to address this increasingly pressing societal issue.

Supporting information

S1 Dataset. All data associated with this study have been provided as a supplementary file named data



  1. 1. Pacheco E, Melhuish N. Online Hate Speech: A Survey on Personal Experiences and Exposure Among Adult New Zealanders. SSRN Electronic Journal. 2018.
  2. 2. Salminen J, Almerekhi H, Milenković M, Jung S, Kawak H, Jansen B. Anatomy of Online Hate: Developing a Taxonomy and Machine Learning Models for Identifying and Classifying Hate in Online News Media. Proceedings of The International AAAI Conference on Web and Social Media (ICWSM 2018). San Francisco; 2018.
  3. 3. Study AM, Selkie EM, Kota R, Chan Y, Moreno M. Use in Female College Students : 2015;18: 79–86. pmid:25684608
  4. 4. Hinduja S, Patchin JW. Bullying, cyberbullying, and suicide. Arch Suicide Res. 2010;14: 206–221. pmid:20658375
  5. 5. Kowalski RM, Limber SP. Electronic Bullying Among Middle School Students. J Adolesc Heal. 2007;41: 22–30. pmid:18047942
  6. 6. Whittaker E, Kowalski RM. Cyberbullying Via Social Media. J Sch Violence. Routledge; 2015;14: 11–29.
  7. 7. Kaakinen M, Oksanen A, Räsänen P. Did the Risk of Exposure to Online Hate Increase After the November 2015 Paris Attacks ? A Group Relations Approach Computers in Human Behavior Did the risk of exposure to online hate increase after the November 2015 Paris attacks? A group relations appr. Comput Human Behav. Elsevier Ltd; 2017;78: 90–97.
  8. 8. Pew Research Center. Online harassment 2017 [Internet]. 2017.
  9. 9. Eddington SM. The Communicative Constitution of Hate Organizations Online: A Semantic Network Analysis of “Make America Great Again.” 2018;
  10. 10. Gagliardone I. Mapping and Analysing Hate Speech Online. Ssrn. 2015;
  11. 11. Olteanu A, Castillo C, Boy J, Varshney KR. The Effect of Extremist Violence on Hateful Speech Online. 2018;
  12. 12. Chau M, Xu J. Mining communities and their relationships in blogs: A study of online hate groups. Int J Hum Comput Stud. 2007;65: 57–70.
  13. 13. Awan I. Cyber-Extremism: Isis and the Power of Social Media. Society; 2017; 138–149.
  14. 14. Minei E, Matusitz J. Cyberspace as a new arena for terroristic propaganda: an updated examination. Poiesis Prax. 2012;9: 163–176. pmid:23205002
  15. 15. Salminen J, Veronesi F, Almerekhi H, Jung S, Jansen B. Online Hate Interpretation Varies by Country, But More by Individual: A Statistical Analysis Using Crowdsourced Ratings. Proceedings of The Fifth International Conference on Social Networks Analysis, Management and Security (SNAMS—2018). The Fifth International Conference on Social Networks Analysis, Management and Security (SNAMS—2018),. Valencia; 2018.
  16. 16. Gerstenfeld PB, Grant DR, Chiang C-P. Hate Online: A Content Analysis of Extremist Internet Sites. Anal Soc Issues Public Policy. 2003;3: 29–44.
  17. 17. Watts LK, Wagner J, Velasquez B, Behrens PI. Cyberbullying in higher education: A literature review. Comput Human Behav. Elsevier Ltd; 2017;69: 268–274.
  18. 18. Correa D, Sureka A. Solutions to Detect and Analyze Online Radicalization: A Survey. 2013;V: 1–30.
  19. 19. Zainudin NM, Zainal KH, Hasbullah NA, Wahab NA, Ramli S, Washington ET, et al. Cyberbullying: A Review of the Literature. Comput Educ. 2015;2: 1–9.
  20. 20. Tokunaga RS. Following you home from school: A critical review and synthesis of research on cyberbullying victimization. Comput Human Behav. Elsevier Ltd; 2010;26: 277–287.
  21. 21. British Institute of Human Rights. Mapping study on projects against hate speech online. 2012;2: 1–57.
  22. 22. Waltman MS, Mattheis AA. Understanding Hate Speech. Oxford Res Encycl Commun. 2017; 1–30.
  23. 23. Gagliardone I. MECHACHAL: Online debates and elections in Ethiopia. From hate speech to engagement in social media. 2016.
  24. 24. Hess D. Science Studies: An advanced introduction. New York: New York University Press; 1997.
  25. 25. Naveed S, Waqas A, Majeed S, Zeshan M, Jahan N, Haaris Sheikh M. Child psychiatry: A scientometric analysis 1980–2016. F1000Research. 2017;6: 1293. pmid:28944045
  26. 26. Chen C. CiteSpace II: Detecting and Visualizing Emerging Trends. J Am Soc Inf Sci. 2006.
  27. 27. Chen C. The centrality of pivotal points in the evolution of scientific networks. Proceedings of the 10th international conference on Intelligent user interfaces—IUI ‘05. 2005.
  28. 28. Clarivate analytics. Web of Science platform [Internet].
  29. 29. Salminen J, Almerekhi H, Dey P, Jansen B. Inter-rater agreement for social computing studies. Proceedings of The Fifth International Conference on Social Networks Analysis, Management and Security (SNAMS—2018). Valencia; 2018.
  30. 30. Perspective API. Alphabet [Internet]. 2018 [cited 21 May 2019].
  31. 31. Small H. Co‐citation in the scientific literature: A new measure of the relationship between two documents. J Am Soc Inf. 1973;24: 265–9. Available:
  32. 32. Raskauskas J, Stoltz AD. Involvement in traditional and electronic bullying among adolescents. Dev Psychol. 2007;43: 564–575. pmid:17484571
  33. 33. Smith PK, Mahdavi J, Carvalho M, Fisher S, Russell S, Tippett N. Cyberbullying: Its nature and impact in secondary school pupils. J Child Psychol Psychiatry Allied Discip. 2008;49: 376–385. pmid:18363945
  34. 34. Dehue F, Bolman C, Völlink T. Cyberbullying: Youngsters’ Experiences and Parental Perception. CyberPsychology Behav. 2008;11: 217–223. pmid:18422417
  35. 35. Slonje R, Smith PK. Cyberbullying: Another main type of bullying?: Personality and Social Sciences. Scand J Psychol. 2008;49: 147–154.
  36. 36. Erdur-Baker Ö. Cyberbullying and its correlation to traditional bullying, gender and frequent and risky usage of internet-mediated communication tools. New Media Soc. 2010;12: 109–125.
  37. 37. Juvonen J, Gross EF. Extending the school grounds?—Bullying experiences in cyberspace. J Sch Health. 2008;78: 496–505. pmid:18786042
  38. 38. Ybarra ML, Mitchell KJ. Prevalence and Frequency of Internet Harassment Instigation: Implications for Adolescent Health. J Adolesc Heal. 2007;41: 189–195. pmid:17659224
  39. 39. Calvete E, Orue I, Estévez A, Villardón L, Padilla P. Cyberbullying in adolescents: Modalities and aggressors’ profile. Comput Human Behav. 2010;26: 1128–1135.
  40. 40. Borrajo E, Gámez-Guadix M, Pereda N, Calvete E. The development and validation of the cyber dating abuse questionnaire among young couples. Comput Human Behav. Elsevier Ltd; 2015;48: 358–365.
  41. 41. Mitchell KJ, Finkelhor D, Wolak J. Youth Internet Users at Risk for the Most Serious Online Sexual Solicitations. Am J Prev Med. 2007;32: 532–537. pmid:17533070
  42. 42. Zweig JM, Dank M, Yahner J, Lachman P. The Rate of Cyber Dating Abuse Among Teens and How It Relates to Other Forms of Teen Dating Violence. J Youth Adolesc. 2013;42: 1063–1077. pmid:23412689
  43. 43. Mishna F, Cook C, Gadalla T, Daciuk J, Solomon S. Cyber Bullying Behaviours among Middle and High School Students. Am J Orthopsychiatry. 2010;80: 362–374. pmid:20636942
  44. 44. Kowalski RM, Giumetti GW, Schroeder AN, Reese HH. Chapter 14 Cyber Bullying Among College Students: Evidence from Multiple Domains of College Life. 2012. pp. 293–321.
  45. 45. Marcum CD, Higgins GE, Ricketts ML. Potential Factors of Online Victimization of Youth: An Examination of Adolescent Online Behaviors Utilizing Routine Activity Theory. Deviant Behav. 2010;31: 381–410.
  46. 46. Ybarra ML, Mitchell KJ. Youth engaging in online harassment: Associations with caregiver-child relationships, internet use, and personal characteristics. J Adolesc. 2004;27: 319–336. pmid:15159091
  47. 47. Aylett R, Vala M, Sequeira P, Paiva A. FearNot!–An Emergent Narrative Approach to Virtual Dramas for Anti-bullying Education. Virtual Storytelling Using Virtual Reality Technologies for Storytelling. Berlin, Heidelberg: Springer Berlin Heidelberg; pp. 202–205.
  48. 48. Finn J, Banach M. Victimization Online: The Downside of Seeking Human Services for Women on the Internet. CyberPsychology Behav. 2000;3: 785–796.
  49. 49. Madigan S, Villani V, Azzopardi C, … DL-J of A, 2018 undefined. The prevalence of unwanted online sexual exposure and solicitation among youth: a meta-analysis. ElsevierPaperpile.
  50. 50. Communication MA-CJ of, 2015 undefined. Towards a cybervictimology: Cyberbullying, routine activities theory, and the anti-sociality of social media. cjc-online.caPaperpile.
  51. 51. Atighetchi M, Pal P, Webber F, Jones C. Adaptive use of network-centric mechanisms in cyber-defense. Proceedings—6th IEEE International Symposium on Object-Oriented Real-Time Distributed Computing, ISORC 2003. 2003.
  52. 52. Agrawal S, Awekar A. Deep Learning for Detecting Cyberbullying Across Multiple Social Media Platforms. 2018. pp. 141–153.
  53. 53. Pitsilis GK, Ramampiaro H, Langseth H. Effective hate-speech detection in Twitter data using recurrent neural networks. Appl Intell. 2018;48: 4730–4742.
  54. 54. Thompson L. “I can be your Tinder nightmare”: Harassment and misogyny in the online sexual marketplace. Fem Psychol. 2018;28: 69–89.
  55. 55. Anzovino M, Fersini E, Rosso P. Automatic Identification and Classification of Misogynistic Language on Twitter. 2018. pp. 57–64.
  56. 56. Farkas J, Schou J, Neumayer C. Platformed antagonism: racist discourses on fake Muslim Facebook pages. Crit Discourse Stud. 2018;15: 463–480.
  57. 57. Chen C. The CiteSpace Manual v1.05. College of Computing and Informatics. 2015.
  58. 58. Gorman JO, Shum DHK, Halford WK, Ogilvie J. World Trends in Psychological Research Output and Impact. 2012;1: 268–283.
  59. 59. Diepeveen S, Larivi V, Pollitt A, Grant J, Montr D. International comparative performance of mental health research, 1980–2011. 2013; 1340–1347. pmid:23452564
  60. 60. Confraria H, Mira M, Wang L. Determinants of citation impact: A comparative analysis of the Global South versus the Global North. Res Policy. Elsevier B.V.; 2016;
  61. 61. Moed HF. The impact-factors debate: the ISI’s uses and limits. Nature. 2002;415: 731. pmid:11845176
  62. 62. Web of Science Core Collection: Web of Science: Summary of Coverage [Internet]. 2019.