Skip to main content
Browse Subject Areas

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Understanding disciplinary vocabularies using a full-text enabled domain-independent term extraction approach

  • Erjia Yan ,

    Roles Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing

    Affiliation College of Computing and Informatics, Drexel University, Philadelphia, Pennsylvania, United States of America

  • Jake Williams,

    Roles Formal analysis, Methodology, Visualization, Writing – original draft, Writing – review & editing

    Affiliation College of Computing and Informatics, Drexel University, Philadelphia, Pennsylvania, United States of America

  • Zheng Chen

    Roles Formal analysis, Methodology, Software, Writing – original draft, Writing – review & editing

    Affiliation College of Computing and Informatics, Drexel University, Philadelphia, Pennsylvania, United States of America


Publication metadata help deliver rich analyses of scholarly communication. However, research concepts and ideas are more effectively expressed through unstructured fields such as full texts. Thus, the goals of this paper are to employ a full-text enabled method to extract terms relevant to disciplinary vocabularies, and through them, to understand the relationships between disciplines. This paper uses an efficient, domain-independent term extraction method to extract disciplinary vocabularies from a large multidisciplinary corpus of PLoS ONE publications. It finds a power-law pattern in the frequency distributions of terms present in each discipline, indicating a semantic richness potentially sufficient for further study and advanced analysis. The salient relationships amongst these vocabularies become apparent in application of a principal component analysis. For example, Mathematics and Computer and Information Sciences were found to have similar vocabulary use patterns along with Engineering and Physics; while Chemistry and the Social Sciences were found to exhibit contrasting vocabulary use patterns along with the Earth Sciences and Chemistry. These results have implications to studies of scholarly communication as scholars attempt to identify the epistemological cultures of disciplines, and as a full text-based methodology could lead to machine learning applications in the automated classification of scholarly work according to disciplinary vocabularies.


The bibliometric community has used scientific publications as an effective instrument to study scholarly communication. Traditionally, bibliometric indicators were employed to assess research impacts [13]. Recent advances in bibliometrics have benefited from the use of network and statistical approaches to map science [46] and identify author communities [710]. Publication metadata, such as authors, journals, and references, were primarily used as the unit of analysis in these prior endeavors. The use of a more content-rich component—full-texts—was largely absent. Consequently, we made great efforts in examining research metadata but not research contents.

The composition of the research landscape is evolving—data, particularly scientific data, are increasing becoming open and accessible. The increased access to data not only provides more efficient means of analyses, but also entails a paradigmatic shift in modes of inquiry as scientists now can form diverse teams surrounded by data and conduct data-intensive research. The success of this transformation requires the use of new methods to extract more granular and content-rich information from large publication data. This need is within the realm of information extraction since computational linguists have developed methods to identify terms that can be used to describe domain-specific concepts from texts. While modern natural language processing techniques have yielded satisfying results on recall and precision, they were primarily employed with the objective of retrieval, as opposed to understanding. Accordingly, systematic approaches are lacking on how to utilize these methods to understand the latent meanings of the texts of scientific publications and how to use them to address questions on scholarly communication.

Thus, the objectives of this paper are two-fold. First, it is motivated to develop a term weighting-based method to extract content-rich terms from full texts. These terms can be broadly perceived as expressions in texts that convey information about the research-relevant aspects of publications, such as methods, theories, and concepts. Second, it uses the extracted terms to compare and contrast disciplines’ vocabularies—these vocabularies are important signifiers of disciplinary discourse patterns and can be used to reveal the epistemological differences in disciplinary cultures, as Hyland [11] argued that “writing…[o]n the contrary, it helps to create those disciplines”. The newly developed term extraction method allows us to examine the epistemological differences in a heretofore unattained extent, which complements the scholarship of the language aspect of disciplinarity studies that were largely confined to analyze samples of articles [12], dissertations [13], textbooks [14], and book reviews [15].

The paper provides insights into disciplinary vocabulary patterns and reveals scholarly communication at a new contextualized level. Conducting content-rich disciplinarity studies has the readily apparent advantage of gaining concrete and fine-grained perceptions of how different scientific concepts are embedded and relate to each other. It also helps us obtain an in-depth understanding of the production and dissemination patterns of scientific knowledge, innovations, and influences. By automatically extracting large and disciplinarily specific vocabularies, the satisfaction of this work’s goals also opens avenues for large-scale applications through algorithms that may use these rich lexica as feature inputs for machine learning.

Literature review

Recent years have witnessed a growing interest in term extraction. The term extraction task is concerned with two concepts, unithood and termhood. Unithood deals with the syntactics of terms and is formally defined as “the degree of strength or stability of syntagmatic combinations of collections” [16]. Termhood focuses on the semantic representation of terms or in Kageura and Umino’s words “the degree that a linguistic unit is related to…domain-specific concepts” [16]. Scholars have employed both linguistic and statistical methods to extract terms with unithood and termhood in mind from a variety of textual genres, such as email correspondences [17], scientific publications [18], and the Web [19]. Applications range from bioinformatics to studies of political parties [20]—giving rise to a new research area called named entity recognition and classification (NERC) [21]. Named entities encompass a variety of actors and artifacts such as people, locations, organizations, and biomedical entities. Three types of NERC methods are present: unsupervised, semi-supervised, and supervised. They are introduced in this section.

Unsupervised methods use lexical resources (e.g., WordNet or Web queries) [22, 23] and lexical patterns (e.g., the “such as” pattern) [21] to extract named entities. This approach has advantage because it provided a high-level validity of unithood. As for termhood, scholars have differentiated the weight of noun phrases according to certain measures, such as an entropy-based index [24], a context-based term weighing method [25], or a uniqueness-based indicator that compares word frequencies between scientific and non-scientific corpora [18, 26]. The idea behind the uniqueness-based indicator (a.k.a “weirdness”) is that terms in scientific and non-scientific corpora have disparate frequencies, making it is possible to use standard non-scientific corpora, such as the British National Corpus, to filter technical terms from scientific corpora [18, 26]. Reports of high precision [18] suggest potential for this method in applications to domain-independent corpora.

Semi-supervised methods typically use a bootstrapping technique. This technique recursively learns the contextual patterns of a small number of seed terms and uses the learned patterns to select new terms. Bootstrapping is an “effective, interpretable” [27] method and has performed well at extracting domain-dependent terms relating to terrorism [28], law [29], and medicine [30]. Supervised methods primarily include maximum entropy models [31], support vector machines [32], decision trees [33], hidden Markov models [34], and conditional random fields [35, 36]. These methods perform well, and extract named entities using labeled data [37]. However, requirements for large training data with entity-class associations result in a high complexity of O(D×R), where D is the number of documents and R is the number of relations [19]. Thus, developing Web-scale, domain-independent methods is recognized as a priority in the NERC community. Milestone events in this vein of research include the KnowItAll [19] and TextRunner systems [38]. KnowItAll was the first published domain-independent system, according to Etzioni, Banko [19]. The performance of the system, however, was impeded by the high volumes of Web query requests and the system-wide adjustment every time a new relation was added. TextRunner resolved these scalability issues and is seen as a “fully implemented” open information extraction (OIE) system [19]. It supported the discovery of new entity-class associations and reduced the complexity to O(D).

Methods and data

Term extraction

This paper employs a new term extraction method for full-text scientific copra developed in our prior research [39]. The method processes texts through StanfordNLP for lemmatization and part-of-speech (POS) tagging. Original texts, lemmas and POS tags then go through the POS matching procedure to identify possible terms, i.e., candidates, which can be a word or a phrase of multiple words. Candidates are further scored by our term extraction algorithm, which will be discussed in the following paragraphs. After POS tagging, each word is associated with a POS tag (Penn Treebank: [40]). POS tag sequences are matched by POS patterns. We chose the following pattern to match candidate terms: (({JJ}|({NN}[{VBG}|{VBN}]?)|{CD})+{IN})?({JJ}|({NN}[{VBG}|{VBN}]?)|{CD})*{NN}, where JJ denotes adjectives, NN denotes nouns, VBG denotes verbs in gerund or present participle, VBN denotes verbs in past participle, CD denotes cardinal numbers, and IN denotes prepositions. This pattern is selected based on our heuristic observations that valid terms are a combination of nouns, prepositions, adjectives, and verbs in present or past participles. Table 1 shows several examples of terms extracted by this pattern.

Table 1. Examples of extracted terms using the defined POS pattern.

This pattern is capable of including as many POS structures as possible; however, because it is designed to maximize recall, the candidate terms are considerably noisy (see the last two examples in Table 1). To select valid terms from these candidates, formula (1) is adopted to score each candidate.

In formula (1), Cd(x) is the C-Value formula given by Frantzi, Ananiadou [25], and W1(x), W2(x) are our weighting functions. C-Value relies on features such as term lengths and document-level frequencies. These features, however, cannot effectively distinguish scientific concepts from ordinary expressions. Thus, we propose an extension to include two types of frequency lists in scoring to improve performance.

We define the type I frequency list as a list of word frequencies acquired from a non-scientific corpus, for example, news articles and novels. In this work, we obtain a type I list from the Project Gutenberg eBooks repository for our method ( We then employed the sigmoid function to weigh down candidate terms that contain high-ranking words in the type I reference list. where ω is a word in the term x, and rω is the rank of ω in the type I frequency list. If ω is not in the list, we let rω = +∞. is a sigmoid function. r0 is a constant that controls when the function takes value 0.5. For example, if r0 = 4000, then g(ω) = 0.5 if rω = 4000, meaning a word ranked 4000 in the frequency list will be mapped to a sigmoid value 0.5. The other constant s1 controls the function’s rate of increasing with respect to rω. An increase of s1 will make the sigmoid increase slower, and a decrease of s1 will make the sigmoid increase faster. which is used to normalize W1(x) between 0 and 1. This reflects the idea of uniqueness of scientific terms—top words in daily language tend not to appear in them.

We define a type II frequency list as a list of word frequencies from a scientific terminology dictionary. For some words, such as “protein” and “behavior”, there presence in a candidate boosts the likelihood of them being a scientific term. Moreover, we observe that the position of a word in a term can also be used to determine the likelihood. For example, the word “central” is unlikely to appear at the end of any valid term and the word “theory” tends to tail valid terms. A second weighting function is designed as: where ω is a word in the term x, and fω is the frequency of ω in the type II frequency list. If ω is not in the list, we let fω = 0. Similar to W1(x), is also a sigmoid function. s2 is a slope parameter as described before for W1(x). Also we let for the purpose of normalization. We define the normalized position (NP) of a word ω in x as its zero-based position in x divided by |x|−1, the length of x minus one. For example, in term “central limit theorem”, “central” is at position 0 and is of NP 0, “limit” is at position 1 and is of NP “0.5”, and “theorem” is at position 2 and of NP “1”. is the average normalized position of word ω of all terms that generate the type II frequency list. As a result is a deviation of ω’s position in x from ω’s average position in the terminology dictionary, or simply the position disagreement, and is the position agreement. In addition, β ∈ [0,1] is a weighting parameter that controls on what level position disagreement affects W2(x).

A comprehensive type II reference list is relatively less available than a type I list. We recommend the use of acronyms to build up a type II list. Currently, the type II list is constructed from the literature itself based on recognized acronyms. About 10 thousand acronyms were recognized from our PLoS ONE corpus by using a simply rule-based approach (i.e., the existence of title case capitalization and parentheses) and we used these acronyms to build a type II list. It is worth noting that this term extraction method gives a higher score to words on the type II list and can find terms outside of this list. The benefit of using a type II list is that it can be considered comprehensive or well-rounded with respect to the literature corpus we analyze, and it reduces the method’s dependence on external resources.

Our method is advantageous because it does not rely on any corpus-level features such like document frequency, and thus it is able to process publications without first processing the whole corpus. We showed in [39] that our term extraction method outperformed the state-of-the-art, C-Value method and summarize this key finding in Table 2.

Precision in Table 2 refers to the ratio of the number of technical relevant terms among the top 20 extracted terms over all top 20 terms. Three human coders conducted the evaluation over 50 documents; a term is considered as a non-technical relevant term when a consensus was reached among all three coders. Recall refers to keyword recall; we used keywords as the gold standard when evaluating recall of the three methods. We see from the evaluation results that our method improved precision by 15% when scored(x) = cd(x) × W1(x) was used and by 17% when scored(x) = cd(x) × W1(x) × W2(x) was used. In regards to recall, our method (scored(x) = cd(x) × W1(x) × W2(x)) boosted the recall by 18% compared with the C-Value method.


The dataset used in this paper contains 52,981 PLoS ONE articles published between 2006 and 2015. The access point to the corpus is provided by PLoS ONE ( and it is freely accessible to the public. For each article in the dataset, we applied the designed term extraction method and selected top 40 terms based on scored. In total, we collected 532,725 unique terms from the dataset.

To examine disciplinary vocabularies, papers in the dataset need to be grouped into appropriate disciplines. PLoS has a classification scheme that assigns a paper to two or more research areas. We noticed that some research areas are quite similar and thus reclassified the research areas into 12 broader disciplines based on research similarities (Table 3).

Multi-counting was adopted in that a paper is counted in each discipline it was assigned into (see S1 Table for the summary of papers with multiple-subject assignments). For instance, a paper assigned into Biology and life sciences and Chemistry in the original scheme was counted once in Biology and once in Chemistry in the reclassified scheme. The advantage of this counting method is that it avoids counting a paper in an arbitrary discipline [41]. The caveat, however, is that it blurred the disciplinary boundaries and we should be cautious when interpreting results on interdisciplinarity. Another limitation of the employed classification scheme is that despite PLoS ONE’s multidisciplinary scope, this journal has a more extensive coverage on biomedical-related topics. This has raised a dilemma to us: on the one hand, we are interested to include in our analysis more representative domain-specific journals; on the other hand, most domain-specific journals outside biomedicine are not open access and thus would hinder the study’s reproducibility. As more journals are having open access options, we see the use of open access, domain-specific journals to examine disciplinary vocabularies as a future research direction.


An overview of disciplinary vocabularies

Table 4 shows the number of publications, number of unique terms, and terms per paper for each discipline.

While Biology and Medicine had the highest numbers of unique terms, Engineering possessed the highest terms per paper (tpp = 35.70). Other disciplines that resulted in high terms per paper include Agriculture (tpp = 20.52), Mathematics (tpp = 19.11), and Physics (tpp = 18.30). The result suggests that these disciplines tended to use nomenclatures more frequently in texts. Meanwhile, Computer and Information Sciences (tpp = 8.05) and Social Sciences (tpp = 8.96) yielded the lowest terms per paper. The result that Computer and Information Sciences had fewer terms per paper comes as a surprise, because this field is often seen as technology-driven. Thus, it is expected that it pertains to more frequent uses of technical terms—similar to the case of Engineering. It is possible that given PLoS ONE’s multidisciplinary nature, papers accepted by it in the area of computer science are primarily on the application track and used fewer technical terms than those on the theory and method tracks that are often published in domain-specific journals and conference proceedings.

To gain an understanding of the basic research themes of each discipline, we show top terms of each discipline. We first calculated the ratio of the occurrence of a term in one discipline against its occurrences in all 12 disciplines. This ratio helps bring terms that have a more salient association with certain disciplines. We then set up a threshold of 1,000 documents, meaning that a term needs to occur in at least 1,000 documents to be considered as top terms. Finally, we ranked the ratios in the descending order, with results shown in Table 5.

Table 5 shows that the designed term extraction method is able to identify both words and phrases from texts. These words and phrases are capable of depicting the general research themes of each discipline. For instance, Computer Science examines algorithms and networks; Earth Sciences is centered with environmental studies; Ecology focuses on biodiversity and conservation, Medicine discusses T cells and vaccinations; and Social Sciences seem to cover topics on both qualitative and quantitative research methods.

Term distributions

In this subsection, we show the distributions of terms over disciplines (Fig 1) and documents (Figs 2 and 3). Fig 1 illustrates the numbers of terms that occurred in one or more disciplines. Note no term was found only in one discipline because papers were assigned to at least two disciplines in PLoS ONE.

Fig 2. The rank-frequency distribution of terms in documents.

Fig 3. The rank-frequency distribution of terms in documents for 12 disciplines.

More than 40% of the terms (223,512) were associated with two disciplines and as the number of disciplines increases, the number of terms declines. There are 1,340 terms (0.25%) that occurred in all 12 disciplines. Among these, the following terms are the ones with the highest document-level occurrence (document-level occurrences in parentheses): gene (33,594), protein (23,143), antibody (18,709), USA (14,180), enzyme (12,855), and apoptosis (12,356). Meanwhile, for terms that only occurred in two disciplines, those with the highest document-level occurrence are CD27 (51), CD8 t-cell response (48), tetramer staining (27), pulmonary macrophage (26), centrosome amplification (24), and hematopoietic stem cell transplantation (24). We can see that terms in the latter group are more granular.

Fig 2 shows the distribution of terms over documents in a log-log scale. The y-axis shows the number of documents a term occurs and the x-axis shows the order of terms, from the term that occurs in the highest number of documents to the terms that occur in just two documents.

A power law pattern is visible in Fig 2 because of the linear distribution between the number of documents and the rank of terms in the log-log scale with base 10. This pattern shows that while most terms only occurred in a small number of documents, some terms occurred in most documents [42], such as gene, protein, antibody, USA, enzyme, apoptosis, and genome that occurred in more than 10,000 documents. In the meantime, there are 202,536 terms that only occurred in two documents.

We now zoom in to examine the term document distribution in each discipline. Fig 3 shows the distributions of terms over documents for all 12 disciplines in a log-log scale with the same axis compositions.

Distributions in Fig 3 share a similar pattern with Fig 2 in that there is a power law relationship between the number of documents and the rank of terms. Curves’ slopes are consistent, exhibiting a parallel form among the curves. The difference, however, is the intercept on the y-axis: Biology, for instance, had the largest number of terms and documents and thus has the highest intercepting value; Earth Sciences, on the other hand, has the lowest intercepting value.

Cross-discipline vocabulary similarities

We employed a principal component analysis (PCA) to measure disciplines’ similarities based on the vocabularies disciplines used in scientific publications. The PCA was applied to a 12 by 532,725 discipline-term matrix where each cell contains the number of occurrences of a term in a discipline. The largest two components in PCA accounted for 97.8% of the total variance (95.4% and 2.4% respectively) and they are visualized in Fig 4.

Fig 4. The plot of the first two components of the principal component analysis.

Several observations can be made regarding the PCA plot. First, all domains are located in the first and fourth quadrants because the loadings for each domain in the first component is non-negative. Second, Mathematics, Medicine and Health Sciences, Research and analysis methods, and Social Sciences are in the first quadrant, while the other domains are in the fourth quadrant. Third, Ecology, Earth Sciences, and Agriculture are closely located in the fourth quadrant. Also closely located in the fourth quadrant are Computer and Information Sciences, Engineering, and Physics. Fourth, Biology, Chemistry, and Medicine are seemingly far part on the plane, but because the first component accounted for more than 95% of the total variances, the three domains are in effect closely located when projecting them on the x-axis. This close relationship can be seen more clearly in Fig 5 where only the first component is visualized. Numbers before the domain names are their ranks projected on the x-axis in that Earth Sciences (0.01, 0) is the closest to the origin and Biology (0.80, 0) the furthest.

Fig 5. The plot of the first component of the principal component analysis.

To show the loadings of the first two components, we provide a stacked bar chart (Fig 6). For consistency, the same color coding of Fig 3 is used for Fig 6.

Biology has the largest loading in the first component, followed by Medicine, and Chemistry. Meanwhile, Earth Sciences, Research and Analysis Methods, and Social Sciences have the smallest loadings in the first component. The results indicate the dominant role of biochemical terms in the first component, which is not surprising given that a large percentage of PLoS ONE publications are classified under Biology, Chemistry, or Medicine. Besides these three domains, Ecology, Engineering, and Physics also made noticeable contributions to the first component. The first component, therefore, is led by biochemistry and physical sciences. While all domains’ loadings are non-negative in the first component, only four domains’ loadings are non-negative in the second component (Mathematics, Medicine, Research and Analysis Methods, and Social Sciences) and others’ have negative loadings. The fact that Medicine has the largest loading while Biology and Chemistry have the largest negative loadings show that the latter two domains are more closely related through their use of terms from Medicine. This difference, however, is very subtle, due to the small variances the second component contributed.


When we projected the 12 domains on the first component, we obtained a discipline similarity chain: Earth Sciences->Research and Analysis Methods->Social Sciences->Mathematics->Agriculture->Computer and Information Sciences->Physics->Ecology and Environmental Sciences->Engineering->Chemistry->Medicine and Health Sciences->Biology. We compare the chain with the “consensus map” in [43], which was created by merging 20 existing science maps and identifying the general proximity patterns of disciplines: starting from mathematics, there are “physics, physical chemistry, engineering, chemistry, earth sciences, biology, biochemistry, infectious diseases, medicine, health services, brain research, psychology, humanities, social sciences, and computer science”.

A few similarities can be identified between the term occurrence map (i.e., the chain) and the consensus map: in the consensus map, social science and computer science are collocated and in the term occurrence map, Social Sciences and Computer and Information Sciences are also closely located to each other, separated by two domains Mathematics and Agriculture. Physics, engineering, and chemistry are collocated in the consensus map while Physics, Ecology, Engineering, and Chemistry are collocated in the term occurrence map. In the consensus map, biology and a few medical science domains are collocated while in the term occurrence map, Medicine and Biology are collocated. A few differences can be found, which are understood through two main factors: first, the PCA plot obtained from this study is based on 12 broad knowledge domains whereas previous science maps were based on more-granular units, such as the 27 Scopus major subject areas [44], 220 Web of Science subject categories [45], thousands of journals [5, 46], or millions of documents [47]. The more-granular units provided the possibility of using richer dimensions to depict the similarity between scientific fields. Second, previous science maps were created upon article- or journal-level co-citation relationships, whereas the plot in Fig 4 was created through discipline-level term occurrence relationships.

This paper found several disciplinary vocabulary use patterns. While Engineering tended to use nomenclatures more frequently in texts, Computer and Information Sciences and Social Sciences had the lowest numbers of terms per paper. In addition, based on a principal component analysis, this paper found that Mathematics and Computer and Information Sciences had a similar vocabulary use pattern, as did the pairing of Engineering and Physics. According to Carnap [48], disciplines in each of the abovementioned pairs pertained to “a very narrow and homogeneous class of terms of the physical thing-language”. Meanwhile, Chemistry and Social Sciences exhibited contrasting vocabulary use patterns, as did Earth Sciences and Chemistry. The results may have implications to studies of scholarly communication as scholars attempt to identify the epistemological cultures of disciplines [11], find disciplinary knowledge paths [49, 50], promote interdisciplinary research and collaborations [51], and design effective indicators to assess research outputs [52].

The term extraction method developed in this paper complements current work in co-word analysis. Co-word analysis can be very useful in portraying the cognitive space of a variety of disciplines [53, 54]. It is predicated upon a few assumptions [55]: keywords symbolize “non-trivial relationship between their referents” and indexers assign reliable keywords to refer to scientific concepts. In reality, however, due to the so-called “indexer effect” [56, 57], these assumptions may not be fully met, and the performance of co-word analyses may be impaired. To alleviate this tension, there is a growing interest of using title or abstract words instead of keywords [58]. The choice over title and abstract words may grant a “more direct access to the view of authors” [55] and may create a richer content to extract scientific concepts and ideas. The presented term extraction method can provide further refinement to co-word analysis from two aspects: first, this method is capable of extracting noun phrases—this holds clear advantage over single word-based extraction because many scientific concepts contain more than one word. Second, because of its weighting mechanism, the presented method can be applied to full texts that are richer in content and the extracted terms of a paper are weighed and ranked and the most distinctive terms can be used in co-word analysis.


This paper employed an efficient, domain-independent term extraction method to extract disciplinary vocabularies from a large multidisciplinary corpus of PLoS ONE publications. The employed method can effectively extract content-rich terms from unstructured bibliometric fields such as full texts used in this study. Extracted terms can help researchers and practitioners contextualize findings and make sense of bibliometric indicators and numbers. Examinations of the extracted terms can help reveal the scholarly communication at a new granular level and address questions on the provenance, diffusion, coevolution, trend, and impact of knowledge at a much improved extent and depth. Analyzing and modeling content-rich terms also complements the state-of-the-art data infrastructure that orients towards network analysis of publications [6, 5961]. In addition, this paper also found a power law pattern of the distribution of terms over documents: a small number of terms occurred in most documents while most others only occurred in a limited number of documents. This distribution pattern was also present for documents in each discipline, indicating the existence of a large and natural quantity of discipline-specific terms sufficient to application of statistical analyses machine learning algorithms.

Developing effective term extraction methods applicable to the full texts of scientific literature is the first step of a greater effort to enable content-aware bibliometric research. Next step will likely include the design and application of automated methods to induct taxonomies to organize extracted terms, in addition to the development of statistical methods and machine learning algorithms that may leverage extracted terms to automatically classify the ever-growing scientific literature. Thus, in the context of increasing interdisciplinarity, this effort should have long-term benefits to knowledge management and information retrieval. Future work will also include the employment of statistical- and network-based methods to understand the lifecycle of innovations as codified by content-rich terms.

Supporting information


  1. 1. Merton RK. The Matthew effect in science. Science. 1968;159(3810):56–63.
  2. 2. Hirst G. Discipline impact factors: A method for determining core journal lists. Journal of the American Society for Information Science. 1978;29(4):171–2.
  3. 3. Cronin B. The citation process. The role and significance of citations in scientific communication. London: Taylor Graham, 1984. 1984;1.
  4. 4. Börner K, Contractor N, Falk-Krzesinski HJ, Fiore SM, Hall KL, Keyton J, et al. A multi-level systems perspective for the science of team science. Science Translational Medicine. 2010;2(49):49cm24–49cm24. pmid:20844283
  5. 5. Boyack KW, Klavans R, Börner K. Mapping the backbone of science. Scientometrics. 2005;64(3):351–74.
  6. 6. Chen C. CiteSpace II: Detecting and visualizing emerging trends and transient patterns in scientific literature. Journal of the American Society for information Science and Technology. 2006;57(3):359–77.
  7. 7. Ahn Y-Y, Bagrow JP, Lehmann S. Link communities reveal multiscale complexity in networks. Nature. 2010;466(7307):761–4. pmid:20562860
  8. 8. Leskovec J, Lang KJ, Dasgupta A, Mahoney MW, editors. Statistical properties of community structure in large social and information networks. Proceedings of the 17th international conference on World Wide Web; 2008: ACM.
  9. 9. Girvan M, Newman ME. Community structure in social and biological networks. Proceedings of the national academy of sciences. 2002;99(12):7821–6.
  10. 10. Wang L, Lou T, Tang J, Hopcroft JE, editors. Detecting community kernels in large social networks. 2011 IEEE 11th International Conference on Data Mining; 2011: IEEE.
  11. 11. Hyland K. Disciplinary Discourses, Michigan Classics Ed.: Social Interactions in Academic Writing. University of Michigan Press; 2004.
  12. 12. Dahl T. Textual metadiscourse in research articles: a marker of national culture or of academic discipline? Journal of pragmatics. 2004;36(10):1807–25.
  13. 13. Demarest B, Sugimoto CR. Argue, observe, assess: Measuring disciplinary identities and differences through socio-epistemic discourse. Journal of the Association for Information Science and Technology. 2015;66(7):1374–87.
  14. 14. Hyland K. Academic attribution: Citation and the construction of disciplinary knowledge. Applied linguistics. 1999;20(3):341–67.
  15. 15. Tse P, Hyland K. Gender and discipline: Exploring metadiscourse variation in academic book reviews. Academic discourse across disciplines. 2006:177–202.
  16. 16. Kageura K, Umino B. Methods of automatic term recognition: A review. Terminology. 1996;3(2):259–89.
  17. 17. Minkov E, Wang RC, Cohen WW, editors. Extracting personal names from email: Applying named entity recognition to informal text. Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing; 2005: Association for Computational Linguistics.
  18. 18. Drouin P. Term extraction using non-technical corpora as a point of leverage. Terminology. 2003;9(1):99–115.
  19. 19. Etzioni O, Banko M, Soderland S, Weld DS. Open information extraction from the web. Communications of the ACM. 2008;51(12):68–74.
  20. 20. McCallum A, Li W, editors. Early results for named entity recognition with conditional random fields, feature induction and web-enhanced lexicons. Proceedings of the seventh conference on Natural language learning at HLT-NAACL 2003-Volume 4; 2003: Association for Computational Linguistics.
  21. 21. Nadeau D, Sekine S. A survey of named entity recognition and classification. Lingvisticae Investigationes. 2007;30(1):3–26.
  22. 22. Alfonseca E, Manandhar S, editors. An unsupervised method for general named entity recognition and automated concept discovery. Proceedings of the 1st international conference on general WordNet, Mysore, India; 2002.
  23. 23. Collins M, Singer Y, editors. Unsupervised models for named entity classification. Proceedings of the joint SIGDAT conference on empirical methods in natural language processing and very large corpora; 1999: Citeseer.
  24. 24. Van Eck NJ, Waltman L, Noyons EC, Buter RK. Automatic term identification for bibliometric mapping. Scientometrics. 2010;82(3):581–96. pmid:20234767
  25. 25. Frantzi K, Ananiadou S, Mima H. Automatic recognition of multi-word terms: the c-value/nc-value method. International Journal on Digital Libraries. 2000;3(2):115–30.
  26. 26. Gillam L, Tariq M, Ahmad K. Terminology and the construction of ontology. Terminology. 2005;11(1):55–81.
  27. 27. Gupta S, Manning CD, editors. Improved Pattern Learning for Bootstrapped Entity Extraction. CoNLL; 2014.
  28. 28. Thelen M, Riloff E, editors. A bootstrapping method for learning semantic lexicons using extraction pattern contexts. Proceedings of the ACL-02 conference on Empirical methods in natural language processing-Volume 10; 2002: Association for Computational Linguistics.
  29. 29. Nallapati R, Manning CD, editors. Legal docket-entry classification: Where machine learning stumbles. Proceedings of the Conference on Empirical Methods in Natural Language Processing; 2008: Association for Computational Linguistics.
  30. 30. Gupta S, Manning CD. Spied: Stanford pattern-based information extraction and diagnostics. Sponsor: Idibon. 2014;38.
  31. 31. Borthwick A, Sterling J, Agichtein E, Grishman R, editors. Exploiting diverse knowledge sources via maximum entropy in named entity recognition. Proc of the Sixth Workshop on Very Large Corpora; 1998.
  32. 32. Krauthammer M, Nenadic G. Term identification in the biomedical literature. Journal of biomedical informatics. 2004;37(6):512–26. pmid:15542023
  33. 33. Prokofyev R, Demartini G, Cudré-Mauroux P, editors. Effective named entity recognition for idiosyncratic web collections. Proceedings of the 23rd international conference on World wide web; 2014: ACM.
  34. 34. Collier N, Nobata C, Tsujii J-i, editors. Extracting the names of genes and gene products with a hidden Markov model. Proceedings of the 18th conference on Computational linguistics-Volume 1; 2000: Association for Computational Linguistics.
  35. 35. Torii M, Hu Z, Wu CH, Liu H. BioTagger-GM: a gene/protein name recognition system. Journal of the American Medical Informatics Association. 2009;16(2):247–55. pmid:19074302
  36. 36. Yan E, Zhu Y. Identifying entities from scientific publications: A comparison of vocabulary-and model-based methods. Journal of Informetrics. 2015;9(3):455–65.
  37. 37. Sutton C, McCallum A. An introduction to conditional random fields for relational learning. Introduction to statistical relational learning. MIT Press; 2006.
  38. 38. Banko M, Cafarella MJ, Soderland S, Broadhead M, Etzioni O, editors. Open Information Extraction from the Web. IJCAI; 2007.
  39. 39. Chen Z, Yan E. Domain-Independent Term Extraction & Term Network for Scientific Publications. iConference 2017; Wuhan, China2017.
  40. 40. Marcus MP, Marcinkiewicz MA, Santorini B. Building a large annotated corpus of English: The Penn Treebank. Computational linguistics. 1993;19(2):313–30.
  41. 41. Yan E, Ding Y, Cronin B, Leydesdorff L. A bird's-eye view of scientific trading: Dependency relations among fields of science. Journal of Informetrics. 2013;7(2):249–64.
  42. 42. Zipf GK. Human behavior and the principle of least effort: An introduction to human ecology: Ravenio Books; 2016.
  43. 43. Klavans R, Boyack KW. Toward a consensus map of science. Journal of the American Society for information science and technology. 2009;60(3):455–76.
  44. 44. Yan E, Yu Q. Using path-based approaches to examine the dynamic structure of discipline-level citation networks: 1997–2011. Journal of the Association for Information Science and Technology. 2016;67(8):1943–55.
  45. 45. Yan E. Finding knowledge paths among scientific disciplines. Journal of the Association for Information Science and Technology. 2014;65(11):2331–47.
  46. 46. Leydesdorff L, Carley S, Rafols I. Global maps of science based on the new Web-of-Science categories. Scientometrics. 2013;94(2):589–93. pmid:23335826
  47. 47. Waltman L, Eck NJ. A new methodology for constructing a publication-level classification system of science. Journal of the American Society for Information Science and Technology. 2012;63(12):2378–92.
  48. 48. Carnap R. Logical foundations of the unity of science. International Encyclopaedia of Unified Science. 1991;1(1):393–404.
  49. 49. Yan E, Yu Q. Using path-based approaches to examine the dynamic structure of discipline-level citation networks: 1997–2011. Journal of the Association for Information Science and Technology. 2015.
  50. 50. Yan E. Disciplinary knowledge production and diffusion in science. Journal of the Association for Information Science and Technology. 2015.
  51. 51. Stokols D, Hall KL, Taylor BK, Moser RP. The science of team science: overview of the field and introduction to the supplement. American journal of preventive medicine. 2008;35(2):S77–S89.
  52. 52. Waltman L, van Eck NJ, Visser M, Wouters P. The elephant in the room: The problem of quantifying productivity in evaluative scientometrics. arXiv preprint arXiv:160502378. 2016.
  53. 53. Callon M, Courtial JP, Laville F. Co-word analysis as a tool for describing the network of interactions between basic and technological research: The case of polymer chemsitry. Scientometrics. 1991;22(1):155–205.
  54. 54. Coulter N, Monarch I, Konda S. Software Engineering as seen through its Research Literature: A Study in co-word Analysis. Journal of the American Society for Information Science. 1998;49(13):1206–23.
  55. 55. He Q. Knowledge discovery through co-word analysis. Library trends. 1999;48(1):133-.
  56. 56. Leydesdorff L, Zaal R. Co-words and citations relations between document sets and environments. 1988.
  57. 57. Whittaker J. Creativity and conformity in science: Titles, keywords and co-word analysis. Social Studies of Science. 1989;19(3):473–96.
  58. 58. Bhattacharya S, Basu P. Mapping a research area at the micro level using co-word analysis. Scientometrics. 1998;43(3):359–72.
  59. 59. Team S. Science of Science (Sci2) Tool.: Indiana University and SciTech Strategies; 2009
  60. 60. Waltman L, van Eck NJ, Noyons EC. A unified approach to mapping and clustering of bibliometric networks. Journal of Informetrics. 2010;4(4):629–35.
  61. 61. De Nooy W, Mrvar A, Batagelj V. Exploratory social network analysis with Pajek. Cambridge University Press; 2011.