Increasing public interest in science information in a digital and 2.0 science era promotes a dramatically, rapid and deep change in science itself. The emergence and expansion of new technologies and internet-based tools is leading to new means to improve scientific methodology and communication, assessment, promotion and certification. It allows methods of acquisition, manipulation and storage, generating vast quantities of data that can further facilitate the research process. It also improves access to scientific results through information sharing and discussion. Content previously restricted only to specialists is now available to a wider audience. This context requires new management systems to make scientific knowledge more accessible and useable, including new measures to evaluate the reach of scientific information. The new science and research quality measures are strongly related to the new online technologies and services based in social media. Tools such as blogs, social bookmarks and online reference managers, Twitter and others offer alternative, transparent and more comprehensive information about the active interest, usage and reach of scientific publications. Another of these new filters is the Research Blogging platform, which was created in 2007 and now has over 1,230 active blogs, with over 26,960 entries posted about peer-reviewed research on subjects ranging from Anthropology to Zoology. This study takes a closer look at RB, in order to get insights into its contribution to the rapidly changing landscape of scientific communication.
Citation: Fausto S, Machado FA, Bento LFJ, Iamarino A, Nahas TR, Munger DS (2012) Research Blogging: Indexing and Registering the Change in Science 2.0. PLoS ONE 7(12): e50109. https://doi.org/10.1371/journal.pone.0050109
Editor: Matjaz Perc, University of Maribor, Slovenia
Received: January 28, 2012; Accepted: October 18, 2012; Published: December 12, 2012
Copyright: © 2012 Fausto et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Funding: The authors have no support or funding to report.
Competing interests: The authors have read the journal's policy and have the following conflicts: co-authors Atila Iamarino, Luiz Bento and Tatiana R. Nahas are administrators for Portuguese version of Research Blogging, which is owned and operated by Seed Media Group, LLC (Seed). They are not employed and do not receive any payment for this work, as per Research Blogging policies. By the time of this paper submission, David S. Munger used to be an administrator at the English version of the site, but he is not related to Research Blogging anymore. There are no further patents, products in development or marketed products to declare. This does not alter the authors' adherence to all the PLOS ONE policies on sharing data and materials, as detailed online in the guide for authors.
The instruments and methodologies from Bibliometrics and Scientometrics traditionally cooperate in and are widely used by development agencies, academic institutions, and even corporations for planning and management of policies for Science and Technology (S&T), identification and promotion of new areas of research, and many other issues in strengthening and growth of S&T activities.
Bibliometrics and Scientometrics tools provide statistics and indicators to generate measures of published scientific output. Although admittedly imperfect –, this field is mainly based on the number of publications and citations. In fact, as S. Arbesman has written,
For too long, the measurement of scientific contribution has centered on the publication. Whether through the number of articles, the citations those articles have by other articles, or even other far more complicated metrics, most scientists are still measured by a derivative of the research article, the basic technology of scientific publishing that is well over 300 years old .
This is a more than 300 year-old modus operandi of science communication, which began with the invention of the scientific journal in the 17th century  and was well suited to communicating scientific research results for a long time in a world where scientists published their findings, theories and ideas to other scientists. But it is insufficient for the current context of an increasing public interest in science information in a digital and 2.0 science era, where the scientific community is witnessing a dramatic, rapid and deep change. The emergence and expansion of information and communication technologies and internet-based tools is opening space for new possibilities to improve both scientific methodology and communication, assessment, promotion and certification .
New technologies allow modern methods of acquisition, manipulation and storage, generating massive data volumes that can further facilitate the research process ,. These technologies also facilitate access to scientific results through information sharing and discussion. Content previously restricted only to specialists is now available to a wider audience.
This context requires new management systems to make scientific knowledge more accessible and useable, including new measures to evaluate the reach of scientific information not only among professionals and specialists but also to the general public. The new science and research quality measures are strongly related to the new online technologies and services based in social media. Tools such as blogs, social bookmarks, online reference managers (CiteULike, Connotea, Mendeley, Zotero), and Twitter offer alternative, transparent and more comprehensive information about the active interest, usage and reach of scientific publications –. External online tools also represent a new form of post-publication review (e.g. Wikipedia referencing of articles is an indicator of future citations ), a result of the filtering done by specialist authors.
All these changes are stimulating the scientific community to reassess its means of communication. For example, the Science Online conference, now in its sixth edition (in January 2012) aims to explore science on the web , encouraging studies have been released , alternative metrics as PLoS Article-Level Metrics have been developed –, and all of these developments have helped to grow movements such as the new field of Altmetrics . These new tools are based on a belief in the failure and insufficiency of the three more traditional filters - peer-review, citation counting analysis, and Journal Impact Factor - to indicate the most relevant and significant sources in a context of an explosive growth of the volume of academic literature in today's internet-age science.
Here we highlight scientific blogs as one important new filter of scientific research. The science blogosphere has grown significantly in recent years. The information gap that was traditionally fulfilled by science journalists and scientifically-curious laymen now has a new protagonist: the scientist. Blogs are one of the most common methods that scientists use to communicate their ideas to other scientists or to the general public . This preference may be due to incentives for scientists to engage with the blogosphere  and face its challenges to traditional peer-reviewed research channels. But these challenges may also be a great opportunity , enabling scientists to make a direct connection to students – and bringing them closer to the general public. Scientific blogs have a positive tendency for aggregation, mainly through blog platforms developed by respected science journals or through new tools that either allow a new system of science publishing  and post-publication filtering or value online peer-reviewed publication.
This study aims to describe the platform Research Blogging, an aggregator of scientific blog citations of peer-reviewed publications, showing its history, current configuration and characterization of languages, covered topics, number of blogs, posts, use of Open Access (OA), and mentions of scientific and other research. We see it as a critical tool in the ever-changing world of scientific communication, with its own important contribution to this change in the science endeavor.
Research Blogging: background, current state and characterization
Research Blogging (RB) was created in 2007 by the scientific blogger Dave Munger, after one of his readers showed appreciation for his use of an icon to distinguish posts about peer-reviewed research from other general or personal messages on his blog. An icon for all scientific blog posts citing peer-reviewed research was developed, and then a central aggregator collected all such marked posts in a collection harvested from across the internet. Soon, hundreds of bloggers were using the site and a new platform [http://researchblogging.org] was developed and is still maintained in collaboration with Seed Media Group. The RB Website aggregates peer-reviewed research posts from several science blogs in seven different languages: English, Spanish, Portuguese, German, Chinese, Polish and Italian. It is a useful source for readers interested in cutting-edge research and first-hand comments and explanations of science, by scientists and experts in their respective fields. In addition, given that the intrinsic structure of the web makes it difficult to a clear distinction between scientific and pseudo-scientific content, RB is a tool to identify serious academic research and avoid the spread of pseudo-scientific contents, serving as a self-regulated organization that helps to collect only academically relevant information. The site now has over 1,230 active blogs, with over 26,960 entries posted about peer-reviewed research on subjects ranging from Anthropology to Zoology, in categorized blogs.
How Research Blogging works
All RB content is user generated. Participating bloggers - often experts in their research area - identify relevant research in their field. When they write substantive posts about the research on their blogs, they can choose to have those posts aggregated by RB. RB serves as a central means of disseminating findings of peer-reviewed research that careful bloggers have found interesting enough to read and closely analyze.
After registration, bloggers decide themselves to which category their blog will belong indicating their blog topics from the available list within RB site:
Once registered in RB, bloggers use a one-line form to create a snippet of code to place in their posts. This snippet not only notifies the RB site about the scientific posts, it also creates a properly formatted research citation for the blog. The RB software automatically scans registered blogs for posts containing RB code snippet. When it finds them, it indexes and displays them on site front page — thousands of posts from hundreds of blogs, organized by topic. RB editors identify the notable posts in each major discipline, publishing the results on news page in the platform. Other services like PubGet [http://pubget.com] index the RB database as well, so every time readers search for a journal article, they can also locate blog posts discussing the article, and RB also uses sharing tools for divulgation through RSS feeds and social media applications (app) as Twitter.
Participating bloggers agree to use the “Blogging on Peer-Reviewed Research” icons and the aggregator at ResearchBlogging.org only when they are writing a thoughtful, original blog post about peer-reviewed research. Just a linking to or quoting a news article or press release is not considered sufficient for inclusion on RB.
Blogs can be a powerful tool for dissemination of scientific information and RB is one of the tools that promote a self-regulated quality control of blog posts. Bloggers must demonstrate to the RB editors and readers that they regularly produce posts that meet the criteria to use a “blog badge” . RB editors ensure that newly-registered blogs follow guidelines based on weeks of discussion at ResearchBlogging.org community to safeguard the quality of the aggregator platform. The site continues to receive further recommendations and suggestions for modifications to these guidelines, which are subject to ongoing revision so as to maintain the spirit of good scholarship. The quality of the posts listed on RB site is monitored by the blogger members. If a post doesn't follow the guidelines, it is removed from RB database, and borderline cases may be discussed publicly on the RB blog as well.
The following extract, taken directly from the RB site, describes the most important guidelines for inclusion:
- The “Blogging on Peer-Reviewed Research” icons are to be used solely to denote individual blog posts about peer-reviewed research;
- Similarly, when a blogger is registered with ResearchBlogging.org and uses our system to generate a citation for purposes of aggregation by our site, the citation is to be used solely to denote individual blog posts about the peer-reviewed research listed in the citation;
- While there is no hard-and-fast definition of “peer-review,” peer reviewed research should meet the following guidelines:
- *Reviewed by experts in field
- *Published with clearly stated publication standards
- *Viewed as trustworthy by experts in field
- *In the case of certain curated archives such as arXiv.org, the “intention” for research to be reviewed may be seen as an adequate proxy for peer review
- Posts using the icon or RB citation code should offer a complete formal citation of the work(s) being discussed;
- The post author should have read and understood the entire work cited;
- The blog post should report accurately and thoughtfully on the research it presents;
- Where possible, the post should link to the original source and/or provide a Document Object Identifier (DOI) or other universal reference number;
- The post should contain original work by the post author — while some quoting of others is acceptable, the majority of the post should be the author's own work;
- Users and readers may report potential abuse of the icons and aggregation system by flagging the post on RB site. Reported abuses may be brought to the attention of readers and discussed publicly online.
There are previous studies about Research Blogging, focusing in its characterization as areas covered, journal titles cited, bloggers' gender and anonymity and other aspects , . Our study expands to a closer look to RB, in order to get insights into its contribution to the changes which we verify in scientific communication.
Data collection and treatment
We conducted an exploratory study, with a quantitative approach to guide the search into posts by the Research Blogging Website. The search was performed in January 2012 and included the entire period available in RB since its inception, considering the posts published between November 1, 2007 and December 31, 2011. We chose to analyze only posts actively discussing peer-reviewed articles published in scientific journals, and excluded posts that merely listed references with no discussion. Citations in posted entries with references to books, conference proceedings, guidelines and other online or offline sources were disregarded. We also disregarded those without an active online address and no longer available – only six blogs with a total of 12 posts.
Data were extracted, we hand-searched reference lists from retrieved posts to verify inconsistencies, and then the treated data were summarized in order to generate quantitative descriptions of the following:
- *number of blogs
- *categories by RB topic
- *distribution among the seven languages adopted by RB
- *number of posts
- *citation distribution (number of articles cited by post, journal titles, in restricted journals and in OA journals)
- *reach by number of views.
In addition to generating automatic references on RB by searching for the DOI from scholarly papers, bloggers can create references manually when DOI is unknown, and thus they do not follow a single standard to refer to the journals, e.g. the Proceedings of the National Academy of Sciences of the United States of America appears in full, abbreviated by Nat. Acad. Sci. USA, and by PNAS. This was the case of some journals in the extracted sample, and this sort of lapse in standardization is a common problem in data mining for informetric research , thus any sample obtained automatically must be checked for find inconsistencies and be previously treated to a valid analysis. Here we confirm the titles by the consultation to Ulrich's Periodicals Directory Online [http://ulrichsweb.serialssolutions.com].
Two of the original topics present in RB, Health and Medicine, were joined into a single topic, Health Sciences, and their data values were added to facilitate the analysis under a single category. For counts of views, we consider unique views for each post, and a view for each article cited in this post; i.e. two articles in one post were considered to be viewed two times, while one view was assigned to the post. For all other analysis, we consider simple counts. The access status of periodicals in search for Open Access journals was accessed by consultation to the Directory of Open Access Journals (DOAJ) [http://www.doaj.org, last accessed in January 2012].
Statistical analysis and comparison among metrics
We counted the blog citations and post visualizations for each scientific journal cited in the RB database. We obtained 7 scientometric measurements for the journals available at Journal Citation Reports (JCR) from Thomson Reuters, namely: Journal Impact Factor, Total Articles, Total Citations, Half-Life, Immediacy Index, Eigenfactor Score and Article Impact. We evaluated the correlations among RB count variables and JCR metrics through the Spearman's ρ statistic. The significance of the correlations was accessed through a permutation test (9,999 permutations) and were evaluated at the level of α = 0.05. Additionally, to investigate if Open Access policy would influence citations (i.e. OA articles were more cited than expected), we compared the proportion of OA blog citations to the proportion of OA articles in the sample through a binomial test. These analyses were performed in the R programing environment v. 2.14.2 .
Our results below were extracted from the raw data which are available in Supporting Information [Spreadsheet S1], in accordance with the scenario for science 2.0, with data spreading and sharing .
Totals by Blogs and Posts by RB topics and Journal Titles by Area
During the period analyzed, the Research Blogging website collected, registered, indexed and shared 26,969 posts by 1,236 blogs considering all entries in total [Fig. 1]. The RB topic with the most posts was Biology, with 9,787 posts (36%), followed by Health Sciences (here combined with Health and Medicine), with 4,177 posts (15%). Psychology had 3,401 posts (13%), Neuroscience had 2,495 (9%), Social Science 1,108 (4%), Anthropology 1,058 (4%), Chemistry 879 (3%), Physics 835 (3%), Geosciences 518 (2%), Research/Scholarship 438 (2%), Astronomy 407 (2%), Computer Science/Engineering 239 (1%), Ecology/Conservation 221 (1%), Philosophy 152 (1%) and finally Mathematics with 77 posts. The Other RB topic category had 1,177 posts (4%) [Fig. 2].
RB posts indexed since its creation.
The most common language was English with 1,008 blogs and 22,660 posts, followed by Portuguese, with 65 blogs and 1,013 posted entries. Spanish had 52 blogs with 1,456 posts, German had 36 blogs and 742 posts, Italian had 32 blogs with 449 posts, Polish had 24 blogs and 512 posts, and Chinese had 19 blogs with 137 posted entries [Fig. 3 and Table 1].
English is supported since RB inception in 2007. The other languages were added gradually (German, August 2008; Spanish, May 2009; Portuguese, June 2009; Chinese, August 2009; Polish, April 2010; Italian, December 2010).
Within the analyzed period 19,000 RB posts cited and linked 26,154 scientific papers published in 3,350 different journals [Fig. 4]. The most-covered subject area by journal titles was the Health Sciences, with 1,071 titles, followed by Applied Social Sciences with 796 titles. Biological Sciences had 599 journal titles, Exact & Earth Sciences, 530 titles while the Multidisciplinary area had 308 titles and the Humanities 46 journal titles [Fig. 5].The journals cited 1,000 times or more were Science (1,829 times), Nature (1,803), Proceedings of the National Academy of Sciences USA – PNAS (1,372) and PLoS ONE (1,156): all general purpose periodicals [Table 2]. This result is similar to the sequence found by Shema, Bar-Ilan & Thelwall  in a minor sample of RB posts, putting these four first journal titles in a “Golden Circle” on the Research Blogging website. The citation trend does not follow a close relation to Impact Factors (IF), and all three groups of most cited journals have some periodicals with high IFs and some with IFs of less than 10.
Only posts citing peer-reviewed research from periodicals were considered.
The second most-cited group of journals spanned from 201 to 350 citations and include six journal titles; the third most-cited group was those with between 101 and 200 citations — 18 titles. There were 36 journals with 51 to 100 citations, 134 journals with 20 to 50 citations, 581 journals with 5 to 19 citations, 1,059 journals with 2 to 4 citations, and 1,512 journals with one citation.
From the 3,350 journals listed in the RB database, 1,822 had scientometric information available at JCR. The correlation matrix shows a moderately modular structure [Fig. 6]. The lowest correlations were associated with Article Half-Life, showing a mean correlation of 0.18 with other metrics and non-significant correlations with both RB count variables. The Total Number of Articles also seems to have generally low correlations with other metrics, with values ranging from 0,16 to 0,26, except for Total Citation and Eigenfactor Score (0.74 and 0.75, respectively). Apart from those variables, all JCR metrics shows correlations among themselves that ranges from 0.54 to 0.97 (0.94, excluding 5year based IF), with an average of 0.58, even if we exclude the 5year based IF. In contrast, RB counts have correlations with the JCR metrics (except Total Articles and Half-Life) that ranges from 0,32 to 0,42, with a mean correlation of 0.37. RB counts showed an average correlation of 0.27 with Total Articles. The correlation between RB citations and RB counts was 0.88.
Open Access Journals
The results showed that 11.7% of the citations (3,054 of 26,154) came from Open Access journals, a value four times larger than that observed in Wikipedia citations - 2.8% . These citations come from the 7.2% OA journals present in our sample (241 of 3,350). The differences between the proportion of OA citations and OA articles available were significant under a binomial test (p = 2.079e−144). Unlike in Wikipedia citations, six of the most cited journals were OA [Table 2]: PLoS ONE in First group (with more than 1,000 citations), Psychological Science and PLoS Biology in Second group (with between 201 and 350 citations) and PLoS Medicine, Pediatrics and PLoS Pathogens in Third group (with 101 to 200 citations). Also, when visits were considered, three of the 10 most visited article links were of OA journals: PLoS One, Psychological Science and PLoS Biology.
As explained in Methods, for view count we considered unique views for each post, and a view for each article cited in this post, i.e. to two articles in one post were two separate views, one for each article [Fig. 7], and only one view for the post. As expected, results showed that more cited journals obtained higher numbers of overall views, but this is also true for some less cited journals, which obtained high number of views too [Fig. 8]. The opposite trend was found to individual article from journals often cited that in some cases did not obtain a high number of views. When we analyzed the views for unique articles - not journals - some surprising views were seen: the most-viewed article was from the Journal of Applied Animal Welfare Science, which has an IF of only 0.71. It received 62,217 views, well ahead of second place, an article in Proceedings of the Royal Society B: Biological Sciences (IF 5.064), which had 15,265 views.
Article views (in thousands) are represented according to citing posts at Research Blogging. Most recent articles have less time to accumulate views.
Blogs and Posts by RB topics
Both the registered blog totals and total number of posts are dominated by Biology (36% of posts). Health Sciences appears in a distant second position (15%), an interesting result since the area of Health Sciences dominates scientific communication, in a number of traditional publications , also verified here when we analyze the cites by the journal titles [Fig. 5]. Other categories have minor representation, with 13% (Psychology) and 9% (Neuroscience). Still others form a long tail of the site, with the remaining eleven categories taking less than 4 percent of the total number of posts on RB [Fig. 2]. It's possible that the topic distribution is due to the early dominance of Biology. Perhaps other disciplines saw RB as primarily a Biology/Health site and opted out.
Frequency of posts per year
The frequency of posts grew vigorously from the establishment of Research Blogging in November 2007, with the number of posts in 2009 doubling over 2008 [Fig. 1]. After a peak in 2010, in 2011 the number of posts declined to levels similar 2009. This increase in 2010 may be related to Research Blogging Awards 2010, since nominations started early February and winners were announced early March, 2010. Following this period, despite the addition of new blogs and languages, the number of posts and views have returned to values equal to or smaller than 2009. We consider the equivalency in posts from 2011 and 2009 an actual decrease in blog posts, since this number results from more blogs and languages that adopted RB during these two years. During the second half of 2011, the automatic aggregation tool of RB was not functional, which may have led to this decrease. This difference could be in part explained by a shift in science divulgation in recent years from blogs to other online platforms, such as social networks (e.g. Twitter and Facebook). These tools have different purposes and functionalities, and mainly in the case of science writing would be more a good complement for spreading scientific blog posts  and to collect and share stories and resources , rather than construction and discussion as observed in blogs, indicating that these new tools are more related to social networks acting in spread and sharing information, linking to contents (including to the blogs), whereas blogs would be considered as information repositories. With faster ways of propagation and discussion of topics in these new tools, the decrease in post numbers may represent shorter comments on articles left out of blogs and posted in social media instead, and that posts are less frequent but used to cover research more thoroughly.
Languages and RB Topics covered
The dominant language on RB is English, with 1,008 blogs posting 22,660 entries, followed by Portuguese (65 blogs and 1,013 posts), Spanish (52 blogs and 1,456 posts), German (36 blogs and 742 posts), Italian (32 blogs and 449 posts), Polish (24 blogs and 512 posts), and Chinese (19 blogs and 137 posts) [Fig. 3, Table 1].
English has been supported for the longest period at RB, having been a part of the system since its inception in 2007. The other languages were added gradually (German, August 2008; Spanish, May 2009; Portuguese, June 2009; Chinese, August 2009; Polish, April 2010; Italian, December 2010), and there is some correlation between when a language was added and the number of posts in that language. However, perhaps because more science publishing and blogging occurs in English, or because the RB interface is in English, English continues to substantially outpace the other languages.
There are also some interesting regional patterns. The general dominance of Biology is not the same in all languages. In Italian, there are more posts in Physics than Biology (174∶141), in Chinese there are more posts about Chemistry (37), Psychology (36) and Computer Science/Engineering (24) than Biology (11). Polish demonstrated an equilibrated distribution of topics [Table 1]. These regional peculiarities show an interesting avenue for future comparisons in scientific communication among different cultures.
Our correlation analysis shows that there is general decoupling of blog metrics and other classical scientific metrics (Fig. 6). This is exemplified by the fact that, generally, the correlation between RB counts and JCR metrics are lower than those observed among the majority of JCR metrics, with the exception of Article Half Life and Total Articles. This suggests that the main factors influencing journal citation in the blogosphere are not the same that determine journal merit, as evaluated through JCR metrics, even though academic merit have a substantial influence on blog citation, as reveled by the presence of significant correlations among almost all JCR metrics and RB counts.
The presence of this imperfect association between classical metrics and blog citation can be exemplified by the fact that high IFs are present in most-cited journals but are not a prerequisite or predictive of journal citations in posts [Table 2]. So, rather more than being more frequently cited due to high relevance due to IF, the “Golden Circle” may also be favored because it consists of multidisciplinary journals, while those with fewer citations are specialized journals, with a more restricted audience. We consider the wide variety of journals that were discussed to be a positive feature of RB, although almost half of the titles was only referenced once in the study time period.
These findings in general draw attention to the importance of new article level metrics and other scientometrics tools for measuring the relevance of papers outside traditional publications –. Also, as articles cited in Wikipedia tend to be more relevant than equivalents, an indicative that the choice of Wiki citations favors relevant research , it may be interesting to follow if citations in blog posts are predictive of future article relevance.
Another interesting finding was that increasingly blogs cite more articles in the same post. One post had 29 citations, 18 of which refer to articles that are part of a series derived from a project proposal to the National Science Foundation (NSF) . This difference supports the argument that blogs promote a deeper understanding of the subjects they cover and the hypothesis that bloggers are writing less frequently, but dealing with more relevant information. In fact, while the number of posts and citations fell in absolute terms in 2011 [Fig. 2, 3], the number of citations per post did not. This number has increased from 1.38 in 2010 to 1.48 in 2011, which may indicate that bloggers are beginning to add more content to each post. Also, blogs tend to cite more types of sources than just peer-reviewed articles, leading to questions about online metrics: Are mentions of published scientific research at blogs or Wikipedia as valid as citations? Should we reconsider what we commonly understand by citation: an article talking about another article? These are important questions, since the process of scientific communication is historically based on procedures which don't necessarily have analogs in a digital and 2.0 context, where we are looking for new, valid metrics for assessing the reach and impact of science and research .
Open Access journals
There is a large, ongoing effort to promote and disseminate Open Access scientific journals, motivated by the idea that scientific information must flow freely to generate more knowledge . Our findings show that the number of OA journals cited by RB posts is much larger than observed for Wikipedia citations , suggesting that blog authors have favored OA content, and blog readers have proportionally more access to the original article discussed at the posts. Recently, there has been an increasing concern in publication policy and public access to research results , , and academic bloggers are especially engaged in these matters , which may reflect in OA trends. The large presence of paid content journals indicates that bloggers still maintain some of the characteristics of traditional scientific discourse as a preference for high-impact and multidisciplinary journals, following findings in others studies about RB , . On the other hand, we suggest that they perform an important social function by exposing and explaining scientific content that is inaccessible to the general public due to the constraints of paid access scientific journals in a transition context permeated by the effort to the greater access to scientific knowledge.
The results corroborate the methodologies of Article-Level Metrics that consider the individual article to determine its value and reach, in contrast to journal-level measures of research quality that have traditionally been made available until now , as an alternative form to verify the quality, importance, and relevance to scientific literature, more immediately than the IF allows. One of the criteria of article-level metrics - the number of views to the article - allows verify the article relevance soon after the publication unlike journal-level measures based in IF.
In addition, as P. Janiszewski points out, citation on blogs may improve the reach of research:
Put another way, the same research which I published in a prestigious medical journal and made basically no impact, was then viewed by over 12,000 sets of eyes because I decided to discuss it online. And it doesn't end there .
The systematic indexing and citation registering adopted by RB is an efficient filter for published research and its dissemination, allowing article views and access statistics agree with blog coverage metric .
Extracting data from RB posts is a challenge, mainly due to the heterogeneous pattern of journals entries by the bloggers, as previously explained. Also regarding RB further improvements, it will be useful to allow its data to be mined by integrating features like its Twitter app with tools like CrowdoMeter ,, improving the categorization of the citations in RB posts, and integrating other tools to promote a joint effort with the scientific community. Additionally, it would be informative to deeper evaluate the regional patterns observed between languages, allowing comparisons in scientific communication among different cultures.
The emergence and rise of more recent online technologies and services based in social media tools such as Twitter may mean that blogs, one of the oldest digital platforms, are losing ground in numbers. We believe that blogging is still an important way to give visibility to science in a more complete and detailed format. It can offer an alternative view of science, one that is more transparent, comprehensive, and comprehensible, while increasing interest, usage and reach of scientific publications; it continues to hold an important place among other new technologies. Platforms like RB not only spread but also record and index published research, as well as having an important social function by bringing restricted publications of science to the general public.
Also, it points to a new path of scientific information spreading. The previous (and somehow still ongoing) path was: 1) scientific data published in traditional scientific journals; 2) press releases; 3) scientific data divulged (not always accurate) in the mass media. An important new ongoing path is: 1) scientific data published in traditional scientific journals and also in open access scientific journals; 2) peer-reviewed posts published in science blogs, which provides updated and accurate scientific information in more accessible language to a non-scientific public. Considering this, it would be a relevant challenge to develop and/or improve new metrics related to tools like RB in order to better evaluate its effective contribution to scientific information reach.
In this sense, our correlation analyses suggest that RB citations and views indeed evaluate different aspects of scientific production. The fact that the correlations between RB counts and JCR metrics is lower then the correlations among JCR variables (with the exception of Total Articles and Article Half-Life) suggests that the overall factors influencing the traditional metrics are not the main factors in defining blogging citations and views. If the pattern found here for JCR metrics are consistent with large-scale studies of correlation between different metrics , than this could be an indicative that RB-based metrics are evaluating a different feature of journal quality, merit or impact. Even if RB counts are connected to Usage metrics (e.g. Closeness Centrality, Degree Centrality, Journal Use Probability), the mean correlation between those and Citation metrics is very high (according to Bollen et al , it ranges 0.68 to 0.73, with the exception of Usage Impact Factor, with a value of 0.27), strongly suggesting that RB counts are evaluating a different aspect of research quality. Specific investigations of the relationship between Usage metrics and RB counts are warrant in order to evaluate the true relation of these metrics. Overall, RB metrics correlations are consistent with findings for other altmetrics , indicating that they should be viewed as such.
Even though RB counts would not be available to all journals (not all journals are cited in blogs), they nevertheless state something about the social impact of those that were cited, and could be of use to journal editors that wish to develop policies to increase their journal outreach. Large publishers (such as Nature group) are already doing this through the establishment of a blogosphere linked to their publications. RB is different in this sense because it is not directly connected to any scientific publishing group and could be seen as a relatively independent source of scientometric information, and a more reliable base for policy-making.
Research Blogging Reports raw data. Excel spreadsheet with Research Blogging data from November 1, 2007 until December 31, 2011. Sheet S1-A: RB Blog Report with blog name, blog URL, status, Research Blogging topic, number of posts and blog language. Sheet S1-B: RB Citations Report with publication date, post title, number of views, blog name, DOI, journal title and Research Blogging topic.
We thank Rogério Mugnaini (Escola de Artes, Ciências e Humanidades, USP) and Gilson L. Volpato (Departamento de Fisiologia, Instituto de Biociências, Caunesp, UNESP) for encouragement and kind advice on data analysis and discussion.
Conceived and designed the experiments: SF LB AI TRN DM. Performed the experiments: SF LB AI DM. Analyzed the data: SF LB AI DM FAM. Contributed reagents/materials/analysis tools: SF LB AI. Wrote the paper: SF LB AI TRN DM FAM.
- 1. MacRoberts M, MacRoberts B (2010) Problems of citation analysis: A study of uncited and seldom–cited influences. Journal of the American Society for Information Science and Technology 61 (1) 1–12
- 2. Anderson K (2009) The impact factor: A tool from a bygone era? The Scholarly Kitchen (29 June). Available: http://scholarlykitchen.sspnet.org/2009/06/29/is-the-impact-factor-from-a-bygone-era. Accessed 15 January 2012.
- 3. Rosner M, Van Epps H, Hill E (2007) Show me the data. Journal of Experimental Medicine 204 (13) 1091–1092
- 4. Arbesman S (2012) New Ways to Measure Science. Social Dimension Wired Science Blogs (9 Jan). Available: http://www.wired.com/wiredscience/2012/01/new-ways-to-measure-science/all/1. Accessed 10 January 2012.
- 5. Wikipedia. Philosophical Transactions of the Royal Society. Available: http://en.wikipedia.org/wiki/Philosophical_Transactions_of_the_Royal_Society. Accessed 10 January 2012.
- 6. Lane J (2010) Let's make science metrics more scientific. Nature 464: 7288.
- 7. Molloy JC (2011) The Open Knowledge Foundation: Open Data Means Better Science. PLoS Biology 9 (12) e1001195
- 8. Schofield PN, Bubela T, Weaver P, Portilla L, Brown SD, et al. (2009) Post-publication sharing of data and tools. Nature 10: 171–173.
- 9. Li X, Thelwall M, Giustini D (2011) Validating online reference managers for scholarly impact measurement. Scientometrics [Online 1st 21 December] 1–11
- 10. Haustein S, Siebenlist T (2011) Applying social bookmarking data to evaluate journal usage. Journal of Informetrics 5 (3) 446–457.
- 11. Eysenbach G (2011) Can Tweets Predict Citations? Metrics of Social Impact Based on Twitter and Correlation with Traditional Metrics of Scientific Impact. Journal of Medical Internet Research 13 (4) e123
- 12. Groth P, Gurney T (2010) Studying Scientific Discourse on the Web using Bibliometrics: A Chemistry Blogging Case Study. In: Proceedings of the WebSci10, Raleigh, NC, US. Available: http://journal.webscience.org/308. Accessed 10 January 2012.
- 13. Bonetta L (2009) Should you be tweeting? Cell 139 (3) 452–453
- 14. Bogers T, van den Bosch A (2008) Recommending scientific articles using CiteULike. Proceedings of the ACM Conference on Recommender Systems, New York: ACM, 287–290. Available: http://ilk.uvt.nl/~toine/publications/bogers.2008.recsys2008-paper.pdf. Accessed 10 January 2012.
- 15. Beltrão P (2006) Connotea tag: Evolution citation report. Public Rambling (21 Nov). Available: http://pbeltrao.blogspot.com/2006/11/connotea-tagevolution-citation-report.html. Accessed 20 January 2012.
- 16. Evans P, Krauthammer M (2011) Exploring the use of social media to measure journal article impact. In: AMIA Annual Symposium Proceedings. 374–8. pmcd: PMC3243242.
- 17. Science Online 2012 Conference, 6th ed. Available: http://scienceonline2012.com. Accessed 22 January 2012.
- 18. Gu F, Widén-Wulff G (2011) Scholarly communication and possible changes in the context of social media. The Electronic Library 29 (6) 762–776
- 19. Public Library of Science (PLoS) Article-Level Metrics. Available: http://article-level-metrics.plos.org. Accessed 10 January 2012.
- 20. Yan K-K, Gerstein M (2011) The Spread of Scientific Information: Insights from the Web Usage Statistics in PLoS Article-Level Metrics. PLoS ONE 6 (5) e19917
- 21. Neylon C, Wu S (2009) Article-Level Metrics and the Evolution of Scientific Impact. PLoS Biology 7 (11) e1000242.
- 22. Priem J, Taborelli D, Groth P, Neylon C (2010) Altmetrics: a manifesto (26 Oct). Available: http://altmetrics.org/manifesto. Accessed 10 January 2010.
- 23. Bonetta L (2007) Scientists Enter the Blogosphere. Cell 129 (3) 443–445.
- 24. Editorial (2009) It's good to blog. Nature 457: 1058
- 25. Kouper I (2010) Science blogs and public engagement with science: practices, challenges, and opportunities. Journal of Science Communication 9 (1) 1–10.
- 26. Deitering A, Gronemyer K (2011) Beyond Peer-Reviewed Articles: Using Blogs to Enrich Students' Understanding of Scholarly Work. Portal: Libraries and the Academy 11 (1) 489–503.
- 27. Powell DA, Jacob CJ, Chapman BJ (2011) Using Blogs and New Media in Academic Practice: Potential Roles in Research, Teaching, Learning, and Extension. Innovative Higher Education [Online 1st 3 December] 1–12
- 28. Batts SA, Anthis NJ, Smith TC (2008) Advancing Science through Conversations: Bridging the Gap between Blogs and the Academy. PLoS Biology 6 (9) 1837–1841.
- 29. Akst J (2010) Publish or post? The Scientist Blog (9 Aug). Available: http://www.the-scientist.com/blog/display/57613. Accessed 10 January 2012.
- 30. Shema H, Bar-Ilan J, Thelwall M (2012) Research Blogs and the Discussion of Scholarly Information. PLoS ONE 7 (5) e35869
- 31. Munger D (2010) Blogging out of balance. Seed Magazine (22 Sept) Available: http://seedmagazine.com/content/article/blogging_out_of_balance. Accessed 27 January 2012.
- 32. Bar-Ilan J (2009) A closer look at the sources of Informetric Research. In: (2009) Proceedings of the 12th International Society for Scientometrics and Informetrics Conference Larsen B, Leta J, editors. Rio de Janeiro: BIREME/PAHO/WHO, UFRJ. 1: 1–9.
- 33. R Development Core Team (2012) R: A Language and Environment for Statistical Computing. Vienna: R Foundation for Statistical Computing.
- 34. Kowalczyk S, Shankar K (2011) Data sharing in the sciences. Annual Review of Information Science and Technology 45: 247–294.
- 35. Moya-Anegón F, Chinchilla-Rodríguez Z, Vargas-Quesada B, Corera-Álvarez H, Muñoz-Fernández FJ, et al. (2007) Coverage analysis of Scopus: A journal metric approach. Scientometrics 73 (1) 53–78
- 36. Fausto S, Iamarino A, Bento LFJ, Nahas TR (2011) Peer-reviewed science from blogs: an option for the Brazilian growing interest in science? In: (2011) Proceedings of the 13th International Society for Scientometrics and Informetrics Conference Noyons E, Ngulube P, Leta J, editors. Durban: ISSI, Leiden University, University of Zululand. 2: 957–977.
- 37. Rowan C (2011) How blogs, Twitter and other social media tools are changing conversations about scientific research. Earth (28 Apr). Available: http://www.earthmagazine.org/article/science-and-social-media. Accessed: 01 may 2012.
- 38. Kwak H, Lee C, Park H, Moon S (2010) What is Twitter, a Social Network or a News Media? In: International World Wide Web Conference Committee (IW3C2). Raleigh, NC. Available: http://an.kaist.ac.kr/~haewoon/papers/2010-www-twitter.pdf. doi:10.1145/1772690.1772751. Accessed: 04 May 2012.
- 39. Garcia M (2011) Forests and Water, part 1: Where there's smoke… Hydro-Logic (23 Jun). Available: http://hydro-logic.blogspot.com/2011/06/forests-and-water-part-1-where-theres.html. Accessed: 10 January 2012.
- 40. Priem J, Hemminger BM (2010) Scientometrics 2.0: Toward new metrics of scholarly impact on the social Web. First Monday, 15(7). Available: http://firstmonday.org/htbin/cgiwrap/bin/ojs/index.php/fm/article/view/2874/2570. Accessed 10 January 2012.
- 41. Harnad S (1991) Post-Gutenberg Galaxy: The Fourth Revolution in the Means of Production of Knowledge. Public-Access Computer Systems Review 2 (1) 39–53.
- 42. Taylor M (2012) Academic publishers have become the enemies of science. The Guardian, (Monday, 14 Jan). Available: http://www.guardian.co.uk/science/2012/jan/16/academic-publishers-enemies-science. Accessed: 06 may, 2012.
- 43. Eisen MB (2012) Research Bought, Then Paid For. The New York Times (10 Jan). Available: http://www.nytimes.com/2012/01/11/opinion/research-bought-then-paid-for.html?_r=3&scp=1&sq=PLoS%20NIH&st=cse. Accessed: 06 May 2012.
- 44. The price of information (2012) The Economist (4th Feb). Available: http://www.economist.com/node/21545974. Accessed: 05 May 2012.
- 45. Janiszewski P (2010) Why all scientists should blog: a case study. Science of Blogging (23 Nov) Available: http://scienceofblogging.com/why-scientists-should-blog-a-case-study. Accessed 15 January 2012.
- 46. CrowdoMeter (2011) Available: http://crowdometer.org/ratings. Accessed: 05 January 2012.
- 47. Fenner M (2011) CrowdoMeter – or trying to understand tweets about journal papers. Gobbledygook (20 Dec) Available: http://blogs.plos.org/mfenner/2011/12/20/crowdometer-or-trying-to-understand-tweets-about-journal-papers. Accessed: 05 January 2012.
- 48. Bollen J, Van de Sompel H, Hagberg A, Chute R (2009) A principal component analysis of 39 scientific impact measures. PLoS ONE 4 (6) e6022
- 49. Priem J, Piwowar HA, Hemminger BM (2012) Altmetrics in the wild: Using social media to explore scholarly impact. arXiv:1203.4745 [preprint]. Available: http://arxiv.org/html/1203.4745v1. Accessed: 22 April 2012.