Academia.eduAcademia.edu

Readership Data and Research Impact

2019, Springer Handbook of Science and Technology Indicators

Reading academic publications is a key scholarly activity. Scholars accessing and recording academic publications online are producing new types of readership data. These include publisher, repository, and academic social network download statistics as well as online reference manager records. This chapter discusses the use of download and reference manager data for research evaluation and library collection development. The focus is on the validity and application of readership data as an impact indicator for academic publications across different disciplines. Mendeley is particularly promising in this regard, although all data sources are not subjected to rigorous quality control and can be manipulated.

Readership data and Research Impact Ehsan Mohammadi1, Mike Thelwall 2 1School of Library and Information Science, University of South Carolina, Columbia, South Carolina, United States of America 2Statistical Cybermetrics Research Group, School of Mathematics and Computer Science, University of Wolverhampton, Wolverhampton, United Kingdom Abstract: Reading academic publications is a key scholarly activity. Scholars accessing and recording academic publications online are producing new types of readership data. These include publisher, repository and academic social network download statistics as well as online reference manager records. This chapter discusses the use of download and reference manager data for research evaluation and library collection development. The focus is on the validity and application of readership data as an impact indicator for academic publications across different discipline. Mendeley is particularly promising in this regard, although all data sources are not subjected to rigorous quality control and can be manipulated. Subject terms: readership; academic social network sites; reference managers; Mendeley 1.1 Introduction and Overview The act of reading an academic publication is a key point at which knowledge is transferred from the author to someone else. With the prevalence of the web and social web, scholars now often read and register academic publications online, leaving electronic records of their activities. This readership data can reveal which outputs are used as well as giving insights how the scientific enterprise works. It is important to exploit such information to improve research evaluation practices and to investigate how science communication is evolving. The main advantage of using readership information rather than citations is that reading occurs before citing and can therefore give more timely information. A second advantage is that the reading public for academic research is wider than the citing public since it includes students, professionals and others. This chapter reviews research about download and readership data for academic outputs from the perspective of its value for research evaluation. It is mainly concerned with journal articles but books and conference papers are also briefly mentioned. The chapter has extended coverage of readership information from online reference managers, such as Mendeley, because of their practical value for research assessment. It also discusses publisher usage data, since “using” an article in this context often means reading it. The usage statistics component complements the Usage Bibliometrics as a Tool to Measure Research Activity chapter in this Handbook, which gives an extended case study of arXiv and analyses research evaluation at a more systemic level. The current chapter also discusses usage data from the academic social network sites ResearchGate and Academica.edu which are changing the way in which articles are discussed and shared. Previous works have reviewed usage bibliometrics (Kurtz & Bollen, 2010), readership metrics (Haustein, 2014), and social media metrics (Thelwall & Kousha, 2015a). This chapter updates the previous readership metrics chapter (Haustein, 2014) with newer topics and findings. It is intended for research evaluators, scientometricians and bibliometricians as well as those interested in recent changes in the scholarly communication ecosystem. This chapter opens with a discussion of reading, terminology and research impact assessment to set the context for the ways in which readership data may be used. It continues with a discussion of online usage data from publishers and repositories, before analysing the effect of online reference managers on the availability of readership data and discussing social networks. Finally, the chapter discusses research evaluation applications. 1.2 Reading research: Background and terminology Although knowledge has been traditionally communicated in human societies orally and by imitation, the written record is a cornerstone of modern science. Academics read scholarly publications to inform their current research, for current awareness purposes, to support their teaching, or to help them fulfil a professional advisory role related to their expertise. In addition, other professionals may read journals to inform their day-to-day practice. This is the case for medical doctors, who need to be aware of the latest developments in medical practice that are relevant to their role or expertise (Tenopir et al., 2007). For example, 73% of non-publishing Canadian physicians read journal articles (McAlister, Graham, Karr, & Laupacis, 1999), and hospital residents consider journal articles to be valuable sources of information for them (Schilling, Steiner, Lundahl, & Anderson, 2005). Information about which academic documents are read and by whom can help the librarians that buy them, the managers and policy makers that need to evaluate the impact of the research produced, and the scholars that investigate science itself. Many different terms have been used for reading-related data, especially within electronic contexts. The most general is perhaps usage, which does not imply a reason why an item was accessed but is often employed in digital contexts as an umbrella term to describe accesses of all kinds of digital resource. It is helpful to use more specific terms, when justifiable, to aid interpretation. The more explicit term download refers to a local copy being taken of an electronic resource, such as a journal article, whether by a human or a robot. The terms hit and view refer to online accesses of electronic information without necessarily downloading it. For example, a digital library visitor might view a webpage containing the title, abstract and metadata of an article and then download a full-text copy of the article to their local computer. The term full-text download (Haque & Ginsparg, 2009) can be used for emphasis. Although the term download usually connotes accessing a full text version of an article, some publishers provide non- downloadable full-text copies of articles and readers may view full-text articles online without downloading them to their local computer storage (e.g., Grabowsky, 2016). The most specific term is read/reader/readership, which implies that the item accessed has been read by a human. This term can be justified for a data source if it is reasonable to believe that the items accessed will usually be read. For example, it would be reasonable to believe that a book is usually read when it is borrowed from a library. Thus, borrowing statistics could be claimed to be (non-exhaustive) reading indicators. In contrast, most accesses of documents on some websites may be from web crawlers and so it would not be reasonable to interpret download counts as readership counts (but see the paragraph below). Moreover, human users may also systematically download journal articles if they are concerned that their access will be interrupted before they know which articles to read (Emrani, Moradi-Salari, & Jamali, 2010). To describe download counts as a readership indicator, evidence would be needed to connect downloads to reading. Thus, if full text downloads from a specific source are likely to be of human origin and frequently lead to reading, then it would be reasonable to refer to full-text downloads as readership data. More specifically, and following an argument for citations (van Raan, 1998), to be a readership indicator, download or access counts should positively correlate with reader numbers, even if not all downloads/accesses lead to reading. If this condition is fulfilled, then the counts convey information in an information theoretic sense about how often an item has been read. To give an example, a positive correlation implies that if article X has more accesses than Y then there is a greater than 50% chance that X has been more read than Y. In general, the higher the correlation, the higher this chance is. Aggregating sets of documents also, in general, increases this chance. Thus, a moderate positive correlation between accesses and reading would give a high change that research group A’s outputs had been read more than research group B’s outputs if they had been accessed more on average. In practice, the exact number of readers of any academic document is never known because of ways of accessing documents that cannot be tracked, such as from print copies of journals. Statistical correlations therefore need to be supported by arguments that the untracked readings are rare compared to the tracked readings or that the tracked readings are likely to be a relatively unbiased sample of all accesses. Alternatively, if the tracked accesses can be directly shown to correlate with research impact or value then this additional step is unnecessary. 1.3 Readership data from libraries Two traditional sources of journal readership data are sales and library circulation information – either the number of libraries that hold a journal or circulation data for that journal. Both give journal-level rather than article-level evidence and are limited by the increasing share of reading that occurs online. For university libraries, the main journal readers are students, who are likely to be readers but not citers (Duy & Vaughan, 2006), and it is not clear whether they are also the main consumers of academic journals, or whether there are disciplinary differences in this. These students are likely to have different uses for academic journals and librarians need to consider this factor when analysing readership evidence to build collections (Rowlands & Nicholas, 2007). Circulation information has been analysed for a long time (Gross & Gross, 1927) from practical (White & McCain, 1989) and mathematical modelling (Egghe & Rousseau, 2000) perspectives to support librarians (Kurtz & Bollen, 2010). Some examples illustrate the approaches used and give background for more current strategies. The earliest usage records kept by libraries were lending or reshelving statistics and these have been used as proxies for usage or readership data for books and journals (King, Tenopir, & Clarke, 2006), even though they do not cover all ways in which they may be read (Davis, 2004). This data has been shown to correlate positively with Journal Impact Factors (JIFs), at least when comparing journals from the same discipline, and journals from specialisms that are similarly served by the library (Stankus & Rice, 1982; Tsay, 1998). This might be due to some readers subsequently citing articles in the journals or more cited journals tending to have more useful information. For example, Nature and Science target non-citing audiences in parallel with the acknowledged high quality of their scholarly content. In contrast, other journals primarily target a professional audience (e.g., nurses, librarians, lawyers) and may be less concerned to attract an academic readership. High correlations between online downloads and local readership information in one context give some evidence that downloads can be good indicators of readership (Emrani, Moradi-Salari, & Jamali, 2010). Inter-library lending information and direct observations of library users can also give readership information, although the latter is too time consuming to be routinely used (Butkovich, 1996). 1.4 Research impact assessment The publication of peer-reviewed research is a critical scholarly activity and the analysis of scholarly publications is important for assessing the research impact of scholars or teams in most disciplines. This assessment can be qualitative, quantitative or both. It can be combined with other sources of evidence or judgements and it can be made for formal evaluations, formative self-evaluations, or to investigate an aspect of science or science communication. 1.4.1 Peer review Peer review is the central evaluation mechanism for modern science (Kostoff, 1997). The value of research is judged by experts from the same field because non-specialists are less able to understand the work or its contribution to scholarship. Ultimately, however, the scholarly community is collectively responsible to the governments or others that fund them and so experts may adjust their expectations in response. For example, this may lead a community to regard work as better if it offers societal benefits. Peer review is also at the heart of the publication system, with journal articles being usually subject to critical evaluation before a decision is made about whether to accept them, and the same is often true for monographs (Wouters, 1997). It is also central to research funding applications (Jayasinghe, Marsh, & Bond, 2003), perhaps as one of a checklist of attributes to assess. In some countries, peer review is used to periodically assess research to allocate block research grants. In the UK, the Higher Education Funding Council for England (HEFCE) uses panels of field experts to evaluate the quality of research from government funded universities and other academic organizations to share out funding on a merit basis (HEFCE, 2011). The main disadvantage of peer review is that it is slow and expensive, consuming a substantial amount of expert time (ABRC, 1990; Smith, 2006). It is also fallible because reviewers may be consciously or unconsciously biased against others’ work based on personal characteristics, research competition or the research paradigm followed (Lee, Sugimoto, Zhang, & Cronin, 2013). Moreover, even expert reviewers can disagree about the merits of papers (Welch, 2014). Nevertheless, peer review is often the most credible single source of evaluation evidence for academic research. Although peer review primarily assesses prior work rather than predicting future performance, experts can also assess applicants’ plans if these are relevant for appointments and promotions. 1.4.2 Citation analysis Citation-based indicators are the primary quantitative tools to evaluate research, whether on their own, to support human judgements or to cross-check reviewers’ opinions (van Raan, 2000). Citation data may come from a recognized citation index, such as the Web of Science (WoS), Elsevier’s Scopus, or Google Scholar. There are also specialist citation indexes for some fields and others with a national or linguistic scope, such as the Chinese Citation Index. The best-known citation indicator is the Journal Impact Factor (JIF), which estimates the average number of citations to recently-published articles in a journal. The JIF is informally used as an indicator of the quality of a journal perhaps because of its simplicity and intuitive reasonableness. Nevertheless, there are numerous problems with the accuracy of the calculations and their ability to reflect impact (Egghe, 1988; Dellavalle, Schilling, Rodriguez, Van de Sompel, & Bollen, 2007). There are also major problems of over-interpretation, leading to inappropriate uses (Opthof, 1997; Boell & Wilson, 2010), such as those that ignore disciplinary differences. The average number of citations per publication is known to vary by field and year and so it is not reasonable to compare the average citation count between groups of publications. Field normalized indicators, such as the Mean Normalized Citation Score (MNCS) (Waltman, van Eck, van Leeuwen, Visser, & van Raan, 2011ab) and Mean Normalized Log-transformed Citation Score (MNLCS) (Thelwall, 2017ab) solve this problem by normalizing citation counts for the publishing field and year so that a score of 1 always means citation impact equal to the world average. For individual authors, the h-index has become popular (Hirsch, 2005), although it is biased towards senior researchers and male researchers. All citation-based indicators suffer from several weaknesses. At a theoretical level, citations may reflect scholars acknowledging prior work that has influenced them (Merton, 1973), but they are not used systematically for this, can be influenced by spurious factors and can be negative (MacRoberts & MacRoberts, 1989; MacRoberts & MacRoberts, 1996). Moreover, citations only reflect knowledge advancement rather than wider contributions to academia or society (Kostoff, 1998; Merkx, Weijden Besselaar, & Spaapen, 2007). For example, academic publications can be used in education, the professions (Schloegl & Stock, 2004), and to inform health (Bennett, Casebeer, Kristofco, & Strasser, 2004; Lewison, 2002). The people affected in these cases may be thought of as “pure readers” in the sense of consuming academic outputs without subsequently citing them (Stock, 2009; Tenopir & King, 2000). Governments and research funders may explicitly state the need to consider non-academic impacts in their evaluations (NSF, 2013; Piwowar, 2013; Research Councils UK, 2011, para.1). An important practical drawback of citation-based indicators is the considerable time that they take to accumulate. A research team might think of an idea, submit a grant proposal, get funded, carry out their research, submit a write-up to a journal, get it reviewed and accepted and then wait for the article to be published online or in a formal journal issue. This process might take several years and several more years would be needed before their article had attracted a reasonable number of citations from others who had read their work and then followed the same process to conduct related research. A citation window of three years is sometimes recommended for research evaluations (Abramo, Cicero, & D’Angelo, 2011). Thus, whilst academic evaluations often have the aim of predicting future research excellence so that it can be funded, or the promising academic can be appointed/promoted, in practice citation-based indicators reflect performance that occurred several years in the past. 2 Online access and download data In response to the time delay and limited impact coverage problems of citation analysis as well as the expense of peer review, science policy makers and managers may seek alternative indicators to reflect wider types of research impact or to give earlier evidence of impact (Drooge, Besselaar, Elsen, & Haas, 2010; De Jong, Van Arensbergen, Daemen, Van der Meulen, & Van den Besselaar, 2011; Priem, Taraborelli, Groth, & Neylon, 2010). Another important driver towards new indicators is the changing nature of scholarly communication, with increasingly diverse types of scholarly output being published online and valued even if they are rarely cited (Czerniewicz, Kell, Willmers, & King, 2014; Van der Vaart et al. 2013). In parallel, an increasing amount of scholarly communication takes place in public and online, leaving traces that may be used to evaluate its importance (Cronin, Snyder, Rosenbaum, Martinson, & Callahan, 1998). Given that an academic publication must be accessed and read to be valuable but that not all readers are citers, it is logical to look to access and readership data for evidence of the wider impacts of publications (Haustein, 2014). This information can also ameliorate the time delay problem of citation analyses because an article must be read before any action can be taken based on its contents. Hence, unless the impact of an article is predicted from its metadata (e.g., from Journal Impact Factors), evidence of downloads or readership gives the earliest impact evidence. Away from research evaluation, this argument has also been made for library collection development. At the level of entire journals, readership and citation may be thought of as two separate, overlapping dimensions of the impact of research and giving librarians or evaluators information about both can help them to make more informed choices (Haustein, 2011; Haustein, 2012). Although rarer in research evaluation contexts, the importance of readers is a natural concern for scholars, who may submit to journals partly based on the readership that they hope to gain (Tenopir & King, 2000), librarians who choose journals primarily to service their potential readers, and editors or publishers that monitor the overall audience or sales for a journal as an indicator of its value or health (Rousseau, 2002). Readership data reflects something different from citations, even when only considering academic journal articles and restricting attention to academic readers. This is because there are some types of articles that are often read but rarely cited, such as a series of annual summaries of astrophysics (Kurtz, Eichhorn, Accomazzi, Grant, Demleitner, Murray, et al., 2005). 2.1.1 Online access and download data for journal usage assessment Statistics about local online accesses of journals are the modern way for librarians to monitor readership levels, although they are not comprehensive. Libraries keep print copies of journals, researchers and faculty may subscribe to individual serials, and educators may photocopy articles for students. In addition, article preprints may be shared from institutional and subject repositories and home pages, as well as by email and post. These uses will not normally be recorded electronically by a library or publisher but online accesses nevertheless are likely to give more comprehensive and timely information than library circulation data. Online usage data may be able to separate out people that browse journals from those that read articles by accessing their full text, giving more substantial information than circulation data. Local data is more relevant to libraries than generic publisher data that covers all uses because each library serves a user community that has specific information needs – including research interests and educational provision. Moreover, local usage data seems to be more tied to reading since robots would presumably rarely access local library copies of articles, even though humans may still systematically download them (Emrani, Moradi-Salari, & Jamali, 2010). Online usage information originates from the log file of a web server recording accesses of journal pages or downloads of electronic copies of articles. There are many technical pitfalls with this information, including accesses by robots and repeated accesses by individuals for spamming purposes or by accident, and so online accesses do not equate with human readership. Since publishers gain extra sales if their access statistics are higher, they do not have an incentive to clean their data from spurious downloads before delivering it to libraries or others. There have been initiatives to standardize the process of cleaning the data to ensure that compliant publishers generate credible and comparable final statistics for their end users. The main initiative for this is COUNTER (Counting Online Usage of NeTworked Electronic Resources), which standardizes the reporting of usage information (COUNTER, 2017; Pesch, 2015). As discussed above, usage data is inaccurate because it is always an incomplete record of readership and there will also be unrecorded readers. In addition, there are other important limitations that apply to some or all contexts. For all applications, the lack of contextual information with most usage data (Mendeley and questionnaire data are exceptions) is an important restriction. Librarians may consider usage by faculty to be more valuable to the university mission than uses by students, at least on an individual level, but can rarely distinguish between the two in their data. In addition, no current major data source gives evidence about how a document was used by the reader (Kurtz & Bollen, 2010). This is an advantage of citations, because, in theory, the reason for a citation can be deduced from the accompanying text. 2.1.2 Online access and download data for research evaluation The shift to electronic publishing has led to the widespread availability of electronic access information at the level of individual articles, such as from publisher websites. This has made usage or readership data a practical source of evidence for research evaluations. In many cases usage and readership information can be used in a similar way to citations for impact assessment, although it has different advantages and limitations. It has also not become as generally accepted as citations for this purpose. For example, the Journal Impact Factor is much more widely reported and recognized than any of the proposed usage-based alternatives. For research evaluations rather than collection development purposes, statistics that are available for entire journals but not individual articles are unhelpful, although monograph circulation data can help research evaluation in the arts, humanities and some social sciences (White, Boell, Yu, Davis, Wilson, & Cole, 2009). Also for research evaluation purposes, the limited nature of local access data from individual libraries for journal articles can be resolved by forming a consortium of libraries to share data (Bollen & Van de Sompel, 2008) or by substituting publisher statistics. The former may be a realistic possibility for libraries that already have a common infrastructure to access electronic journals and so that data sharing can be added as an additional service rather than a completely new contribution. Whatever the source of usage data, its principal advantage over citations for research evaluations is timeliness because usage logically comes before the publication of citations. A second advantage is scope because usage data includes, but does not differentiate, readers that do not cite the work. Hence usage data may provide a timelier source of impact evidence with a wider scope. The reason why it is rarely preferred to citations is that it is much easier to manipulate and so it is not credible enough for formal research evaluation purposes, even if from a COUNTER-compliant source. Nevertheless, it can be valuable for informal evaluations, selfevaluations and assessments of the science system as well as to cross-check the results of peer review or citation analysis. Because of the accuracy limitations of usage data, it is important to assess whether it gives evidence of academic impact before it is used for article-level research evaluations. The primary strategy so far for this is to assess the extent to which article-level usage statistics correlate with citation counts. A perfect correlation cannot be expected because of the absence of pure readers from citation statistics, but a moderate or high correlation would suggest that the usage source assessed is not too affected by manipulation or fake data from robots. In contrast, a correlation close to zero would suggest that either there are many readers that have very different needs to citers or that the results have been intentionally or unintentionally manipulated. Correlation analyses have mostly found moderate or high correlations between downloads and citations, which tends to confirm the value of usage data. A small study (n=153) of article downloads in their first week found a moderate correlation (Pearson r=0.5) with WoS citations five years later (Perneger, 2004). Similar correlations have been found for downloads of articles in the physics preprint server arXiv in the first six months and their citations after two years (Harnad & Carr 2006), for RePEc economics preprint downloads and citations (Chu, Krichel, & Blvd, 2007), for PLoS PDF downloads and citations (Yan & Gerstein, 2011. In contrast, a correlation of only 0.1 was found between early downloads (2 months) and later citations (25 months) for the fast, organic chemistry journal Tetrahedron Letters, suggesting that for this journal, early accesses represent a different type of use to citation (Moed, 2005). For downloads within Elsevier’s ScienceDirect and Scopus citation counts, the two correlate in all disciplines at the level of journals and articles; early downloads also correlate with later citations. These correlations vary in strength by discipline, are lowest in the arts and humanities (0.2-0.3) and reach as high as 0.8 (life sciences). Despite this, the most downloaded articles tend to differ from the most cited articles for individual journals (Moed & Halevi, 2016). Confusingly, a study of Chinese journals found higher correlations between downloads and citations within the arts, humanities and social sciences than for other academic disciplines (Vaughan, Tang, & Yang, 2017). At the level of journals, various download-based indicators have been defined in similar ways to the JIF, including the Usage Impact Factor (Bollen & Sompel, 2008) and the Download Immediacy Index (Wan, Hua, Rousseau, & Sun, 2010). Correlation tests have been used to help assess the value and validity of download-based indicators, with typically weaker results than at the level of individual articles (e.g., Wulff & Nixon, 2004), and with some negative correlations. Usage data for Rouen University Hospital digital library had a low positive correlation with JIFs in one study (Darmoni & Roussel, 2002), and correlations were not significantly different from zero for JIFs and Concordia University chemistry and biochemistry journal usage data (Duy & Vaughan, 2006). A comparison of JIFs with aggregate full text download data for a set of universities found low negative correlations, suggesting that journals most used by students (the main downloaders) were the least cited (Bollen, Van de Sompel, & Rodriguez, 2008; see also: Bollen and Sompel, 2008). Thus, whilst download data seems to reflect a high degree of scholarly impact at the level of individual articles, when articles are aggregated into journals, scholarly impact is substantially less important and download data may predominantly reflect educational value. Electronic usage data can sometimes incorporate information about the origins of the users from the internet address of their computers. It is therefore possible to break down the readers of an article by institution and country and perhaps also organization type, if this data is made available by publishers or web server operators (Duin, King, & Van Den Besselaar, 2012; Kurtz, Eichhorn, Accomazzi, Grant, Demleitner, Murray, et al., 2005). This can reveal where articles and journals have had impact. This may be relevant to national funding bodies that want to demonstrate international impact or, conversely, want to make sure that the home nation rather than a competitor is benefiting from their investment. 2.1.3 Limitations of online usage data As discussed above, usage data can include false hits, whether robot accesses or downloads by people that did not intend to read the article, and articles can be accessed from multiple print and online sources (Haustein & Siebenlist, 2011). These limitations apply unequally between journals and even between articles so it is not fair to compare the impact of articles using any source of download data. For example, one article’s publisher download count may be half that of another because it is available free online from the author, or is available in print in addition to electronic versions (Anderson, Sack, Krauss, & O’Keefe, 2001). The main disadvantage of download counts from a research evaluation perspective is that they are easy to manipulate unless extensive measures are taken to protect them (Zimmermann, 2013). An additional disadvantage is that the data is not transparent because publishers do not share the identities of those that accessed an article and so authors and evaluators have no means of verifying downloads. 2.2 Readership data from online reference managers In addition to manual methods to collect readership information, such as surveys, reader observation and reshelving information, and computerized methods, such as library, publisher or repository download statistics, the web has made possible an additional indirect method to find whether an article has many readers: online reference managers. A reference manager is a program that records metadata about some or all the resources that a person is interested in, typically to keep track of what they have read and to automatically generate reference lists for their documents, whether they are journal articles, conference papers, books or reports. Reference managers like EndNote, RefWorks, CiteULike, Connotea, Mendeley, Bibsonomy and Zotero all perform this role in different ways. If it is assumed that reference manger users tend to record articles that they have read, then the collective databases of reference managers form a large source of information about what has been read by whom. Some reference managers do not share this information but others, such as Mendeley and Bibsonomy, do, and so reference manager records are an alternative source of readership information (Neylon & Wu, 2009). At first glance, reference manager data is an unpromising source of readership evidence. Not all readers use reference managers and so they form an incomplete readership record. No reference manager is dominant and so if one is used as a data source then its information will be partial even with respect to all reference manager users. Reference manager users are likely to be a biased subset of all readers because technophobes might avoid learning a new program and people that do not write documents that include references would have little need for them. Nevertheless, some reference managers have an advantage over download data: their information is freely available from a single source (rather than multiple publishers), they are not affected by multiple copies of articles being available (e.g., preprints in repositories) and they seem to give more definite evidence of readership than downloads because the latter could be from a crawler or a casual user. For this reason, they can be a more realistic data source for research evaluations than download data. Data from reference managers that are also social websites and provide an Applications Programming Interface (API), such as Mendeley, CiteULike and Bibsonomy, falls within the scope of altmetrics (Holmberg, 2015). These are indicators derived from data harvested from social web sites via APIs. The altmetrics movement has led to the creation of many new indicators. Indicator companies, such as Altmetric.com, ImpactStory.org and Plum Analytics, systematically collect altmetric data (including from reference managers) and make it available to publishers, scholars and institutions (Wouters & Costas, 2012). Altmetric.com, for example, attempts to provide accurate and transparent article-level indicators (Liu & Adie, 2013). Although it includes readership data from Mendeley, it treats this as a secondary data source since it is not transparent (i.e., does not reveal the identity of readers). In addition, there are public initiatives to harvest and share altmetric data, such as one from PLoS (Fenner, 2013). The promise of altmetrics is that it will deliver faster impact evidence that encapsulates wider types of impact (Priem, Piwowar, & Hemminger, 2012; Priem, Taraborelli, Groth, & Neylon, 2010). Within this, reference manager data fits as a fast and wider source of evidence since reference manager users may be students (e.g., Winslow, Skripsky, & Kelly, 2016) and other nonpublishing article readers. Each altmetric has its own strengths and weaknesses, and potentially reflects a different type of impact. For example, tweet citations (Eysenbach, 2011) seem to reflect attention rather than impact and are probably the fastest indicator to accrue. Reference manager data can therefore be thought of as an alternative to download counts as a source of readership evidence, or as a type of altmetric to be analysed in parallel with other altmetrics. 2.2.1 Online reference managers: background Online reference managers have broadly similar functions but each has its own software design, individual features and user demographics. The national, disciplinary and age composition of the adopters of each one is likely to be influenced by its age, national and disciplinary origins and the fit of its affordances within disciplinary missions. For example, most have Western origins and did not prioritise language support, which may have alienated potential users in China, Russia and Japan. User demographics are unknown for most, however. The descriptions below of some of the major social sites give a flavour of their differences but their capabilities evolve over time and so may have changed now. All are online social reference managers in the sense that they manage references, are online, and allow users to create a public profile.  Bibsonomy (www.bibsonomy.org) manages web bookmarks as well as references and incorporates social features (Benz, Hotho, Jäschke, Krause, Mitzlaff, Schmitz, & Stumme, 2010; Jäschke et al., 2007; Ortega, 2016). Users typically have about 20% more references than bookmarks (Ortega, 2016). Probably like the other sites, most references are for journal articles (Borrego & Fry, 2012). Bibsonomy provides a copy of all its data free for researchers (www.kde.cs.uni-kassel.de/bibsonomy/dumps/).  CiteULike (citeulike.org) is free, has the basic reference manager capabilities described above and allows users to annotate references and share them with others (Reher & Haustein, 2010). It also has discussion fora and blogs (Bogers & Van Den Bosch, 2008). Because of its communication and sharing capabilities it is also a type of academic social web site.  Mendeley (Mendeley.com) is a free reference manager (Henning & Reichelt, 2008) that has been bought by Elsevier and offers social networking features, such as the ability to follow other users, as well as discussion groups and the ability for users to list their own publications. It therefore serves as an academic social network site as well as a reference manager, although its social features do not seem to be extensively used (Jeng, He, & Jiang, 2015). Mendeley offers a free and comprehensive API to access its readership data for all articles in its catalogue so that anyone can find out how many Mendeley users have recorded any given document within their libraries. Although Mendeley does not report which users have registered each document, it gives a breakdown of user types by status (e.g., undergraduate, professor, other professional), geographic location (country) and main academic discipline (Gunn, 2013).  Zotero (www.zotero.org) is a free, open source reference manager that originated as a Firefox web browser plugin but is now available as a separate program. It has features to support group discussions and group reference sharing. In addition to the above, RefWorks is a reference manager owned by ProQuest, and EndNote is owned by Thomson Reuters. Neither share readership data publicly at the time of writing. 2.2.2 Online reference managers: Coverage Readership data from online social reference managers needs to be publicly available, or at least shared with researchers, and to have many records to be useful. If a site has few users, then these are likely to be a very biased subset of readers so the results may be misleading. For example, article readers tend to be from the same country as the authors (Thelwall & Maflahi, 2015) so any national readership biases will translate into international readership indicator biases. If most articles do not have a record of readers in the site, then its data is unlikely to be powerful enough for research evaluation purposes unless used on a large scale to compare average reader counts (e.g., using the EMNPC: Thelwall, 2017bd). Of the online reference managers sharing data, Mendeley has the widest coverage and probably the most users. It had records for 80% of PLoS articles compared to 31% for CiteULike (Priem, Piwowar and Hemminger, 2012) and indexed more Nature and Science articles (Li, Thelwall, & Giustini, 2012).  Bibsonomy: Bibsonomy has much lower coverage of physics journal articles 2004-2008 than CiteULike and probably less than 1% (Haustein & Siebenlist, 2011). Journal articles comprise half (47%) of the items recorded, with conference papers (25%) and books (12%) also being common (Borrego & Fry, 2012).  CiteULike: Most (65%) PloS Biology articles have a record in CiteULike (Fenner, 2013). Less than 3% of physics articles 2004-2008 are in CiteULike (Haustein & Siebenlist, 2011).  Mendeley: Virtually all (97%) articles from Journal of the American Society for Information Science and Technology 2001-2011 (Bar-Ilan, 2012) and PloS Biology (95%) have a record in Mendeley (Fenner, 2013). Most (66%) PubMed articles 2010-2012 that are also in the Web of Science have a Mendeley record (Haustein, Larivière, Thelwall, Amyot, & Peters, 2014). For Scopus medical fields, 78% of articles had at least one reader (Thelwall & Wilson, 2016). Another study found high coverage for WoS Medicine 2008 articles (72%) but lower (about a third) for Physics, Chemistry, Engineering and Technology (Mohammadi, Thelwall, Haustein, et al., 2015). Less than half of recent social sciences WoS articles (44%) are in Mendeley, varying from Psychology (54%) to Linguistics (34%) and only 13% of humanities articles were indexed, from Education (34%) to Literature (4%) (Mohammadi & Thelwall, 2014). Nevertheless, 61% of Swedish humanities journal articles from 2012 were in Mendeley (Hammarfelt, 2014). Compared to other altmetrics from Altmetric.com, Mendeley had the highest coverage (63%) of a large sample of Web of Science articles (Zahedi, Costas, & Wouters, 2014). Very few books have records: only 7% of recent WoS science and medicine volumes (Kousha & Thelwall, 2016). Thus, whilst Mendeley has wide coverage overall and particularly for medicine, it is weak in the humanities and very weak in some disciplines and for books. This may have changed since the studies described here however.  Zotero. No coverage information is available. 2.2.3 Online reference managers: Correlation with citation counts When a new indicator is proposed for an aspect of research evaluation then the logical first method to assess whether it has any value is to calculate its correlation with citation counts on the basis that a positive result would be evidence that the data was not random and related to scholarly impact in some way (Sud & Thelwall, 2014). Even though a negative or zero correlation is also consistent with a new indicator reflecting a completely different type of impact, in practice most types of impact relate to each other to some extent and so this test is reasonable. There is extensive evidence of this type for Mendeley and a little for CiteULike. For Mendeley, readership counts correlate positively and moderately strongly with citation counts (and peer review judgements) in most fields, with the arts being the main exception.  CiteULike records and citations have a significant positive correlation for Science and Nature (Li, Thelwall, & Giustini, 2012). Usage data dominated by CiteULike have low Spearman correlations (0.1) with Journal Impact Factors for physics journals (Haustein & Siebenlist, 2011)  Mendeley records and citations have a significant positive correlation for Science and Nature (Li, Thelwall, & Giustini, 2012), for PLoS ONE, PLoS Biology and PLoS Pathogens articles (Priem, Piwowar, & Hemminger, 2012) and for selected genetics and genomics articles (Li & Thelwall, 2012). Mendeley readers have a moderate overall correlation (0.5) with WoS article citations (Zahedi, Costas, & Wouters, 2014). For PubMed articles 20102012 in WoS, Spearman correlations between Mendeley readers WoS citations were positive and statistically significant in all broad disciplines except the Arts. They varied from 0.2 (Humanities) to 0.6 (Engineering and Technology), with an average of 0.5 (Haustein, Larivière, Thelwall, Amyot, & Peters, 2014). For WoS articles from 2008, five social science fields had Spearman correlations of 0.4-0.6 and five humanities fields had Spearman correlations of 0.4 or 0.5 (Mohammadi & Thelwall, 2014; see also: Thelwall, in press). Similar correlations were found for science and medicine fields (0.4 or 0.5) except for Engineering and Technology (0.3) (Mohammadi, Thelwall, Haustein, & Larivière, 2015). Within Medicine the correlations later (and for narrower fields) rose to 0.7 (Thelwall & Wilson, 2016). The most systematic analysis so far checked 325 narrow Scopus fields, finding strong positive correlations in almost all (Thelwall, 2017f). For books, correlations between Mendeley reader counts and citations are about 0.1 (Kousha & Thelwall, 2016). Engineering conference papers have a very low correlation with citation counts (Aduku, Thelwall, & Kousha, 2017). A more direct source of evidence of the value of readership counts is their correlation with peer review scores. Peer review judgements are impractical to obtain for large samples of articles unless the data is a by-product of a research evaluation. For articles published in 2008 and submitted for evaluation by subject experts in the UK’s Research Excellence Framework (REF) 2014, correlations between Mendeley reader counts and expert ratings in 33 of the 36 fields examined were positive, with the highest being for clinical medicine (0.4) and the lowest for Music, Drama, Dance and Performing Arts (-0.1) (HEFCE, 2015). Given that these articles were selected by their authors for being high quality, the correlations are likely to substantially underestimate the underlying level of agreement between peer judgement and Mendeley reader counts and so this is strong evidence that in most fields Mendeley reader counts reflect the quality of journal articles. A weaker corroborating piece of evidence is that UK clinical guideline references have more Mendeley readers than do comparable articles (Thelwall & Maflahi, 2016) 2.2.4 Online reference managers and reading References can be logged within a reference manager by users that have not read them (Weller & Peters, 2012) or as part of training exercises (MacMillan, 2012) and so it is not clear that it is reasonable to characterize reference manager data as “readership counts”. The best way to find out why users enter reference data is to ask them. A survey of Mendeley users found that most (85%) added articles to cite them, but many also added articles for professional (50%) or teaching (25%) purposes. Moreover, most added articles that they had read or intended to read. Thus, Mendeley readership data clearly represents readership and a wider type of impact than scholarly impact, although mainly still within a broad academic context (Mohammadi, Thelwall, & Kousha, 2015). Some articles are added for educational reasons, having many readers but few citations (Thelwall, 2017c). Since undergraduates use reference managers, it is logical to look to readership data for evidence of educational reading. This is undermined by evidence that undergraduates and researchers tend to register similar articles (Thelwall, 2017a) 2.2.5 Online reference managers: Reader types and demographics Readers of research can originate from any country in the world, from any discipline, from any academic status. They can also be professionals using the information for their work or could be members of the public with an interest in a specific topic or fact. Within these groups, some read more academic research than others, and even when considering academic researchers alone, younger researchers read and cite more (Larivière, Sugimoto, & Bergeron, 2013; Tenopir, King, Spencer, & Wu, 2009). Undergraduates sometimes read scientific papers but their reading is often directed by their lecturers (Korobili, Malliari, & Zapounidou, 2011; Whitmire, 2002). To interpret the type of impact reflected by readership data, it is therefore important to investigate the nature of people that use online reference managers. Partial information is known for Bibsonomy and Mendeley. In terms of geography, almost half of all Bibsonomy users are from Germany (Ortega, 2016), undermining its value for general impact assessment. Probably all the major services have relatively low numbers of users from China and from countries with little scientific publishing or a low level of technology use in universities. In terms of work status, Mendeley includes substantial numbers of undergraduates and Master’s students and few non-academic users. In science, it is dominated by young users: PhD students, postgraduates and postdoctoral researchers (Mohammadi, Thelwall, Haustein, & Larivière, 2015). In contrast, successful senior researchers seem to avoid it (Mas-Bleda, Thelwall, Kousha, & Aguillo, 2014) and so there is an age/seniority bias. 2.2.6 Online reference managers: Timeliness Mendeley readers appear about a year before citations, on average. For four library and information science (LIS) journals, the number of citations reaches the number of readers after about seven years (Maflahi & Thelwall, 2016). A similar pattern of initially higher readership counts than citation counts has been found for 50 fields, although the number of years needed for citations to overtake readers varies by discipline (Thelwall & Sud, 2016). Early Mendeley readership counts are also better predictors of later high citation counts than are journal impact factors or citations (Zahedi, Costas & Wouters, 2015). All this evidence supports the conclusion that Mendeley readership counts give statistically stronger impact evidence than citation counts in the first few years after publication. It is common for articles to have Mendeley readers as soon as they are formally published because of the prior sharing of preprints (Maflahi & Thelwall, 2018). This makes it possible to conduct evaluations of them immediately upon publication, if these evaluations do not require the statistical power of high average readership counts. Most importantly, the higher number of Mendeley readers than citations in the year following publication makes Mendeley reader counts correlate more strongly than citation counts with peer review judgements of the quality of journal articles (HEFCE, 2015). 2.2.7 Online reference managers: Research evaluation applications Readership data is useful for research evaluation applications where timeliness is important and there is little risk of deliberate manipulation of the data. This excludes formal exercises where those evaluated are told the data sources in advance but allows their use for more routine academic progress monitoring. Mendeley readership counts are useful for national-level evaluations for governments to compare their progress against that of their competitors. The goal of such evaluations is to inform policy decisions or to assess the effect of recent policy changes. For the latter case in particular, timely data is essential. Mendeley readership data is preferable to citations because of its timeliness but has the limitation that it is influenced by different levels of national uptake from its users. This is a problem because of the tendency for people to read articles from their own country. It is possible to circumvent this issue with a modelling approach to measure the amount of bias in caused by the readership demographics and then correct for them, but this strategy is imperfect because it requires assumptions or information about the evolution of national uptake of the site over time (Fairclough & Thelwall, 2015). Funding councils are also logical users of readership data. These may monitor the average impact of the research that they fund to identify which funding streams are most successful and whether the average impact of their funded research has changed over time. Web indicators can be useful for these because the time lag of citations would delay decisions about changing ineffective funding strategies (Thelwall, Kousha, Dinsmore, & Dolby, 2016) as well as for evidence of societal research impact (Dinsmore, Allen, & Dolby, 2014). National biases in uptake and the bias towards younger users are relatively minor problems for this and so Mendeley readership data is a better source than citations (e.g., Thelwall, 2017b), although it does not reflect societal benefits or many professional uses. One recent application harnesses readership data purely for its early impact evidence in comparison to citation counts, emphasizing the importance of the publication lag for citations. It used reader counts for early evidence of the success of an article promotion strategy in a randomised controlled trial (Kudlow, Cockerill, Toccalino, Dziadyk, Rutledge, et al., 2017). 2.2.8 Illustration of Mendeley data Three documents were compared with Mendeley data to illustrate some of the features of Mendeley and some of the factors that should be considered when interpreting its data. Three important scientometric papers were selected for this comparison. The first is an old Nature article discussing citation analysis from the perspective of non-scientometricians concerned about its uses. The second is the well-known Leiden Manifesto. The third is an article about altmetrics (Table 1). All three articles have high numbers of readers and Google Scholar citations. Other factors being equal, older articles should be more cited so it seems likely that the second and third articles, from 2015, will eventually be more cited than the first one, from 2002. The two newer articles already have more Mendeley readers than the first article (Counting house). This is partly because Mendeley identified readers before citations and so newer articles take less time to catch up with older articles in terms of reader counts. It is also partly because the Counting house article was published years before Mendeley was released so its peak years of use would have preceded the existence of substantial numbers of Mendeley users. Table 1. Mendeley readers and Google Scholar citations for three scientometrics articles. Title and authors Citation analysis: The counting house by Adam, D. The Leiden Manifesto for research metrics by Hicks, D., Wouters, P., Waltman, L., De Rijcke, S., Rafols, I. Do altmetrics correlate with citations: Extensive comparison of altmetric indicators with citations from a multidisciplinary perspective by Costas, R., Zahedi, Z., Wouters, P. Year Readers GS cites Reads/Cites 2002 173 576 0.30 2015 634 474 1.34 2015 389 251 1.55 Mendeley data includes users’ professions (Figure 1). Most strikingly, the altmetric article is most used by librarians. Presumably this is due to the proliferation of altmetrics in publisher websites. In contrast, professors seem to be more concerned with traditional citation-based indicators. Figure 1. All Mendeley reader occupations for three scientometric articles. Mendeley data includes users’ declared country of origin or work (Figure 2). It seems that some countries that are taking citation analysis seriously, such as Brazil, are not concerned with altmetrics. In contrast, Canada and Netherlands seem to be more interested in altmetrics than citation analysis although both countries have active researchers working in both areas. The Counting house article seems to be particularly influential in the USA but it is not clear why. Figure 2. Countries of Mendeley readers for countries with at least 10 readers. There are substantial disciplinary differences in the uptake of the articles (Figure 3). The altmetrics article has attracted the most attention in the Social Sciences and Computer Science, although both categories might be due to library and information science researchers since this field falls within both. The citation analysis articles are of interest in the Agricultural and Biological Sciences. This is unsurprising given the origins of the San Francisco Declaration on Research Assessment (DORA) within the life sciences (the American Society for Cell Biology), indicating an unease with misuses of citation analysis within this discipline. Figure 3 also confirms that all three articles have attracted substantial interest outside of their home disciplines. Figure 3. Disciplines of Mendeley readers for the top 10 disciplines. 2.2.9 Investigating science with online reference manager data An interesting application of readership data is to track the flow of knowledge between fields. There is a long tradition of using citations to track knowledge flows by interpreting a citation from an article in field A to a paper in field B as knowledge flowing from B to A (Small, 1997). The same is possible for readership data when the domains of the readers of an article are known. The advantages of using readership data for this are timeliness and its ability to capture slightly wider impact because of the inclusion of students. A lot of data is needed to give good results however, which was a problem with one CiteULike study (Jiang, He, & Ni, 2011). An investigation comparing knowledge flows based on Mendeley readership data with citation-based knowledge flows found differences suggesting that researchers in some fields, including Business, read widely but cited narrowly (Mohammadi & Thelwall, 2014). A related application is the discovery of research clusters by identifying groups of articles read by the same user and then clustering them based on co-readership information (Kraker, Körner, Jack, & Granitzer, 2012), although this seems to be no longer possible with Mendeley. Readership data has also been used to investigate academics through their reference lists, when these are public (Borrego & Fry, 2012), to evaluate journals through the extent to which they are read (Haustein & Siebenlist, 2011), and to support literature search systems (Heck, Peters, & Stock, 2011). 2.2.10 Advantages and disadvantages of reference manager data compared to citation counts The timeliness, wider impact, and reader demographic information advantages of readership data from all sources have already been mentioned, as have the disadvantages that it is sometimes not transparent and always open to manipulation, with a biased user base. Some additional factors are important to consider when evaluating readership data. The advantages are listed first.  Traditional citation indexes, such as WoS and Scopus, have national biases and limits in coverage (Archambault, Campbell, Gingras, & Larivière, 2009) whereas there are no restrictions on the articles that may be added to reference managers.  Readership data is often free whereas citation indexes, except Google Scholar and Microsoft Academic, tend to charge for access.  Readership data is relatively easy to access on a large scale from sites with an API. For example, the free Webometric Analyst software can download Mendeley records via its API from the articles DOIs and/or metadata.  Readership data tends to be more numerous than citation counts (e.g., Thelwall, 2017e), except for older articles, and tests using it can therefore be statistically more powerful. There are also additional disadvantages with readership data.  Whereas, in theory, it is possible to find out how a work has been cited by reading the text accompanying the citation, references are rarely annotated with information that reveals why a publication was selected. Reviews are annotated readings and these are available from sites like Goodreads and Amazon for books (Kousha & Thelwall, 2016; Kousha, Thelwall, & Abdoli, 2017).  Despite the recognised national biases in citation indexes, Mendeley readership data seems to be more nationally biased than citation counts (Mas-Bleda & Thelwall, 2016).  Some altmetric sources of readership data can give inconsistent results (Fenner, 2014) and there is a need for standardisation between data providers and sources (Konkiel, 2013).  Younger readers are more represented in Mendeley (Mohammadi, Thelwall, Haustein, & Larivière, 2015) and the share of younger readers may vary by narrow field and publication year.  Differences in adoption levels and behaviours across disciplines is a complicating factor when interpreting the results of any multidisciplinary analysis (Wouters & Costas, 2012).  Some publication information entered by users to record their references is incomplete, leading to missed data (Bornmann & Haunschild, 2015). This may be more frequent for documents with mathematical titles or in languages that are not represented by the ASCII character set. 2.3 Usage data from academic social network sites The online environment for science communication is continually evolving and usage data is now not only available from publishers, academic repositories and reference managers but also from some academic social network sites. Both Academia.edu and ResearchGate allow members to log their own papers and add them to their profile pages (as Mendeley also does now). They also provide usage data on these records in the form of download or view counts. They differ from reference managers by focusing on each author’s own publications rather than their references of (presumably) mainly other scholars’ works. Thus, their usage data has essentially the same nature as the download and access statistics of publishers or repositories, even though their appearance is more like Mendeley. Academic social network sites are in competition with publishers as sources of published academic research and, because of this, undermine the comprehensiveness of publisher data, apparently irrespective of copyright concerns (Jamali, 2017). For research evaluation purposes, academic social network sites are not good sources of usage indicators because they have an incomplete collection of articles and do not make their usage data easily available for researchers. Nevertheless, they are important because they have many mamebers and their scores are apparently taken seriously by many researchers (Jordan, 2015, 2017; Orduna-Malea, Martín-Martín, Thelwall, & López-Cózar, 2017). ResearchGate article views correlate positively with Scopus citations and seem to reflect a wider set of users than publishing academics, putting them on a par with other sources of usage data (Thelwall & Kousha, 2017). ResearchGate also provides citation counts for uploaded articles by extracting citations from all articles uploaded to the site. Although it indexes less citations than Google Scholar, it finds more early citations than WoS and Scopus, suggesting that many authors upload preprints to the site (Thelwall & Kousha, 2017b). There are differing national levels of uptake of ResearchGate, which will bias its data, but despite being a type of web social network site, its data does not seem to favour younger users (Thelwall & Kousha, 2015b). There is less research about Academia.edu, but, like ResearchGate, its scores seem to favour senior academics. They also tend to favour women, perhaps due to their greater communication expertise in the social web (Thelwall & Kousha, 2014). 2.4 Summary This chapter has summarized research into readership data, including usage data, with a focus on research evaluation applications but also covering collection development applications. In theory, this data is preferable to citation counts because it captures more uses of scholarly documents, such as from students and professionals. Although there is a little evidence to support his conjecture, readership data seems to primarily reflect scholarly uses in most fields. Both readership data from reference managers and usage (download/view) data from publishers have the advantage of giving early impact evidence compared to citations because of the delays associated with the publication cycle. This is due to articles being read a year or more before the citations generated by the reading, if any, appear in a citation index. Nevertheless, both download data and reference manager data can be manipulated and whilst they are useful for informal evaluations and investigations into science itself, they should not be used for formal evaluation when those assessed can influence the data. Readership data from reference managers has the additional promise that it can reveal something about the demographics of the readers, including their discipline, nation and job type. This can help with investigations of science communication. Download data in many cases has the practical limitation that a set of articles may originate from many different publishers, which complicates accessing it and the data may not be fully comparable. In contrast, it is reasonable to collect reference manager readership data from a single site, such as Mendeley via its API, making it a practical source of readership information. 2.5 References Abramo, G., Cicero, T., & D’Angelo, C. A. (2011). Assessing the varying level of impact measurement accuracy as a function of the citation window length. Journal of Informetrics, 5(4), 659–667. Aduku, K. J., Thelwall, M., & Kousha, K. (2017). Do Mendeley reader counts reflect the scholarly impact of conference papers? An investigation of computer science and engineering. Scientometrics, 112(1), 573-581. Advisory Board for the Research Councils (ABRC). (1990). Peer Review. A Report to the Advisory Board for the Research Councils from the Working Group on Peer Review. London. Anderson, K., Sack, J., Krauss, L., & O’Keefe, L. (2001). Publishing Online-Only Peer-Reviewed Biomedical Literature: Three Years of Citation, Author Perception, and Usage Experience. The Journal of Electronic Publishing, 6(3). Archambault, É., Campbell, D., Gingras, Y., & Larivière, V. (2009). Comparing bibliometric statistics obtained from the Web of Science and Scopus. Journal of the Association for Information Science and Technology, 60(7), 1320-1326. Bar-Ilan, J. (2012). JASIST 2001–2010. Bulletin of Association for Information Science and Technology, 38(6), 24–28. Bennett, N. L., Casebeer, L. L., Kristofco, R. E., & Strasser, S. M. (2004). Physicians’ Internet information-seeking behaviors. The Journal of Continuing Education in the Health Professions, 24(1), 31–8. Benz, D., Hotho, A., Jäschke, R., Krause, B., Mitzlaff, F., Schmitz, C., & Stumme, G. (2010). The social bookmark and publication management system Bibsonomy. The VLDB Journal—The International Journal on Very Large Data Bases, 19(6), 849-875. Boell, S. K., & Wilson, C. S. (2010). Journal Impact Factors for evaluating scientific performance: use of h-like indicators. Scientometrics, 82(3), 613–626. Bogers, T., & Van Den Bosch, A. (2008). Recommending scientific articles using citeulike. In Proceedings of the 2008 ACM conference on Recommender systems RecSys 08 (pp. 287– 290). ACM Press. Bollen, J., & Sompel, H. Van de. (2008). Usage impact factor: The effects of sample characteristics on usage-based impact metrics. Journal of the American Society for Information Science and Technology, 59(1), 136–149. Bollen, J., Van de Sompel, H., & Rodriguez, M. A. (2008). Towards usage-based impact metrics. In Proceedings of the 8th ACM/IEEE-CS joint conference on Digital libraries (pp. 231–240). New York, New York, USA: ACM Press. Bornmann, L., & Haunschild, R. (2015). Which people use which scientific papers? An evaluation of data from F1000 and Mendeley. Journal of Informetrics, 9(3), 477–487. Borrego, Á., & Fry, J. (2012). Measuring researchers’ use of scholarly information through social bookmarking data: A case study of BibSonomy. Journal of Information Science, 38(3), 297– 308. Brody, T., Harnad, S., & Carr, L. (2006). Earlier Web Usage Statistics as Predictors of Later Citation Impact. Journal of the American Society for Information Science and Technology, 57(8), 1060–1072. Butkovich, N. J. (1996). Use Studies: A Selective Review. Library Resources & Technical Services, 40(4), 359–368. Chu, H., Krichel, T., & Blvd, N. (2007). Downloads vs. Citations: Relationships, Contributing Factors and Beyond. In H.F. Moed (Eds.) (Ed.), Proceedings of the 11th International Society for Scientometrics and Informetrics Conference (pp. 207–215). Madrid: CINDOC. COUNTER. (2017). About COUNTER. Retrieved January 1, 2017, from https://www.projectcounter.org/about Cronin, B., Snyder, H. W., Rosenbaum, H., Martinson, A., & Callahan, E. (1998). Invoked on the Web. Journal of the American Society for Infor- Mation Science, 49(14), 1319–1328. Czerniewicz, L., Kell, C., Willmers, M., & King, T. (2014). Changing Research Communication Practices and Open Scholarship: A Framework for Analysis. Scholarly Communication in Africa Programme. Retrieved from https://open.uct.ac.za/handle/11427/9068 Darmoni, S., & Roussel, F. (2002). Reading factor: a new bibliometric criterion for managing digital libraries. Journal of the Medical Library, 90(3), 323–327. Davis, P. M. (2004). Informationseeking behavior of chemists: A transaction log analysis of referral URLs. Journal of the American Society for Information Science and Technology, 55(4), 326–332. de Jong, S. P. L., van Arensbergen, P., Daemen, F., van der Meulen, B., & van den Besselaar, P. (2011). Evaluation of research in context: an approach and two cases. Research Evaluation, 20(1), 61–72. Dellavalle, R. P., Schilling, L. M., Rodriguez, M. A., Van de Sompel, H., & Bollen, J. (2007). Refining dermatology journal impact factors using PageRank. Journal of the American Academy of Dermatology, 57(1), 116–9. Dinsmore, A., Allen, L., & Dolby, K. (2014). Alternative perspectives on impact: The potential of ALMs and altmetrics to inform funders about research impact. PLoS Biology, 12(11), e1002003. Drooge, L., Besselaar, P., Elsen, G., & Haas, M. (2010). Evaluating the societal relevance of academic research: A guide. The Hague: Rathenau Instituut. Retrieved from http://depot.knaw.nl/9026/ Duin, D., King, D., & Van Den Besselaar, P. (2012). Identifying audiences of e-infrastructures-tools for measuring impact. PloS one, 7(12), e50943. Duy, J., & Vaughan, L. (2006). Can electronic journal usage data replace citation data as a measure of journal use? An empirical examination. The Journal of Academic Librarianship, 32(5), 512– 517. Egghe, L. (1988). Mathematical relations between impact factors and average number of citations. Information Processing & Management, 24(5), 567–576. Egghe, L., & Rousseau, R. (2000). The influence of publication delays on the observed aging distribution of scientific literature. Journal of the American Society for Information Science, 51(2), 158–165. Emrani, E., Moradi-Salari, A., & Jamali, H. R. (2010). Usage Data, E-journal Selection, and Negotiations: An Iranian Consortium Experience. Serials Review, 36(2), 86–92. Eysenbach, G. (2011). Can Tweets Predict Citations? Metrics of Social Impact Based on Twitter and Correlation with Traditional Metrics of Scientific Impact. Journal of Medical Internet Research, 13(4), e123. Fairclough, R., & Thelwall, M. (2015). National research impact indicators from Mendeley readers. Journal of Informetrics, 9(4), 845–859. Fenner, M. (2013). What Can Article-Level Metrics Do for You? PLoS Biology, 11(10), e1001687. Retrieved from http://dx.plos.org/10.1371/journal.pbio.1001687 Fenner, M. (2014). Altmetrics and Other Novel Measures for Scientific Impact. In S. Bartling & S. Friesike (Eds.), Opening Science (pp. 179–189). Springer International Publishing. Gross, P. L. K., & Gross, E. M. (1927). College Libraries and Chemical Education. Science. Grabowsky, A. (2016). Expanding access: An evaluation of ReadCube access as an ILL alternative. Medical Reference Services Quarterly, 35(1), 16-26. Gunn, W. (2013). Social signals reflect academic impact: What it means when a scholar adds a paper to Mendeley. Information standards quarterly, 25(2), 33-39. Hammarfelt, B. (2014). Using altmetrics for assessing research impact in the humanities. Scientometrics, 101(2), 1419–1430. Haque, A., & Ginsparg, P. (2009). Positional effects on citation and readership in arXiv. Journal of the American Society for Information Science and Technology, 60(11), 2203–2218. Haustein, S., Larivière, V., Thelwall, M., Amyot, D., & Peters, I. (2014). Tweets vs. Mendeley readers: How do these two social media metrics differ? IT – Information Technology, 56(5), 207–215. Haustein, S., & Siebenlist, T. (2011). Applying social bookmarking data to evaluate journal usage. Journal of Informetrics, 5(3), 446–457. Haustein, S. (2011). Taking a Multidimensional Approach Toward Journal Evaluation. In Proceedings of the ISSI Conference (Vol. 2, pp. 280–291). Durban, South Africa. Haustein, S. (2012). Multidimensional Journal Evaluation Analyzing Scientific Periodicals beyond the Impact Factor. Berlin / Boston: De Gruyter Saur. Haustein, S. (2014). Readership Metrics. In B. Cronin & C. Sugimoto (Eds.), Beyond Bibliometrics: Harnessing Multidimensional Indicators of Scholarly Impact (pp. 327–344). Cambridge, MA: MIT Press. Heck, T., Peters, I., & Stock, W. G. (2011). Testing collaborative filtering against co-citation analysis and bibliographic coupling for academic author recommendation. In B. Mobasher & J. Burke (Eds.), Proceedings of the 3rd ACM Workshop on Recommender Systems and the Social Web (pp. 16–23). NewYork: ACM Press. Hemphill. HEFCE. (2011). Research Excellence Framework (Vol. 2011). London, England: Higher Education Funding Council for England. HEFCE. (2015). The Metric Tide Supplementary Report II: Correlation analysis of REF2014 scores and metrics. Henning, V., & Reichelt, J. (2008). Mendeley-A Last. fm for research? In: IEEE Fourth International Conference on eScience (eScience'08) (pp. 327-328). Los Alamitos, CA: IEEE Press. Hirsch, J. E. (2005). An index to quantify an individual’s scientific research output. Proceedings of the National Academy of Sciences of the United States of America, 102(46), 16569. Holmberg, K. J. (2015). Altmetrics for information professionals: Past, present and future. Oxford, UK: Chandos Publishing. Jamali, H. R. (2017). Copyright compliance and infringement in ResearchGate full-text journal articles. Scientometrics.doi:10.1007/s11192-017-2291-4 Jäschke, R., Hotho, A., Schmitz, C., Grahl, M., Krause, B., & Stumme, G. (2007). Organizing publications and bookmarks in BibSonomy. CEUR Workshop Proceedings, 273, 2–5. Jayasinghe, U. W., Marsh, H. W., & Bond, N. (2003). A multilevel cross-classified modelling approach to peer review of grant proposals: the effects of assessor and researcher attributes on assessor ratings. Journal of the Royal Statistical Society: Series A (Statistics in Society), 166(3), 279–300. Jeng, W., He, D., & Jiang, J. (2015). User participation in an academic social networking service: A survey of open group users on Mendeley. Journal of the Association for Information Science and Technology, 66(5), 890-904. Jiang, J., He, D., & Ni, C. (2011). Social reference: Aggregating Online Usage of Scientific Literature in CiteULike for Clustering Academic Resources. In Proceeding of the 11th annual international ACM/IEEE joint conference on Digital libraries - JCDL ’11 (p. 401). New York, New York, USA: ACM Press. Jordan, K. (2015). Exploring the ResearchGate score as an academic metric: reflections and implications for practice. In Quantifying and Analysing Scholarly Communication on the Web (ASCW’15). Oxford, UK. Jordan, K. (2017). Understanding the structure and role of academics’ ego-networks on social networking sites. The Open University. Retrieved from http://oro.open.ac.uk/48259/ King, D., Tenopir, C., & Clarke, M. (2006). Measuring total reading of journal articles. D-Lib Magazine, 12(10). Konkiel, S. (2013). Altmetrics: A 21st-Century Solution to Determining Research Quality. Online Searcher, (July/August). Korobili, S., Malliari, A., & Zapounidou, S. (2011). Factors that Influence Information-Seeking Behavior: The Case of Greek Graduate Students. The Journal of Academic Librarianship, 37(2), 155–165. Kostoff, R. N. (1997). The principles and practices of peer review. Science and Engineering Ethics, 3(1), 19–34. Kostoff, R. N. (1998). The use and misuse of citation analysis in research evaluation. Scientometrics, 43(1), 27–43. Kousha, K., & Thelwall, M. (2016). Can Amazon.com reviews help to assess the wider impacts of books? Journal of the Association for Information Science and Technology, 67(3), 566–581. Kousha, K., Thelwall, M., & Abdoli, M. (2017). Goodreads reviews to assess the wider impacts of books. Journal of the Association for Information Science and Technology, 68(8), 2004-2016. Kraker, P., Körner, C., Jack, K., & Granitzer, M. (2012). Harnessing User Library Statistics for Research Evaluation and Knowledge Domain Visualization. In Proceedings of the 21st international conference companion on World Wide Web WWW 12 Companion (p. 1017). ACM. Kudlow, P., Cockerill, M., Toccalino, D., Dziadyk, D. B., Rutledge, A., Shachak, A., & Eysenbach, G. (2017). Online distribution channel increases article usage on Mendeley: a randomized controlled trial. Scientometrics, 112(3), 1537-1556. Kurtz, M. J., & Bollen, J. (2010). Usage bibliometrics. Annual Review of Information Science and Technology, 44(1), 1–64. Kurtz, M. J., Eichhorn, G., Accomazzi, A., Grant, C., Demleitner, M., Murray, S. S., … Elwell, B. (2005). The bibliometric properties of article readership information. Journal of the American Society for Information Science and Technology, 56(2), 111–128. Larivière, V., Sugimoto, C. R., & Bergeron, P. (2013). In their own image? a comparison of doctoral students’ and faculty members’ referencing behavior. Journal of the American Society for Information Science and Technology, 64(5), 1045–1054. Lee, C. J., Sugimoto, C. R., Zhang, G., & Cronin, B. (2013). Bias in peer review. Journal of the American Society for Information Science and Technology, 64(1), 2–17. Lewison, G. (2002). From biomedical research to health improvement. Scientometrics, 54(2), 179–192. Li, X., Thelwall, M., & Giustini, D. (2012). Validating online reference managers for scholarly impact measurement. Scientometrics, 91(2), 461–471. Li, & Thelwall, M. (2012). F1000, Mendeley and Traditional Bibliometric Indicators. In 17th International Conference on Science and Technology Indicators (Vol. 2, pp. 451–551). Liu, J., & Adie, E. (2013). Five challenges in altmetrics: A toolmaker’s perspective. Bulletin of the American Society for Information Science and Technology, 31–34. MacMillan, D. (2012). Mendeley: Teaching scholarly communication and collaboration through social networking. Library Management, 33(8/9), 561-569. MacRoberts, M. H., & MacRoberts, B. R. (1989). Problems of citation analysis: A critical review. Journal of the American Society for Information Science, 40(5), 342–349. MacRoberts, M. H., & MacRoberts, B. R. (1996). Problems of citation analysis. Scientometrics, 36(3), 435–444. Maflahi, N., & Thelwall, M. (2018). How quickly do publications get read? The evolution of Mendeley reader counts for new articles. Journal of the Association for Information Science and Technology, 69(1), 158–167. Maflahi, N., & Thelwall, M. (2016). When are readership counts as useful as citation counts? Scopus versus Mendeley for LIS journals. Journal of the Association for Information Science and Technology, 67(1), 191–199. Mas-Bleda, A., & Thelwall, M. (2016). Can alternative indicators overcome language biases in citation counts? A comparison of Spanish and UK research. Scientometrics, 109(3), 2007– 2030. Mas-Bleda, A., Thelwall, M., Kousha, K., & Aguillo, I. F. (2014). Do highly cited researchers successfully use the social web? Scientometrics, 101(1), 337–356. McAlister, F. A., Graham, I., Karr, G. W., & Laupacis, A. (1999). Evidence-based medicine and the practicing clinician. Journal of General Internal Medicine, 14(4), 236–242. Merkx, F., Weijden, I. van der, Oostveen, A.-M., Besselaar, P. van den, & Spaapen, J. (2007). Evaluation of research in context: A quick scan of an emerging field (second edition). Rathenau Institute, Department of Science System Assessment (Vol. www.rathen). The Hague: Merton, R. (1973). The Sociology of Science: Theoretical and Empirical Investigations. Chicago, IL: University of Chicago Press, Chicago, IL. Moed, H. F., & Halevi, G. (2016). On full text download and citation distributions in scientific‐ scholarly journals. Journal of the Association for Information Science and Technology, 67(2), 412-431. Moed, H. F. (2005). Statistical relationships between downloads and citations at the level of individual documents within a single journal. Journal of the American Society for Information Science and Technology, 56(10), 1088–1097. Mohammadi, E., & Thelwall, M. (2014). Mendeley readership altmetrics for the social sciences and humanities: Research evaluation and knowledge flows. Journal of the Association for Information Science and Technology, 65(8), 1627–1638. Mohammadi, E., Thelwall, M., Haustein, S., & Larivière, V. (2015). Who reads research articles? An altmetrics analysis of Mendeley user categories. Journal of the Association for Information Science and Technology, 66(9), 1832–1846. Mohammadi, E., Thelwall, M., & Kousha, K. (2016). Can Mendeley bookmarks reflect readership? A survey of user motivations. Journal of the Association for Information Science and Technology, 67(5), 1198–1209. Neylon, C., & Wu, S. (2009). Article-Level Metrics and the Evolution of Scientific Impact. PLoS Biology, 7(11), 6. NSF. (2013). Proposal and award policies and procedures guide; Changes to the Grant Proposal Guide (GPG) Effective January 14, 2013. Virginia. Retrieved from http://www.nsf.gov/pubs/policydocs/pappguide/nsf13001/gpg_index.jsp Opthof, T. (1997). Sense and nonsense about the impact factor. Cardiovascular Research, 33(1), 1–7. Orduna-Malea, E., Martín-Martín, A., Thelwall, M., & Delgado López-Cózar, E. (2017). Do ResearchGate Scores create ghost academic reputations? Scientometrics. doi:10.1007/s11192-017-2396-9 Ortega, J. L. (2016). Social Network Sites for Scientists: A Quantitative Survey. Cambridge, MA: Chandos Publishing. Perneger, T. V. (2004). Relation between online “hit counts” and subsequent citations: prospective study of research papers in the BMJ. British Medical Journal, 329(7465), 546– 547. Pesch, O. (2015). Implementing SUSHI and COUNTER: A primer for librarians. The Serials Librarian, 69(2), 107-125. Piwowar, H. (2013). Altmetrics: Value all research products. Nature, 493(7431), 159–159. Priem, J., Piwowar, H. A., & Hemminger, B. M. (2012). Altmetrics in the wild: Using social media to explore scholarly impact. Arxiv Preprint arXiv:1203.4745. Retrieved from http://arxiv.org/html/1203.4745v1 Priem, J., Taraborelli, D., Groth, P., & Neylon, C. (2010). Alt-metrics: a manifesto. Retrieved from http://altmetrics.org/manifesto RCUK. (2011). Typology of Research Impacts. Retrieved from http://www.rcuk.ac.uk/documents/impacts/TypologyofResearchImpacts-pdf Reher, S., & Haustein, S. (2010). Social bookmarking in STM: Putting services to the acid test. Online Leading Magazine for Information Professionals, 34(6), 34–42. Rousseau, R. (2000). Exponential decline in the use distribution of medical journals. Journal of Documentation, 56(4), 454–5. Rowlands, I., & Nicholas, D. (2007). The missing link: journal usage metrics. Aslib Proceedings, 59(3), 222–228. Schilling, L. M., Steiner, J. F., Lundahl, K., & Anderson, R. J. (2005). Residents’ patient-specific clinical questions: opportunities for evidence-based learning. Academic Medicine: Journal of the Association of American Medical Colleges, 80(1), 51–6. Schloegl, C., & Stock, W. G. (2004). Impact and relevance of LIS journals: A scientometric analysis of international and German-language LIS journals - Citation analysis versus reader survey. Journal of the American Society for Information Science and Technology, 55(13), 1155–1168. Small, H. (1997). Update on science mapping: Creating large document spaces. Scientometrics, 38(2), 275–293. Smith, R. (2006). Peer review: a flawed process at the heart of science and journals. Journal of the Royal Society of Medicine, 99(4), 178–82. Stankus, T., & Rice, B. (1982). Handle with Care. Collection Management, 4(1–2), 95–110. Stock, W. G. (2009). The inflation of impact factors of scientific journals. ChemPhysChem, 10(13), 2193–2196. Sud, P., & Thelwall, M. (2014). Evaluating altmetrics. Scientometrics, 98(2), 1131–1143. Tenopir, C., King, D. W., Clarke, M. T., Na, K., & Zhou, X. (2007). Journal reading patterns and preferences of pediatricians. Journal of the Medical Library Association : JMLA, 95(1), 56– 63. Tenopir, C., King, D. W., Spencer, J., & Wu, L. (2009). Variations in article seeking and reading patterns of academics: What makes a difference? Library & Information Science Research, 31(3), 139–148. Tenopir, C., & King, D. W. (2000). Towards Electronic Journals: Realities for Scientists, Librarians, and Publishers. Washington, D. C.: Special Libraries Association. Thelwall, M., & Kousha, K. (2014). Academia.edu: Social network or academic network? Journal of the Association for Information Science and Technology, 65(4), 721–731. Thelwall, M., & Kousha, K. (2015a). ResearchGate: Disseminating, communicating, and measuring Scholarship? Journal of the Association for Information Science and Technology, 66(5), 876– 889. Thelwall, M., & Kousha, K. (2015b). Web indicators for research evaluation, part 1: Citations and links to academic articles from the web. El Profesional de La Información, 24(5), 587–606. Thelwall, M., & Kousha, K. (2017a). ResearchGate articles: Age, discipline, audience size, and impact. Journal of the Association for Information Science and Technology, 68(2), 468–479. Thelwall, M., & Kousha, K. (2017b). ResearchGate versus Google Scholar: Which finds more early citations? Scientometrics. doi:10.1007/s11192-017-2400-4 Thelwall, M., Kousha, K., Dinsmore, A., & Dolby, K. (2016). Alternative metric indicators for funding scheme evaluations. Aslib Journal of Information Management, 68(1), 2–18. Thelwall, M., & Maflahi, N. (2015). Are scholarly articles disproportionately read in their own country? An analysis of Mendeley readers. Journal of the Association for Information Science and Technology, 66(6), 1124–1135. Thelwall, M., & Maflahi, N. (2016). Guideline references and academic citations as evidence of the clinical value of health research. Journal of the Association for Information Science and Technology, 67(4), 960–966. Thelwall, M., & Sud, P. (2016). Mendeley readership counts: An investigation of temporal and disciplinary differences, 67(12), 3036–3050. Thelwall, M., & Wilson, P. (2016). Mendeley readership altmetrics for medical articles: An analysis of 45 fields. Journal of the Association for Information Science and Technology, 67(8), 1962– 1972. Thelwall, M. (2017a). Does Mendeley provide evidence of the educational value of journal articles? Learned Publishing, 30(2), 107–113. Thelwall, M. (2017b). Three practical field normalised alternative indicator formulae for research evaluation. Journal of Informetrics, 11(1), 128–151. Thelwall, M. (2017c). Why do papers have many Mendeley readers but few Scopus-indexed citations and vice versa? Journal of Librarianship and Information Science, 49(2), 144-151. Thelwall, M. (2017d). Web indicators for research evaluation: A practical guide. San Rafael, CA: Morgan & Claypool. Thelwall, M. (2017e). Microsoft Academic: A multidisciplinary comparison of citation counts with Scopus and Mendeley for 29 journals. Journal of Informetrics, 11(4), 1201-1212. Thelwall, M. (2017f). Are Mendeley reader counts useful impact indicators in all fields? Scientometrics, 113(3), 1721–1731. Thelwall, M. (in press). Do Mendeley reader counts indicate the value of arts and humanities research? Journal of Librarianship & Information Science. Tsay, M. Y. (1998). The relationship between journal use in a medical library and citation use. Bulletin of the Medical Library Association, 86(1), 31–9. Van der Vaart L, Van Berchum M, Bruce R, Burgess M, Hanganu G, Jacobs N, Lecarpentier D, Pinfield S & Stokes P (2013) ‘Open’ as the default modus operandi for research and higher education. European Network for Co-ordination of Policies and Programmes on eInfrastructures. Available at: http://e-infranet.eu/wp-content/uploads/2013/03/eInfraNet-Open-as-the-Default-Modus-Operandi-for- Research-and-Higher-Education.pdf. van Raan, A. F. J. (1998). In matters of quantitative studies of science the fault of theorists is offering too little and asking too much. Scientometrics, 43(1), 129-139. van Raan, A. F. J. (2000). The Pandora’s Box of citation analysis: Measuring scientific excellence the last evil? In E. Garfield, B. Cronin, & H. B. Atkins (Eds.), The Web of Knowledge: A Festschrift in Honor of Eugene Garfield (pp. 301–320). Medford, New Jersey: ASIS Monograph Series: Information Today. Vaughan, L., Tang, J., & Yang, R. (2017). Investigating disciplinary differences in the relationships between citations and downloads. Scientometrics, 111(3), 1533-1545. Waltman, L., van Eck, N. J., van Leeuwen, T. N., Visser, M. S., & van Raan, A. F. J. (2011). Towards a new crown indicator: Some theoretical considerations. Journal of Informetrics, 5(1), 37– 47. Wan, J., Hua, P., Rousseau, R., & Sun, X. (2010). The journal download immediacy index (DII): experiences using a Chinese full-text database. Scientometrics, 82(3), 555–566. Welch, I. (2014). Referee recommendations. Review of Financial Studies, 27(9), 2773–2804. Weller, K., & Peters, I. (2012). Citations in Web 2.0. In A. Tokar, M. Beurskens, S. Keuneke, M. Mahrt, I. Peters, C. Puschmann, K. Weller (Eds.), Science and the Internet (pp. 209–222). Düsseldorf. Germany: Düsseldorf University Press. White, H. D., Boell, S. K., Yu, H., Davis, M., Wilson, C. S., & Cole, F. T. (2009). Libcitations: A measure for comparative assessment of book publications in the humanities and social sciences. Journal of the American Society for Information Science and Technology, 60(6), 1083-1096. White, H. D., & McCain, K. W. (1989). Bibliometrics. Annual Review of Information Science and Technology, 24, 119–186. Whitmire, E. (2002). Disciplinary differences and undergraduates’ information-seeking behavior. Journal of the American Society for Information Science and Technology, 53(8), 631–638. Winslow, R. R., Skripsky, S. L., & Kelly, S. L. (2016). Not just for citations: Assessing Zotero while reassessing research. In B. J. D’Angelo, S. Jamieson, B. Maid, & J. R. Walker (Eds.), Information literacy: Research and collaboration across disciplines (pp. 287–304). Fort Collins, CO: WAC Clearinghouse and University Press of Colorado. Wouters, P. (1997). Citation cycles and peer review cycles. Scientometrics, 38(1), 39–55. Wouters, P., & Costas, R. (2012). Users, narcissism and control – tracking the impact of scholarly publications in the 21st century. Image Rochester NY, 50 pages. Retrieved from http://www.surffoundation.nl/en/publicaties/Pages/Users_narcissism_control.aspx Wulff, J. L., & Nixon, N. D. (2004). Quality markers and use of electronic journals in an academic health sciences library. Journal of the Medical Library Association, 92(3), 315–22. Xue-li, L., Hong-ling, F., & Mei-ying, W. (2011). Correlation between Download and Citation and Download-citation Deviation Phenomenon for Some Papers in Chinese Medical Journals. Serials Review, 37(3), 157–161. Yan, K.-K., & Gerstein, M. (2011). The spread of scientific information: insights from the web usage statistics in PLoS article-level metrics. PLoS ONE, 6(5), e19917. Zahedi, Z., Costas, R., & Wouters, P. (2014). How well developed are altmetrics? A crossdisciplinary analysis of the presence of “alternative metrics” in scientific publications. Scientometrics, 101(2), 1491–1513. Zahedi, Z., Costas, R., & Wouters, P. (2015). Do Mendeley readership counts help to filter highly cited WoS publications better than average citation impact of journals (JCS)? In A. A. Salah, Y. Tonta, A. A. A. Salah, C. Sugimoto, & U. Al (Eds.), Proceedings of the 15th International Conference of the International Society for Scientometrics and Informetrics (pp. 16–25). Istanbul, Turkey: Boğaziçi University Printhouse. Zimmermann, C. (2013). Academic rankings with RePEc. Econometrics, 1(3), 249–280.