The number of citations is the most commonly used metric for quantifying the importance of scientific publications. However, we all have anecdotal experiences that citations alone do not characterize the importance of a publication. Some of the shortcomings of using citations as a universal measure of importance include the following.
-
(1) It ignores the importance of citing papers: a citation from an obscure paper is given the same weight as a citation from a ground-breaking and highly cited work.
-
(2) The number of citations is ill suited to compare the impact of papers from different scientific fields. Due to factors such as size of a field and disparate citation practices, the average number of citations per paper varies widely between disciplines. An average paper is cited ∼6 times in life sciences, 3 times in physics, and <1 times in mathematics.
-
(3) Many groundbreaking older articles are modestly cited due to a smaller scientific community when they were published. Furthermore, publications on significant discoveries often stop accruing citations once their results are incorporated into textbooks. Thus, citations consistently underestimate the importance of influential old papers.
These and related shortcomings of citation numbers are partially obviated by Google's PageRank algorithm (Brin and Page, 1998). As we shall discuss, PageRank gives higher weight to publications that are cited by important papers and also weights citations more highly from papers with few references. Because of these attributes, PageRank readily identifies a large number of scientific “gems”: modestly cited articles that contain ground-breaking results.
In a recent study (Chen et al., 2007), we applied Google's PageRank to the citation network of the premier American Physical Society (APS) family of physics journals (Physical Review A–E, Physical Review Letters, Reviews of Modern Physics, and Physical Review Special Topics). Our study was based on all 353,268 articles published in APS journals since their inception in 1893 until June 2003 that have at least one citation from within this dataset. This set of articles has been cited a total of 3,110,839 times by other APS publications. Our study is restricted to internal citations—citations to APS articles from other APS articles. Other studies (Bollen et al., 2006; Bergstrom, 2007) use the PageRank algorithm on a coarse-grained scale of individual scientific journals to formulate an alternative to the Impact Factor.
We can think of the set of all APS articles and their citations as a network, with nodes representing articles and a directed link between two nodes representing a citation from a citing article to a cited article. In Google's PageRank algorithm (Brin and Page, 1998), a random surfer is initially placed at each node of this network and its position is updated as follows: (1) with probability 1 − d, a surfer hops to a neighboring node by following a randomly selected outgoing link from the current node; (2) with probability d, a surfer “gets bored” and starts a new search from a randomly selected node in the entire network. This update is repeated until the number of surfers at each node reaches a steady value, the Google number. These nodal Google numbers are then sorted to determine the Google rank of each node.
The original Brin-Page PageRank algorithm used the parameter value d = 0.15 based on the observation that a typical web surfer follows of the order of six hyperlinks, corresponding to a boredom attrition factor d = 1/6 ≃ 0.15, before aborting and beginning a new search. The number of citation links followed by researchers is generally considerably <6. In Chen et al. (2007) and Walker et al. (2007), we argued that in the context of citations the appropriate choice is d = 1/2, corresponding to a citation chain to two links.
The Google number G and number of citations k to a paper are approximately proportional to each other for k ≳ 50 (Chen et al., 2007) so that, on average, they represent similar measures of importance (Fortunato et al., 2006a,b). The outliers with respect to this proportionality are of particular interest (Fig. 1). Although the top APS papers by PageRank are typically very well cited, some are modestly cited and quite old. The disparity between PageRank and citation rank arises because the former involves both the number of citations and quality of the citing articles. The PageRank of a paper is enhanced when it is cited by its scientific “children” that themselves have high PageRank and when these children have short reference lists. That is, children should be influential and the initial paper should loom as an important “father figure” to its children.
Scatter plot of the Google number versus number of citations for APS publications with >100 citations. The top Google-ranked papers (filled) are identified by author(s) initials and their details are given in Table 1. The solid curve is the average Google number of papers versus number of citations when binned logarithmically.
The top articles according to Google's PageRank would be recognizable to almost all physicists, independent of their specialty (Table 1). For example, Onsager's (1944) exact solution of the two-dimensional Ising model (Fig. 1, the point labeled with O) was both a calculational tour de force as well as a central development in critical phenomena. The paper by Feynman and Gell-Mann (FG) introduced the V − A theory of weak interactions that incorporated parity nonconservation and became the “standard model” of the field. Anderson's paper (A), “Absence of Diffusion in Certain Random Lattices” gave birth to the field of localization and was cited for the 1977 Nobel prize in physics. Particularly intriguing are the articles “The Theory of Complex Spectra,” by J. C. Slater (S) and “On the Constitution of Metallic Sodium” by E. Wigner and F. Seitz (WS) with relatively few APS citations (114 and 184, respectively, as of June 2003). Slater's paper introduced the determinant form of the many-body wave-function that is so ubiquitous that the original work is no longer cited. Similarly, the latter paper introduced the Wigner–Seitz construction that constitutes an essential curriculum component in any solid-state physics course.
Top non-review-article APS publications ranked according to PageRank (PR), citation number (CNR), and CiteRank (CR) (Walker et al., 2007)
The PageRank algorithm was originally developed to rank web pages that are connected by hyperlinks and not papers connected by citations. The most important difference between these two networks is that, unlike hyperlinks, citations cannot be updated after publication. The constraint that a paper may only cite earlier works introduces a time ordering to the citation network topology. This ordering makes aging effects much more important in citation networks than in the World-Wide Web. The habits of scientists looking for relevant scientific literature are also different from web surfers. Apart from the already-mentioned shorter depth search (1/d = 2), scientists often start literature searches with a paper of interest that they saw in a recent issue of a journal or heard presented at a conference. From this starting point a researcher typically follows chains citations that lead to progressively older publications.
These observations led us to modify the PageRank algorithm by initially distributing random surfers exponentially with age, in favor of more recent publications (Walker et al., 2007). This algorithm, CiteRank, is characterized by just two parameters: d (the inverse of the average citation depth) and τ (the time constant of the bias toward more recent publications at the start of searches). The optimal values of these parameters that best predict the rate at which publications acquire recent citations are d = 0.5 and τ = 2.6 years for APS publications. With these values, the output of the CiteRank algorithm quantifies the relevance of a given publication in the context of currently popular research directions, whereas that of PageRank corresponds to its “lifetime achievement award.” The database with the results of application of both algorithms to the citation network of APS publications can be accessed online at http://www.cmth.bnl.gov/∼maslov/citerank.
Google's PageRank algorithm and its modifications hold great promise for quantifying the impact of scientific publications. They provide a meaningful extension to traditionally used importance measures, such as the number of citations to articles and the impact factor for journals. PageRank implements, in a simple way, the logical notion that citations from more important publications should contribute more strongly to the rank of a cited paper. PageRank also effectively normalizes the impact of papers in different scientific communities (Xie et al., 2007). Other ways of attributing a quality for citations would require detailed contextual information about citation themselves, features that are presently unavailable. PageRank implicitly includes context by incorporating the importance of citing publications. Thus, PageRank represents a computationally simple and effective way to evaluate the relative importance of publications beyond simply counting citations.
We conclude with some caveats. It is very tempting to use citations and their refinements, such as PageRank, to quantify the importance of publications and scientists (Hirsch, 2005), especially as citation data becomes increasingly convenient to obtain electronically. In fact, the h-index of any scientist, a purported single-number measure of the impact of an individual scientific career, is now easily available from the Web of Science (http://apps.isiknowledge.com/). However, we must be vigilant for the overuse and misuse of such indices. All the citation measures devised thus far pertain to popularity rather than to the not-necessarily-coincident attribute of intrinsic intellectual value. Even if a way is devised to attach a high-fidelity quality measure to a citation, there is no substitute for scientific judgment to assess publications. We need to avoid falling into the trap of relying on automatically generated citation statistics for accessing the performance of individual researchers, departments, and scientific disciplines, and especially of allowing the evaluation task to be entrusted to administrators and bureaucrats (Adler et al., 2008).
Footnotes
-
Editor's Note: The misuse of journal impact factor in hiring and promotion decisions is a growing concern. This article is one in a series of invited commentaries in which authors discuss this problem and consider alternative measures of an individual's impact.
-
Work by S.M. at Brookhaven National Laboratory was performed under contract DE-AC02-98CH10886, Division of Material Science, U.S. Department of Energy. S.R. gratefully acknowledges financial support from U.S. National Science Foundation Grant DMR0535503.
- Correspondence should be addressed to Sidney Redner at the above address. redner{at}bu.edu