Tuesday, 28 July 2015

Network-based Citation Metrics: Eigenfactor vs. SJR | The Scholarly Kitchen

 Source: http://scholarlykitchen.sspnet.org/2015/07/28/network-based-citation-metrics-eigenfactor-vs-sjr/

Network-based Citation Metrics: Eigenfactor vs. SJR

Journal Citation Network (or just a duck).
Is the influence of a journal best measured by the number of
citations it attracts or by the citations it attracts from other
influential journals?


The purpose of this post is to describe, in plain English, two
network-based citation metrics: Eigenfactor[1] and SCImago Journal Rank
(SJR)[2], compare their differences, and evaluate what they add to our
understanding of the scientific literature.


Both Eigenfactor and SJR are based on the number of citations a
journal receives from other journals, weighted by their importance, such
that citations from important journals like Nature are given
more weight than less important titles. Later in this post, I’ll
describe exactly how a journal derives its importance from the network.


In contrast, metrics like the Impact Factor do not weight citations:
one citation is worth one citation, whatever the source. In this sense,
the Eigenfactor and SJR get closer to measuring importance as a social
phenomenon, where influential people hold more sway over the course of
business, politics, entertainment and the arts. For the Impact Factor,
importance is equated with popularity.


Eigenfactor and SJR are both based on calculating something called eigenvector centrality,
a mathematical concept that was developed to understand social networks
and first applied to measuring journal influence in the mid-seventies.
Google’s PageRank is based on the same concept.


Eigenvector centrality is calculated recursively, such
that values are transferred from one journal to another in the network
until a steady-state solution (also known as an equilibrium) is
reached. Often 100 or so iterations are used before values become
stable. Like a hermetically sealed ecosystem, value is neither created
nor destroyed, just moved around.


There are two metaphors used to describe this process: The first conceives of the system as a fluid network,
where water drains from one pond (journal) to the next along citation
tributaries. Over time, water starts accumulating in journals of high
influence while others begin to drain. The other metaphor conceives of a
researcher taking a random walk from one journal to the next
by way of citations. Journals visited more frequently by the wandering
researcher are considered more influential.


However, both of these models break down (mathematically and
figuratively) in real life. Using the fluid analogy, some ponds may be
disconnected from most of the network of ponds; if there is just one
stream feeding this largely-disconnected network, water will flow in,
but not out. After time, these ponds may swell to immense lakes, whose
size is staggeringly disproportionate to their starting values. Using
the random walk analogy, a researcher may be trapped wandering among a
highly specialized collection of journals that frequently cite each
other but rarely cite journals outside of their clique.


The eigenvector centrality algorithm can adjust for this problem by
“evaporating” some of the fluid in each iteration and redistributing
these values back to the network as rain. Similarly, the random walk
analogy uses a “teleport” concept, where the researcher may be
transported randomly to another journal in the system–think of Scotty
transporting Captain Kirk back to the Enterprise
before immediately beaming him down to another planet.


Before I continue into details and differences, let me summarize thus
far: Eigenfactor and SJR are both metrics that rely on computing,
through iterative weighting, the influence of a journal based on the
entire citation network. They differ from traditional metrics, like the
Impact Factor, that simply compute an unweighted average.


In practice, eigenvector centrality is calculated upon an adjacency matrix
listing all of the journals in the network and the number of citations
that took place between them. Most of the values in this very large
table are zero, but some will contain very high values, representing
large flows of citations between some journals, for instance, between
the NEJM, JAMA, The Lancet, and BMJ.


The result of the computation–a transfer of weighted values from one
journal to the next over one hundred or so iterations–represents the
influence of a journal, which is often expressed as a percentage of the total influence in the network. For example, Nature‘s
2014 Eigenfactor was 1.50, meaning that this one journal holds 1.5% of
the total influence of the entire citation network. In comparison,
a smaller, specialized journal, AJP-Renal Physiology, received an Eigenfactor of 0.028. PLOS ONE’s Eigenfactor was larger than Nature’s (1.53) as a result of its immense size. Remember that Eigenfactor measures total influence in the citation network.


When the Eigenfactor is adjusted for the number of papers published in each journal, it is called the Article Influence Score. This is similar to SCImago’s SJR. So, while PLOS ONE had an immense Eigenfactor, its Article Influence Score was just 1.2 (close to average performance), compared to 1.1 for AJP-Renal Physiology, and 21.9 for Nature.


This year, Thomson Reuters began publishing a Normalized Eigenfactor, which expresses the Eigenfactor as a multiplicative
value rather than a percent. A journal with a value of 2 has twice as
much influence as the average journal in the network, whose value would
be one. Nature‘s Normalized Eigenfactor was 167, PLOS ONE was 171, while AJP-Renal Physiology was 3.


There are several differences between how the Eigenfactor and SJR are
both calculated, meaning they cannot be used interchangeably:


  1. Size of the network. Eigenfactor is based on the
    citation network of just over 11,000 journals indexed by Thomson
    Reuters, whereas the SJR is based on over 21,000 journals indexed in
    the Scopus database. Different citation networks will result in
    different eigenvalues.
  2. Citation window. Eigenfactor is based on citations
    made in a given year to papers published in the prior five years, while
    the SJR uses a three-year window. The creators of Eigenfactor argue that
    five years of data reduces the volatility of their metric from year to
    year, while the creators of the SJR argue that a three-year window
    captures peak citation for most fields and is more sensitive to the
    changing nature of the literature.
  3. Self-citation. Eigenfactor eliminates
    self-citation, while SJR allows self-citation but limits it to no more
    than one-third of all incoming citations. The creators of Eigenfactor
    argue that eliminating self-citation disincentivizes bad referencing
    behavior, while the creators of the SJR argue that self-citation is part
    of normal citation behavior and wish to capture it.
There are other small differences, such as the scaling factor (a
constant that defines how much “evaporation”or “teleporting”) that takes
place in each iteration. While both groups provide a full description
of their algorithm (Eigenfactor here; SJR here)
it is pretty clear that few of us (publishers, editors, authors) are
going to replicate their work. Indeed, these protocols assume that
you’ve already indexed tens of thousands of journals publishing several
million papers listing tens of millions of citations before you even
begin to assemble your adjacency matrix. And no, Excel doesn’t have a
simple macro for calculating eigenvalues. So while each group is fully
transparent in its methods, the shear enormity and complexity of the
task prevents all but the two largest indexers from replicating their
results. A journal editor really has no recourse but to accept the
numbers provided to him.


If you scan performances of journals, you’ll notice that journals
with the highest Impact Factor also have the highest Article
Influence and SJR scores, leaving one to question whether popularity in
science really measures the same underlying construct as influence.
 Writing in the Journal of Informetrics, Massimo Francechet reports
that for the biomedical, social sciences, and geosciences, 5-yr Impact
Factors correlate strongly with Article Influence Scores, but diverge
more for physics, material sciences, computer sciences, and engineering.
For these fields, journals may perform well one one metric but poorly
on the other. In another paper focusing
on the SJR, the authors noted some major changes in the ranking of
journals, and reported that eigenvalues tended to concentrate in fewer
(prestigious) journals. Considering how the metric is calculated,
this should not be surprising.


In conclusion, network-based citation analysis can help us more
closely measure scientific influence. However, the process is complex,
not easily replicable, harder to describe and, for most journals, gives
us the same result as much simpler methods. Even if not widely adopted
for reporting purposes, the Eigenfactor and SJR may be used for
detection purposes, such as identifying citation cartels
and other forms of citation collusion that are very difficult to detect
using traditional counting methods, but may become highly visible using
network-based analysis.


Notes:


1. Eigenfactor (and Article Influence) are terms trademarked by the University of Washington. Eigenfactors and Article Influence scores are published in the Journal Citation Report (Thomson Reuters) each June and are posted freely on the Eigenfactor.org
website after a six-month embargo. To date, the University of
Washington has not received any licensing revenue from Eigenfactor
metrics.


2. The SCImago Journal & Country Rank is based on Scopus data (Elsevier) and made freely available from: http://www.scimagojr.com/


About Phil Davis

I am an independent researcher and publishing consultant
specializing in the statistical analysis of readership and citation
data. I am a former postdoctoral researcher in science communication and
former science librarian. http://phil-davis.org/

Discussion

10 thoughts on “Network-based Citation Metrics: Eigenfactor vs. SJR




  1. Thanks Phil, this is an interesting post.


    The network based metrics are intellectually interesting, but as you
    point they are effectively black boxes, we have to take it on trust that
    the calculations have been done correctly and right number of citations
    included or excluded. Whereas it is reasonably straight forward to
    estimate an Impact Factor from the Web of Science. These metrics also
    fail in terms of the elevator pitch: I can explain an Impact Factor or
    other simple metrics in 30 seconds, the network based metrics take
    minutes to explain and even that glosses over the details.


    My own quick analysis of the Web of Science metrics shows two groups
    that are closely correlated by journal rank, Total Citations and
    Eigenfactor in one and the article weighted metrics such as Impact
    Factor and Article Influence Score in the other. The ever growing range
    of citation metrics doesn’t appear to add much extra information but
    does give journals another way to claim to be top. I also think that
    these metrics apply less in the social sciences and humanities where the
    ‘high prestige journals’ often have less difference in citation
    profile.









    Posted by James Hardcastle (@JwrHardcastle) | Jul 28, 2015, 6:10 am





  2. The exclusion of self citation seems rather strange. A specialized
    journal, which many are, may well have most of the articles on its
    topic. Later articles will certainly cite numerous prior articles, in
    the normal course of citation. This is a strong measure of the journal’s
    local importance. So excluding self citation appears to penalize that
    specialization which serves a specific research community.









    Posted by David Wojick | Jul 28, 2015, 7:17 am





  3. Thank you for an excellent explanation of this complex but
    important subject. One question I’d like to see explored: How well do
    the three indices (Eigenfactor, SJR, JIF) correlate? That is, how much
    does choosing one or the other affect a journal’s relative ranking with
    competing journals?









    Posted by Ken Lanfear | Jul 28, 2015, 10:12 am





  4. For the Impact Factor, importance is equated with popularity.
    Given that the IF (like all other citation metrics, as far as I’m
    aware) makes no effort to discriminate between approving and
    disapproving citations, wouldn’t it be more accurate to say that for the
    IF, importance is equated with notoriety rather than popularity?









    Posted by Rick Anderson | Jul 28, 2015, 10:26 am





    • My conjecture is that, in the physical sciences and engineering at
      least, negative citations are rare enough to be negligible. HSS may be
      different. It is a good research question, so I wonder if any work has
      been done on it. There is a lot of research on distinguishing positive
      and negative tweets, but maybe not citations.









      Posted by David Wojick | Jul 28, 2015, 12:05 pm





  5. Phil, I like the evaporation metaphor for the flow/voting
    interpretation of Eigenvector centrality. It’s a nice complement to the
    teleport metaphor for the random walk interpretation. It’s fun to see
    each interpretation of the algorithm in action, so you and your readers
    may enjoy the demo that Martin Rosvall and Daniel Edler put together to
    illustrate both the random walk interpretation and the flow
    interpretation. See http://www.mapequation.org/apps/MapDemo.html


    To play with the demo, click on “rate view” at the top center of the
    screen. Then you can click on “random walker” at the top to look at the
    random walk interpretation, using the “step” or “start” buttons to set
    the random walker into motion. Then reset and click on “Init Votes” to
    restart. Click on “Vote” or “Automatic Voting” to view the flow
    interpretation. In both cases, the bar graph at right shows each process
    converge to the leading eigenvector of the transition matrix.


    By the way, my view is that the most important difference between the
    Eigenfactor algorithm and the SJR approach is that in the Eigenfactor
    algorithm, the random walker takes one final step along the citation
    matrix *from the stationary distribution*, without teleporting. This
    ensures that no journal receives credit for teleportation (or
    evaporation and condensation) — the only credit comes from being cited
    directly. We’ve found this step extremely important in assigning
    appropriate ranks to lower-tier journals, whose ranks otherwise heavily
    influenced by the teleport process. In the demo linked above, you can
    see how this affects the final rankings by pressing the “Eigenfactor”
    button.


    Carl Bergstrom

    eigenfactor.org / University of Washington









    Posted by Eigenfactor Project (@eigenfactor) | Jul 28, 2015, 5:16 pm


Trackbacks/Pingbacks


  1. Pingback: Network-based Citation Metrics: Eigenfactor vs. SJR | Nader Ale Ebrahim - Jul 28, 2015



Leave a Reply



Network-based Citation Metrics: Eigenfactor vs. SJR | The Scholarly Kitchen

No comments:

Post a Comment