Continuous, Dynamic and Comprehensive Article-Level Evaluation of Scientific Literature
It is time to make changes to the current research evaluation system, which is built on the journal selection. In this study, we propose the idea of continuous, dynamic and comprehensive article-level-evaluation based on article-level-metrics. Different kinds of metrics are integrated into a comprehensive indicator, which could quantify both the academic and societal impact of the article. At different phases after the publication, the weights of different metrics are dynamically adjusted to mediate the long term and short term impact of the paper. Using the sample data, we make empirical study of the article-level-evaluation method.
💡 Research Summary
The paper critiques the prevailing reliance on journal‑based metrics such as Impact Factor, Web of Science, and Nature Index for research evaluation, arguing that these approaches cannot capture the heterogeneous impact of individual articles. To address this, the authors propose a “continuous, dynamic, and comprehensive” article‑level evaluation framework that integrates multiple metrics reflecting both academic and societal influence. They identify seven metric categories—citations, HTML views, PDF downloads, bookmarks (CiteULike), readership (Mendeley), and social media mentions (Facebook, Twitter)—and collect longitudinal data for 46 articles published in PLOS Computational Biology (June 2012) from eight ALM reports spanning two years.
Factor analysis (principal component with Varimax rotation) reveals two latent factors: an “academic impact” factor loading heavily on CiteULike, Mendeley, PDF downloads, and Scopus citations, and a “societal impact” factor dominated by Facebook and Twitter. HTML views load on both factors, indicating a dual role. Temporal analysis shows that usage metrics (downloads, views) and social media peak early (within months), while citations rise sharply after about a year, confirming distinct dynamic patterns.
Based on these observations, the authors design a phase‑dependent weighting scheme (0‑6 months, 6 months‑2 years, 2‑5 years, >5 years). In early phases, usage metrics receive the highest weight; later phases shift emphasis toward citations and scholarly bookmarks. Weights are derived using the Analytic Hierarchy Process (AHP), constructing pairwise comparison matrices for each phase and applying the 1‑9 Saaty scale. The resulting weight hierarchies are presented in tabular form.
The paper also highlights the limitations of existing databases: Web of Science’s journal‑centric indexing fails to differentiate article‑level performance, and the Nature Index’s narrow journal selection excludes a large share of scientific output. Empirical evidence is offered (e.g., 21 % of 2000‑year chemical engineering articles in SCI received zero citations).
Methodological shortcomings are acknowledged: the sample is confined to PLOS, the AHP judgments are not disclosed, and only two factors emerge from a modest dataset, limiting generalizability. The authors suggest future work should expand to multi‑disciplinary, multi‑publisher datasets, employ data‑driven weighting (e.g., machine‑learning regression), incorporate field‑normalized citation and altmetric scores, and validate the composite indicator against real‑world evaluation outcomes such as grant decisions or promotion.
In sum, the study introduces a conceptual model for article‑level assessment that dynamically balances short‑term societal attention with long‑term scholarly impact, offering a potential pathway toward more nuanced and timely research evaluation.
Comments & Academic Discussion
Loading comments...
Leave a Comment