Five Types of Citation Gaming: Data-Driven Taxonomy for Research Integrity and Academic Fraud
March 30, 2026 | By Billy Wong
In the current scholarly ecosystem, academic success is increasingly mediated by quantitative proxies: citation counts, the h-index, and the Journal Impact Factor (JIF). However, as these metrics have transitioned from descriptive tools to career-defining targets, they have become subject to Goodhart’s Law: When a measure becomes a target, it ceases to be a good measure.
This report provides a formal taxonomy of Citation Gaming (also known as citation manipulation)—the systematic and unethical inflation of metrics to distort perceived scientific influence. By leveraging our high-performance proprietary measuresHE analytics pipeline, we categorize these manipulations into five distinct archetypes, providing a framework to identify and reference primary data signatures of research misconduct and academic fraud through unethical citation practices.
Executive Summary: Citation Gaming Taxonomy Archetypes
| Archetype | Primary Target | Detection Signature |
|---|---|---|
| Self-Citation Outliers | Author H-Index | Excessive (>30%) self-citation rate; citations focused on "h-threshold" papers. |
| Citation Cartels | Group/Department Rank | High reciprocity between unrelated authors; closed-loop citation graphs. |
| Coercive Citation | Journal Impact Factor | Disproportionate citations to a specific journal as a condition of peer review. |
| Paper Mills & Digital Fraud | Volume-based Metrics | Fabricated DOIs; illicit merging of records on Google Scholar; fake preprints. |
| Affiliation Hijacking | Institutional Rankings | Highly cited scholars listed with numerous short duration affiliations |
1. Author-Level Manipulation: Self-Citation and H-Index Targeting
The most prevalent form of gaming involves excessive self-citation. While a baseline of self-referencing is natural for ongoing research, our analysis of the OpenAlex dataset identifies outliers where authors disproportionately cite their own work to artificially inflate their "cited by" count.
Recent metadata trends show a sharp increase in mean and maximum self-citations per paper. A specialized sub-tactic is h-index gaming: authors strategically cite their specific papers that sit just below their current h-index threshold, pushing the metric higher with surgical precision.
Our report The Great Citation Heist details some examples and tactics used by these fraudsters.
2. Collaborative Misconduct: Citation Cartels and Swapping
Citation manipulation is frequently carried out through collaborative networks known as citation cartels, rings, or farms. This occurs when groups of colleagues or seemingly unrelated authors agree to preferentially and routinely cite each other's articles. In extreme cases, a very small, concentrated circle may generate 30% to 80% of the total citations a specific researcher receives. These closed-loop networks are clearly visible when mapping the citation graph of the OpenAlex dataset.
3. Gateway Coercion: Editor and Reviewer-Driven Gaming
Gatekeepers in the peer-review process are also heavily involved in citation gaming. A prominent tactic is coercive citation, where a journal editor or referee forces an author to add spurious citations to the journal's own articles as a condition for publication. These requests typically lack scientific justification and exist only to artificially increase the Journal Impact Factor (JIF).
Other tactics include:
- Reviewers abusing anonymity by suggesting authors cite the reviewer's own work.
- Editors writing editorials that disproportionately cite their own journal’s articles.
- Collusion between journals to cross-cite each other in a practice known as "citation stocking."
- Universities compelling graduates to cite the university's own publications to inflate its ranking performance.
4. High-Tech Manipulation: Paper Mills and Metadata Fraud
As scholarly publishing moves fully into the digital age, a lucrative black market has evolved. A major threat is the use of paper mills, i.e. for-profit companies that produce and sell fabricated research papers, to authors willing to buy authorship slots. These fake papers, alongside fake preprints, are then systematically used to artificially bolster citation counts.
Fraudsters also use tech-savvy methods like indexing false papers and illicitly merging distinct papers on platforms like Google Scholar to inflate metrics. In one bizarre case, researchers successfully listed papers on Google Scholar that had been "authored" by a cat, and then cited them in fake papers posted on ResearchGate, highlighting the extreme fragility of purely digital metrics.
5. Institutional Academic Fraud: Ranking-Chasing via Affiliation Hijacking
The obsession with rankings extends to entire universities trying to climb global league tables. Some institutions engage in ranking-chasing behaviours by paying highly cited, prestigious scholars to simply list the university as their affiliation on their next major paper. This allows the university to benefit from the scholar's citations without having actually contributed to or supported the research. This practice, termed "affiliation hijacking," significantly distorts the competitive landscape of higher education.
Conclusion and Recommendation
Ultimately, the rampant gaming of citation metrics threatens the integrity of the scientific record. It demonstrates a critical need for the academic community and indexing services to rethink how they evaluate and reward scholarly impact.
In the upcoming weeks, we will release further reports to examine each of the citation gaming archetypes in more details.
Primary Data Resource: OpenAlex provides the necessary datasets and Motherduck the query capabilities for identifying these citation gaming archetypes across all scholarly domains.
Glossary
- Affiliation Hijacking: Affiliation hijacking is an institutional ranking-chasing practice where universities pay highly cited scholars to list the university as their affiliation on new papers, which allows the institution to absorb the benefit of the scholar's citations without having supported the research, thereby distorting the competitive landscape of higher education.
- Citation Cartels: Citation cartels, also known as rings or farms, are collaborative networks of colleagues or seemingly unrelated authors who agree to preferentially and routinely cite each other's articles to artificially inflate their citation counts and boost their group or department rank. These closed-loop networks, which can generate 30% to 80% of a specific researcher's total citations in extreme cases, are visible through their high reciprocity between authors when mapping the citation graph.
- Citation Gaming: Citation Gaming, also known as citation manipulation, is defined as the systematic and unethical inflation of quantitative metrics, such as citation counts, the H-index, and the Journal Impact Factor, to distort perceived scientific influence and mediate academic success. The practice is categorized into distinct archetypes, including coercive citation and self-citation outliers, and provides primary data signatures of research misconduct and academic fraud through unethical citation practices.
- Coercive Citation: Coercive citation is a prominent, unethical tactic used by gatekeepers in the peer-review process, where a journal editor or referee forces an author to add spurious citations to the journal's own articles as a condition for publication, serving only to artificially increase the journal's Impact Factor (JIF). These requests typically lack scientific justification.
- Goodhart’s Law: “When a measure becomes a target, it ceases to be a good measure”. It is named after British economist Charles Goodhart, who is credited with expressing the core idea of the adage in a 1975 article on monetary policy in the United Kingdom.
- H-index: The H-index is a quantitative metric used in research evaluation to simultaneously measure a researcher's productivity and citation impact, defined as the largest number h such that h of their published articles have received at least h citations each. In the scholarly ecosystem, this metric acts as a quantitative proxy for academic success, which has made it a primary target for manipulation.
- Journal Impact Factor (JIF): A metric that measures the average number of times articles published in a scientific journal are cited in the two years following their publication. It is calculated annually as:
- Impact Factor = (Number of citations in year Y to articles published in the journal in years Y-1 and Y-2) ÷ (Number of citable articles published in the journal in years Y-1 and Y-2).
Tags: Academic Fraud Citation Gaming Research Integrity Scholarly Publishing