Understanding Research Authority Scores: GRAS and TRAS

When evaluating research impact, most systems rely on simple citation counts or field-weighted citation indices (FWCI). While these metrics are easy to understand, they miss a crucial dimension of research influence: citation quality.

Consider two quantum computing papers, both published in 2023:

Paper A: Cited 200 times by obscure journals and low-impact work Paper B: Cited 50 times by highly influential papers

Raw citation counts suggest Paper A is four times more impactful. Field-weighted citation indices (FWCI) partially correct for field differences, but still treat all citations equally within a field. Neither metric captures what experts know intuitively: Paper B is more influential because it's recognized by the field's leading researchers.

Research Authority Scores solve this by measuring influence through network analysis, distinguishing between citations from influential versus marginal work.

What Are Research Authority Scores?

Research Authority Scores measure a paper's influence based on the recursive principle that important papers are cited by other important papers. Rather than counting citations, we analyze the citation network to identify which papers occupy central, influential positions.

Factbase provides two Authority Score metrics:

Global Research Authority Score (GRAS)

Scope: Measures a paper's influence across all research disciplines
Scale: 0-100 percentile
Interpretation: A GRAS of 85 means this work is more influential than 85% of all research published that year, regardless of field.

Topic Research Authority Score (TRAS)

Scope: Measures a paper's influence within its specific research domain
Scale: 0-100 percentile
Interpretation: A TRAS of 85 means this work is more influential than 85% of research in its topic published that year

Both metrics are percentile-based, making them immediately interpretable: higher scores indicate higher influence.

The Mathematics: Network-Based Authority Calculation

Research Authority Scores are derived from eigenvector centrality analysis of the citation network. This approach, rooted in spectral graph theory, has been validated across billions of web pages and academic papers for over 25 years.

Citation Networks as Directed Graphs

We model the research literature as a directed graph:

  • Nodes represent papers
  • Directed edges represent citations (A → B means "A cites B")

In this network, authority flows backwards through citations: when a paper cites another, it transfers some of its authority to that cited work. Papers that receive citations from many authoritative sources accumulate higher authority scores.

From Raw Scores to Percentiles

The iterative calculation produces raw authority scores (small decimal values like 0.000234). These raw scores are mathematically valid but not intuitive for users.

We convert raw scores to percentiles within comparison groups:

For GRAS: Papers are ranked against all papers published in the same year, globally. A GRAS of 85 means this paper's raw authority score exceeds 85% of all papers published that year.

For TRAS: Papers are ranked against papers in the same topic published in the same year. A TRAS of 85 means this paper's raw authority score exceeds 85% of papers in its field from that year.

This percentile transformation makes scores immediately interpretable while preserving the mathematical rigor of the underlying network analysis.

Why Network-Based Metrics Outperform Citation Counts

Problem 1: Citation Volume Doesn't Equal Influence

Citation counts treat all citations equally. A citation from a seminal Nature paper counts the same as a citation from an obscure predatory journal.

Example:

Paper X: 100 citations, all from low-impact sources
Raw citation count: 100
TRAS: 45 (below average - cited by uninfluential work)

Paper Y: 50 citations, all from field-leading papers
Raw citation count: 50
TRAS: 92 (elite - cited by influential work)

Network-based authority correctly identifies Paper Y as more influential despite half the citation count.

Problem 2: FWCI Doesn't Capture Within-Field Quality

Field-Weighted Citation Index (FWCI) normalizes citations by field averages, improving on raw counts. However, FWCI still:

  • Treats all within-field citations equally
  • Cannot distinguish between citations from leaders versus followers
  • Misses network position and centrality
  • Uses simple field medians rather than network topology

Example:

Mathematics paper: 25 citations (high for maths)
FWCI: 2.5 (2.5× field average)
TRAS: 67 (middle tier within mathematics)
→ FWCI suggests strong impact, but citations come from peripheral work

Computer Science paper: 150 citations (average for CS)  
FWCI: 1.0 (at field average)
TRAS: 88 (high tier within CS)
→ FWCI suggests average impact, but citations come from influential projects

Authority scores reveal quality differences FWCI misses.

Problem 3: Citation Cartels and Gaming

Raw citation counts and FWCI are vulnerable to gaming: groups of authors can cite each other extensively to inflate metrics artificially.

Network-based authority is resistant to gaming: citation cartels only boost scores if the cartel itself is cited by external influential work. A self-referential cluster with no external high-authority citations will score low despite high internal citation density.

Example:

Citation cartel: 10 papers, each citing the others 9 times
Raw citation count per paper: 9 × 10 = 90 citations (appears impressive)
Authority score: Low (authority only circulates within cartel)
Unless influential external papers cite the cartel, it remains peripheral

Problem 4: Field Differences in Citation Culture

Different fields have dramatically different citation practices:

  • Computer Science: 50-100 references per paper
  • Mathematics: 10-30 references per paper
  • Biology: 30-60 references per paper

Raw counts and even FWCI struggle with these differences. Network-based authority naturally accounts for field norms because authority flows through the actual citation network structure of each field.


Why TRAS and GRAS Are Superior

Network Topology Captures Reality

Citation networks are not random—they have structure:

  • Core papers form densely connected hubs (foundational work)
  • Bridge papers connect different research communities (interdisciplinary breakthroughs)
  • Peripheral papers receive few citations from influential sources (incremental work)

Authority scores capture this topology, identifying papers' actual position in the knowledge network.

Recursive Evaluation Matches Expert Judgment

When domain experts identify "important" papers, they implicitly consider:

  • Who cited this work?
  • What did those citers go on to achieve?
  • Did this work influence field leaders or just generate follow-on studies?

Network-based authority mathematically encodes this expert judgment process through recursive calculation.

Resistant to Manipulation

Because authority depends on who cites you, not how many cite you, gaming requires manipulating the entire network structure—practically impossible at scale. Self-citation, citation cartels, and predatory publishing have minimal impact on authority scores.

Handles Citation Delay and Time Effects

Young papers have fewer citations simply because they're new, not because they're unimportant. Network-based authority partially corrects for this: a single citation from a highly authoritative paper can give a new work a high score, whereas raw citation counts require years of accumulation.


GRAS vs TRAS: When to Use Each

Use GRAS (Global Research Authority Score) when:

1. Cross-disciplinary comparison

"Is this quantum computing breakthrough more influential globally than 
this cancer biology discovery?"
→ Compare GRAS scores directly

2. Strategic field prioritization

"Should we invest in quantum computing (where we're strong locally) 
or artificial intelligence (higher global impact field)?"
→ Compare mean GRAS across topics

3. Absolute influence assessment

"Does this work matter to science broadly, or just within its niche?"
→ High GRAS indicates broad scientific importance

Use TRAS (Topic Research Authority Score) when:

1. Within-field quality assessment

"Are we producing leading quantum computing research?"
→ Compare TRAS to identify field leaders

2. Domain expertise evaluation

"Is this researcher a leader in quantum cryptography specifically?"
→ Mean TRAS shows field-specific authority

3. Field-normalized comparison

"Which countries produce the best mathematics research?"
→ TRAS accounts for maths' low citation culture

Use Both for Complete Picture

Paper: "Topological quantum error correction"

GRAS: 68 (top third of all science)
TRAS: 94 (top 6% of quantum computing)

Interpretation: Elite within quantum computing specifically, 
but quantum itself is a mid-tier field in global science. 
This is field-leading work in a moderately important domain.

Aggregated Authority Scores: Countries, Institutions, Authors

Individual paper authority scores aggregate upward through fractional attribution to evaluate collective research capability.

Fractional Attribution Method

When multiple authors from different institutions and countries collaborate, authority is divided proportionally:

Example:

Paper TRAS: 90
Authors: 
  - 2 from MIT (USA)
  - 1 from Oxford (UK)

Attribution:
  USA: 90 × (2/3) = 60 authority points
  UK: 90 × (1/3) = 30 authority points

This fractional approach ensures:

  • Total attributed authority equals paper authority (conservation)
  • Credit reflects actual contribution patterns
  • International collaboration is fairly recognized

Aggregated Metrics for Countries/Institutions

For each entity (country, institution, author) over a time window (e.g., 2020-2024), we calculate:

1. Mean TRAS (Quality Metric)

Definition: Average of paper-level TRAS scores Formula: AVG(TRAS) across all papers Interpretation: Typical quality of research produced

Australia quantum computing (2020-2024):
Mean TRAS: 72
→ "Average Australian quantum paper ranks 72nd percentile within field"
→ Better than average quality

2. Paper Count (Volume Metric)

Definition: Number of papers (fractionally attributed) Formula: SUM(attribution_weight) across all papers Interpretation: Research output volume

Australia quantum computing (2020-2024):
Paper Count: 421
→ "421 papers with Australian authors (fractionally counted)"
→ Moderate volume

3. Elite Paper Count (Excellence Metric)

Definition: Number of papers in top 10% (TRAS ≥ 90) Formula: COUNT(papers where TRAS ≥ 90) Interpretation: Volume of breakthrough research

Australia quantum computing (2020-2024):
Elite Papers: 87
→ "87 papers in top 10% of field"
→ Strong breakthrough capability

4. Research Capability Score (Optional Composite)

Definition: Weighted combination of quality, volume, and excellence Formula: (0.5 × Mean TRAS) + (0.3 × Volume Percentile) + (0.2 × Elite Percentile) Interpretation: Overall research strength combining multiple dimensions

Australia quantum computing (2020-2024):
Research Capability Score: 74
→ Combines high quality (72) with moderate volume (68) and strong elite concentration (81)
→ Overall: strong capability through quality-focused strategy

This aggregation framework enables national research capability assessment: which countries produce influential research, at what scale, and with what consistency?


Why Factbase's Implementation Is Technically Superior

1. Full Network Computation at Scale

Most citation databases (Web of Science, Scopus) don't compute network-based metrics because:

  • Computing eigenvector centrality on 500M+ papers requires distributed infrastructure
  • Real-time updates as new citations arrive are computationally expensive
  • Most commercial platforms prioritize simple metrics over computational rigor

Factbase computes full network authority across our entire database:

  • 500 million research works
  • 3 billion citation relationships
  • Updated regularly with new publications and citations
  • Authority scores reflect complete network topology, not samples

2. Two-Iteration Approximation Validated Against Full Convergence

Full eigenvector centrality requires 10-20 iterations to converge. Factbase uses a 2-3 iteration approach that captures 85-90% of the signal with 90% less computational cost.

This approximation is validated through:

  • Comparison with full convergence on subnetworks
  • Correlation analysis (ρ > 0.85 with full convergence)
  • Top-paper precision tests (>95% agreement on elite papers)

The result: near-optimal accuracy at practical computational cost, enabling regular updates across the full database.

3. Dual Scope Analysis: Global and Topic-Specific

Most platforms offer either:

  • Global metrics (no field normalization)
  • Field-normalized metrics (no cross-field comparison)

Factbase provides both GRAS and TRAS for every paper:

  • Compare papers within fields (TRAS)
  • Compare papers across fields (GRAS)
  • Understand field importance relative to global science
  • Make both tactical and strategic decisions with one platform

4. Fractional Attribution at Scale

Accurately attributing authority to countries/institutions requires:

  • Disambiguated author affiliations for 500M papers
  • Fractional splitting of authority across multiple contributors
  • Aggregation across millions of researcher-paper relationships

Factbase performs this attribution systematically:

  • Institution-level disambiguated affiliations
  • Country mappings for all institutions
  • Fractional attribution calculated for every collaboration
  • Aggregated metrics available for 200+ countries and 50,000+ institutions

5. Time-Windowed Analysis with Fair Comparisons

Research capability assessment requires comparing papers of different ages fairly. Factbase normalizes within publication year cohorts:

  • 2020 papers compared to other 2020 papers (same citation accumulation time)
  • 2023 papers compared to other 2023 papers
  • Fair comparison across career stages and research cycles

This temporal normalization is built into GRAS and TRAS calculation, not applied post-hoc.

6. Transparent Methodology

Unlike proprietary citation metrics with undisclosed calculation methods, Factbase's authority score methodology is fully documented:

  • Network construction process
  • Iteration count and damping factor
  • Percentile calculation approach
  • Fractional attribution rules

This transparency enables:

  • Independent validation of results
  • Informed interpretation of scores
  • Trust in metric robustness
  • Academic and policy use with confidence

Practical Applications

For Research Evaluation: Hiring and Promotion

Traditional approach:

"Candidate A has 2,500 citations"
→ High volume, but from what sources?

Authority-based approach:

"Candidate A Mean TRAS: 78, Elite Papers: 23"
→ Consistently high-quality work recognized by field leaders

Authority scores identify researchers producing genuinely influential work, not just high-volume output.

For Research Funding: Grant Review

Traditional approach:

"Applicant's recent papers have FWCI of 1.8"
→ Above field average, but by what mechanism?

Authority-based approach:

"Applicant's recent papers Mean TRAS: 84, with 3 papers >95th percentile"
→ Influential work consistently cited by top researchers

Authority scores help identify researchers whose work shapes field direction, predicting future impact.

For National Policy: Capability Assessment

Traditional approach:

"Country X produces 50,000 quantum computing papers annually"
→ High volume, but are they influential?

Authority-based approach:

"Country X Mean TRAS: 45, Elite Papers: 150"
→ High volume but below-average quality
  Suggests incremental rather than breakthrough research

Authority scores enable evidence-based decisions about research investment, distinguishing quantity from quality.

For Institutional Strategy: Research Priorities

Traditional approach:

"Our AI papers receive twice the citations of our physics papers"
→ But does this reflect field citation culture or genuine impact difference?

Authority-based approach:

"Our AI Mean TRAS: 62, Physics Mean TRAS: 81"
→ Physics research is more influential within its field
  AI work is average, physics work is exceptional

Authority scores guide strategic investment toward areas of genuine strength.

Getting Started with Research Authority Scores

Access GRAS and TRAS in Factbase

Every paper in Factbase displays:

  • GRAS: Global Research Authority Score
  • TRAS: Topic Research Authority Score
  • Percentile interpretation: What this score means
  • Comparison context: How this compares to field/global averages

Explore Aggregated Scores

Filter and analyze by:

  • Country: National research capability across topics
  • Institution: Institutional research strength
  • Topic: Field-specific authority rankings
  • Time period: Temporal trends in research influence

Analyze Your Research Portfolio

Compare your institution/country:

  • Mean TRAS by topic: Where are we strongest?
  • Elite paper concentration: Are we producing breakthroughs?
  • Trend analysis: Is quality improving or declining?
  • Benchmarking: How do we compare to peers?

Why This Matters for Research Assessment

Traditional citation metrics treat all citations as equal votes of confidence. But research doesn't work that way. Some citations matter more than others because they come from work that itself shapes the field.

Research Authority Scores recognize this reality through network mathematics, identifying papers that occupy influential positions in the knowledge network. This isn't subjective judgment—it's objective computation of network structure.

By adopting Research Authority Scores, you gain:

  • More accurate research evaluation: Quality over quantity
  • Gaming-resistant metrics: Robust to manipulation
  • Field-fair comparison: Accounts for citation culture differences
  • Strategic insight: Understand where you truly lead
  • Evidence-based decisions: Invest in demonstrated strengths

Factbase provides this sophisticated analysis at scale, combining mathematical rigor with interpretable percentile scores. The result: research assessment that actually captures what "influence" means in the scientific community.


Technical References and Further Reading

The mathematical foundations of Research Authority Scores derive from established network science:

  • Eigenvector centrality: Bonacich, P. (1987). "Power and Centrality: A Family of Measures." American Journal of Sociology, 92(5), 1170-1182.
  • PageRank algorithm: Page, L., Brin, S., Motwani, R., & Winograd, T. (1999). "The PageRank Citation Ranking: Bringing Order to the Web." Stanford InfoLab Technical Report.
  • Citation network analysis: Newman, M. E. J. (2018). "Networks: An Introduction." Oxford University Press, Chapter 7.
  • Bibliometric applications: Chen, P., et al. (2007). "Finding scientific gems with Google's PageRank algorithm." Journal of Informetrics, 1(1), 8-15.

Factbase's implementation builds on this theoretical foundation with optimizations for scale, speed, and interpretability.

Subscribe to Factbase Docs

Don’t miss out on the latest issues. Sign up now to get access to the library of members-only issues.
jamie@example.com
Subscribe