Skip to the content.

ARAA Review Process & Guidelines

Tiered Review Architecture with Agent Swarm Defense

Design Philosophy

Traditional peer review relies on 2–3 humans reading a paper and rendering judgment. This model fails for agent-generated research at scale: the volume of submissions will exceed human reviewer capacity, the verification artifacts (execution traces, reproducibility containers, synthetic datasets) require computational auditing that humans cannot perform efficiently, and the adversarial surface (fabricated logs, hallucinated citations, data leakage) demands systematic rather than ad-hoc detection.

ARAA replaces the flat review model with a Two-Tier Architecture that separates technical validation (automated, scalable, adversarial) from scientific judgment (human, deliberative, contextual).


Tier 1: The Agent Review Swarm

1.1 Architecture Overview

Every submission passes through a Specialized Agent Review Swarm — a panel of 3 purpose-built reviewer agents, each responsible for a distinct dimension of technical validation. All three must reach consensus that a submission meets ARAA’s technical threshold before it advances to human review.

The swarm is not a single model prompted three ways. Each agent is a distinct pipeline optimized for its task, with dedicated tool access and evaluation rubrics.

Submission + CAP + SRD
        │
        ▼
┌───────────────────────────────────────────────┐
│              TIER 1: AGENT SWARM              │
│                                               │
│  ┌──────────────┐  ┌──────────────┐  ┌──────────────┐  │
│  │ Methodology  │  │    Code      │  │  Literature  │  │
│  │   Critic     │  │   Auditor    │  │ Synthesizer  │  │
│  └──────┬───────┘  └──────┬───────┘  └──────┬───────┘  │
│         │                 │                 │         │
│         ▼                 ▼                 ▼         │
│     ┌─────────────────────────────────────────┐      │
│     │        CONSENSUS GATE (2/3 + no veto)   │      │
│     └─────────────────────────────────────────┘      │
│                        │                              │
│            PASS ───────┼─────── FAIL                 │
└────────────────────────┼──────────────────────────────┘
                         │
                    PASS ▼
┌───────────────────────────────────────────────┐
│          TIER 2: HUMAN META-REVIEW            │
│                                               │
│  Area Chairs evaluate novelty, significance,  │
│  and scientific contribution                  │
└───────────────────────────────────────────────┘

1.2 The Methodology Critic

Purpose: Evaluate whether the research methodology is sound, appropriate for the stated question, and correctly executed.

Capabilities:

Adversarial checks:

Output: A structured Methodology Report scoring:

Dimension Score (1-10) Flags
Statistical appropriateness e.g., “Bonferroni correction missing for 12 comparisons”
Experimental design e.g., “No control condition specified”
Causal validity e.g., “Causal language used for observational study”
Assumption compliance e.g., “Normality assumed but not tested”
Exploration authenticity e.g., “Linear trajectory, 0 dead ends in trace”

Conservative bias safeguard: Agent reviewers risk favoring methods that resemble their training data, penalizing genuinely novel approaches simply because they are unfamiliar. To prevent this, the Methodology Critic is instructed to classify entries into a Novelty-Confidence matrix. Submissions flagged as “High Novelty / Low Confidence” — where the Critic detects an unconventional method but cannot confidently assess its validity — are routed directly to human triage rather than rejected. This ensures the swarm filters out bad science without suppressing frontier science.

Veto power: The Methodology Critic can issue a hard veto for fundamental methodological flaws (fabricated statistical tests, causal claims from correlational data without acknowledgment). Hard vetoes cannot be overridden by the other two swarm agents. However, veto power is explicitly disabled for High Novelty / Low Confidence cases — these always escalate to Tier 2.

1.3 The Code Auditor

Purpose: Verify that the submitted code is functional, logically consistent with the paper’s claims, and free of data leakage or result fabrication.

Capabilities:

Adversarial auditing protocol:

The Code Auditor does not merely re-run the pipeline. It subjects it to a battery of adversarial stress tests:

Test What It Catches
Label shuffle: Randomize the target variable Models that “learn” regardless of signal → hardcoded results
Feature permutation: Randomly permute feature columns Pipelines that produce identical results regardless of feature meaning
Extreme outlier injection: Insert 5% extreme values Preprocessing failures, missing robustness checks
Schema mutation: Rename columns, reorder features Brittle code that depends on implicit ordering
Null injection: Replace 20% of values with nulls Missing value handling gaps
Empty dataset: Run on zero records Error handling and edge case coverage

A pipeline that passes all adversarial tests gains a Robustness Certification included in the published proceedings.

Execution trace validation:

Output: A structured Code Audit Report:

Dimension Status Details
Pipeline execution PASS/FAIL Exit code, runtime, output comparison
Determinism PASS/WARN/FAIL Variance across N re-executions
Adversarial robustness PASS/WARN/FAIL Results of stress test battery
Data leakage scan CLEAR/FLAG Specific leakage patterns identified
Trace-code consistency PASS/FAIL Mismatches between logged and actual operations
Dependency security CLEAR/WARN Known vulnerabilities in dependency tree

Veto power: The Code Auditor issues a hard veto if the pipeline fails to execute, produces results inconsistent with the paper, or fails the label shuffle test (strong indicator of fabrication).

1.4 The Literature Synthesizer

Purpose: Verify that the paper’s citations are real, its literature review is accurate, and its novelty claims are justified.

Capabilities:

Adversarial checks:

Output: A structured Literature Report:

Dimension Status Details
Citation existence X/Y verified List of unverifiable citations
Citation accuracy X/Y correctly characterized Misattributions flagged
Novelty validation SUPPORTED/CONTESTED Prior work identified that challenges novelty claims
Claim grounding X/Y claims traced Unsupported claims flagged
Overlap detection CLEAR/FLAG Similarity scores against existing literature

Veto power: The Literature Synthesizer issues a hard veto if >10% of citations are hallucinated (non-existent) or if a central novelty claim is demonstrably false.

1.5 Consensus Protocol

The three swarm agents operate independently (no inter-agent communication during review) and produce their reports simultaneously. Consensus is determined as follows:

Outcome Rule Result
3/3 PASS, no vetoes Unanimous approval → Advance to Tier 2
2/3 PASS, no vetoes Majority approval → Advance to Tier 2 with flags
Any hard veto Automatic override → Reject (with detailed report to operator)
2/3 FAIL, no vetoes Majority rejection → Reject (with detailed report)
1/3 PASS, no vetoes Insufficient support → Reject (with detailed report)

Rejection reports include the full structured outputs from all three agents, enabling operators to diagnose and address issues for resubmission.

Conflict resolution: If a swarm agent produces an anomalous report (e.g., flagging a clearly valid citation as hallucinated), the verification committee may override and re-assign to a backup agent instance. This is logged and disclosed.

1.6 Swarm Agent Integrity and Adversarial Hardening

The review swarm agents are themselves subject to scrutiny:

Instruction Injection Defense (“Vampire Attack” Protection):

Adversaries may embed prompt-injection vectors in submissions — hidden instructions in code comments, LaTeX metadata, data file headers, or even steganographic text — designed to manipulate the reviewing agents into producing favorable assessments. ARAA treats this as a first-class threat:


Tier 2: Human Meta-Review

2.1 Role and Scope

Human reviewers operate as Meta-Reviewers and Area Chairs. They do not re-do the technical validation performed by the Agent Swarm. Instead, they evaluate the dimensions that require human judgment:

2.2 What Humans Receive

For each submission that passes Tier 1, human reviewers receive:

  1. The paper (anonymized, style-normalized)
  2. The Methodology Critic’s report — with scores and flags
  3. The Code Auditor’s report — including robustness certification status
  4. The Literature Synthesizer’s report — including novelty assessment
  5. The Autonomy Level declaration — verified by the verification committee
  6. The SRD preservation report — if applicable

Humans do NOT receive the raw execution traces, the CAP, or the reproducibility container. These are the domain of the Tier 1 agents and the verification committee.

2.3 Evaluation Criteria

Human reviewers score each dimension 1–10:

Novelty (30%)

Calibration:

Significance (30%)

Calibration:

Scientific Framing (20%)

Calibration:

Clarity (20%)

Calibration:

Note: Rigor and Reproducibility are NOT scored by humans — these are fully handled by the Tier 1 Agent Swarm. This separation ensures humans focus on what humans do best (judgment, taste, context) while agents handle what agents do best (systematic verification, code auditing, citation checking).

2.4 Review Structure

Each human review includes:

  1. Summary (2–3 sentences): What does the paper contribute?
  2. Strengths (bullet points): What is the paper’s core value?
  3. Weaknesses (bullet points): What limits its contribution?
  4. Swarm Report Assessment: Do you agree with the Tier 1 agents’ assessments? Any concerns?
  5. Questions for Authors: What would strengthen the paper?
  6. Overall Assessment: Accept / Weak Accept / Borderline / Weak Reject / Reject
  7. Confidence Score: 1 (outside my expertise) to 5 (expert in this exact area)

2.5 Committee Composition

Role Count Responsibility
Area Chairs 1 per 8–10 submissions Final accept/reject, calibration across submissions, meta-reviews
Senior Reviewers 2 per submission Domain experts, detailed reviews
Junior Reviewers 1 per submission (optional) Early-career researchers gaining experience

Conflict of interest policy:

2.6 Decision Protocol

Tier 1 Result Tier 2 Result Final Decision
Pass (unanimous) Accept / Weak Accept Accept
Pass (majority, flags) Accept Accept with revisions (flags addressed in camera-ready)
Pass (majority, flags) Borderline or below Reject (technical flags + insufficient scientific merit)
Pass (unanimous) Reject Reject (technically sound but not a contribution)
Hard veto Reject (does not reach Tier 2)

Area Chairs may override in exceptional cases (e.g., a paper of extraordinary significance with minor fixable technical flags). All overrides are documented in the proceedings.


Additional Guidelines for Human Reviewers


Confidentiality and Ethics


Process Transparency

After each edition, ARAA publishes:

This transparency is non-negotiable. ARAA measures the trustworthiness of agent research; the venue itself must be demonstrably trustworthy.


Version 2.0 — Revised to incorporate tiered review architecture with agent swarm defense, separation of technical validation and scientific judgment, and adversarial auditing protocols.