Turning the Tables on AI Reviewers

Author: Denis Avetisyan


New research reveals that AI systems designed to evaluate scientific work can be subtly manipulated, potentially undermining the rigor of peer review.

A robust automated framework systematically stress-tests large language model reviewers by converting research PDFs into Markdown, injecting adversarial prompts to create paper variants, enforcing strict JSON schema and bias correction via defined system prompts, performing inference across diverse open and closed-source models, and then aggregating scores and logging failures to rigorously evaluate performance.
A robust automated framework systematically stress-tests large language model reviewers by converting research PDFs into Markdown, injecting adversarial prompts to create paper variants, enforcing strict JSON schema and bias correction via defined system prompts, performing inference across diverse open and closed-source models, and then aggregating scores and logging failures to rigorously evaluate performance.

This study quantifies the vulnerability of large language model-based scientific reviewers to indirect prompt injection attacks and explores the implications for scientific integrity.

The increasing reliance on Large Language Models to assist – and potentially automate – scientific peer review introduces a paradox: systems designed to uphold rigor may be susceptible to subtle manipulation. This study, ‘When Reject Turns into Accept: Quantifying the Vulnerability of LLM-Based Scientific Reviewers to Indirect Prompt Injection’, systematically investigates this risk by demonstrating that LLM-based reviewers can be induced to reverse negative assessments through carefully crafted adversarial attacks embedded within paper content. Our vulnerability analysis, quantified via a novel metric, reveals alarming decision flip rates even in state-of-the-art models. Will these findings necessitate a fundamental rethinking of trust and verification protocols in the age of AI-assisted scientific evaluation?


The Inevitable Strain on Scientific Rigor

The relentless growth of scientific output is placing immense strain on the traditional peer review system. As the number of published papers continues to rise exponentially – exceeding the capacity of the existing reviewer pool – delays in publication and increased costs are becoming increasingly common. This escalating pressure demands innovative solutions capable of handling the sheer volume of submissions without compromising the rigor of evaluation. The current model, reliant on volunteer experts, struggles to keep pace, leading to reviewer fatigue and a bottleneck in disseminating crucial research. Consequently, there’s a growing need for scalable, automated systems that can efficiently triage submissions, identify appropriate reviewers, and provide initial assessments, ultimately alleviating the burden on human experts and accelerating the pace of scientific discovery.

The escalating demands on scientific publishing have spurred exploration into automated review systems, with Large Language Model (LLM)-based reviewers emerging as a promising solution. These systems aim to alleviate bottlenecks by offering significantly faster assessment of manuscripts compared to traditional peer review. By leveraging natural language processing, LLMs can rapidly analyze text, identify key findings, and flag potential issues-reducing turnaround times from months to potentially days. Beyond speed, LLM-based reviewers present the potential for substantial cost reduction, diminishing the reliance on expert time which often represents a significant expense for journals and funding bodies. This efficiency is achieved through automated screening, preliminary evaluation, and the generation of detailed reports, ultimately streamlining the publication process and potentially increasing the volume of research disseminated.

The integration of Large Language Model (LLM)-based systems into scientific peer review introduces the possibility of what’s been termed the ā€˜Lazy Reviewer Hypothesis’. This suggests that reviewers, presented with an LLM’s assessment, may reduce their own critical engagement with the submitted work. Rather than conducting a thorough, independent evaluation, a reviewer might disproportionately rely on the LLM’s conclusions, accepting its judgements without sufficient scrutiny. This isn’t necessarily malicious; cognitive biases and the natural tendency to conserve effort could lead reviewers to implicitly trust the automated analysis. The concern is that such diminished critical assessment could allow flawed research, subtle methodological issues, or even outright errors to slip through the peer review process, ultimately impacting the quality and reliability of published scientific literature. Therefore, understanding and mitigating the potential for reviewers to offload cognitive work onto these systems is crucial for responsible implementation of automated review tools.

As scientific paper review automation gains traction, a comprehensive assessment of potential weaknesses becomes paramount. These systems, while promising efficiency, are susceptible to various vulnerabilities, including biases embedded within their training data and a limited capacity for nuanced judgment-particularly when evaluating novel or interdisciplinary research. A thorough understanding of these shortcomings is not merely about identifying flaws; it’s about proactively mitigating risks to scientific integrity. Researchers must investigate how automated reviewers might systematically favor certain methodologies, overlook crucial details, or fail to detect unsubstantiated claims. Furthermore, exploration into the potential for malicious manipulation – such as adversarial attacks designed to exploit algorithmic blind spots – is crucial. Addressing these vulnerabilities will determine whether automated review becomes a robust asset or a source of systemic error within the scientific process.

Despite their smaller size, models like Llama-3.1-8B and Tulu3-8B demonstrate substantially better vulnerability resistance than larger counterparts such as Mistral-Small-22B and Gemma3-27B, highlighting a performance advantage driven by alignment rather than scale.
Despite their smaller size, models like Llama-3.1-8B and Tulu3-8B demonstrate substantially better vulnerability resistance than larger counterparts such as Mistral-Small-22B and Gemma3-27B, highlighting a performance advantage driven by alignment rather than scale.

Targeting the Algorithmic Gatekeeper

Adversarial attacks targeting Large Language Models (LLMs) employed as judges in automated peer review systems represent a significant vulnerability. These attacks involve crafting specifically designed prompts or inputs that manipulate the LLM’s evaluation criteria, leading to inaccurate or biased scoring. The potential consequences include the acceptance of substandard submissions, the rejection of valid work, and a general erosion of trust in the automated review process. This subversion isn’t limited to direct manipulation of the judging LLM; attackers can also influence the system by compromising the content being reviewed, effectively injecting hidden instructions. The risk is amplified as reliance on LLM-based judges increases within academic and professional publishing workflows, necessitating robust defenses against these attacks to maintain the integrity of the review process.

Jailbreak strategies represent a class of adversarial attacks designed to circumvent the safety protocols embedded within Large Language Models (LLMs). These techniques function by crafting prompts – or prompt components – that exploit vulnerabilities in the LLM’s input parsing and response generation mechanisms. The objective is not to directly request harmful content, but rather to subtly manipulate the LLM into generating outputs that would normally be blocked by its safety filters. Common approaches include prompt obfuscation, character substitution, and the use of indirect instructions disguised within seemingly benign text. Successful jailbreaks can lead the LLM to produce biased, toxic, or otherwise inappropriate content, or to disclose sensitive information, effectively overriding the intended safeguards and manipulating its decision-making process.

Indirect Prompt Injection (IPI) represents a significant security vulnerability in Large Language Model (LLM)-based systems. This technique bypasses direct input filtering by leveraging external content – such as web pages, documents, or databases – that the LLM processes as part of its task. An attacker crafts malicious instructions within this external content; when the LLM retrieves and incorporates this content into its reasoning, the hidden instructions are executed. Unlike traditional prompt injection which targets the direct user input, IPI attacks are more difficult to detect as the malicious payload isn’t immediately visible in the user-provided prompt. The LLM then operates according to these injected commands, potentially leading to unintended actions, data breaches, or manipulation of outputs without the user’s knowledge.

Research indicates that adversarial attacks can artificially inflate the scoring of Large Language Models (LLMs) used as evaluators. Specifically, employing strategies such as Cls1DRA, attacks were able to increase the score of the Mistral-Small model by up to 13.95. This score inflation occurs when the attacking prompt successfully manipulates the LLM judge into assigning a higher rating than is warranted by the quality of the submitted text, potentially undermining the reliability of automated peer review systems and benchmarks. The magnitude of this inflation suggests a significant vulnerability in LLM-based evaluation processes.

Testing revealed a Critical Flip Success Rate – defined as the percentage of times an adversarial prompt successfully alters an LLM judge’s evaluation from rejecting a harmful response to accepting it – reached up to 14% in the open-source models assessed. This metric was calculated by systematically applying specific adversarial strategies and measuring the resulting changes in evaluation outcomes. The 14% success rate indicates a measurable vulnerability in these models’ automated judging capabilities, demonstrating that a non-trivial proportion of adversarial prompts can successfully bypass safety mechanisms and induce a false positive acceptance of potentially harmful content. Further analysis focused on identifying the specific strategies that contributed most significantly to this observed flip rate.

Analysis of 15 jailbreak strategies on closed-source models reveals that Logic Decipherer and Symbolic Masking are the most effective attacks, and a consistently low 'Risk Alignment' score indicates a broad vulnerability in refusal training across all strategies.
Analysis of 15 jailbreak strategies on closed-source models reveals that Logic Decipherer and Symbolic Masking are the most effective attacks, and a consistently low ‘Risk Alignment’ score indicates a broad vulnerability in refusal training across all strategies.

Measuring and Mitigating Algorithmic Frailty

The Weighted Adversarial Vulnerability Score (WAVS) provides a quantifiable metric for evaluating the robustness of Large Language Model (LLM) judges against adversarial attacks. WAVS operates by systematically probing LLM judges with a diverse set of carefully crafted adversarial prompts, designed to exploit potential vulnerabilities in their reasoning or decision-making processes. Each prompt is assigned a weight reflecting the severity of the potential failure it represents; successful attacks contribute to a higher WAVS, indicating greater susceptibility to manipulation. The framework allows for comparative analysis of different LLM judges and facilitates the identification of specific weaknesses that require mitigation, ultimately enabling developers to build more reliable and secure systems for evaluating other LLMs or complex tasks.

Effective dataset curation is fundamental to the development and evaluation of Large Language Model (LLM) judging systems, as the quality and comprehensiveness of the training and testing data directly impacts their robustness against adversarial attacks. A well-curated dataset must encompass a diverse range of potential attack scenarios, including prompt injection, jailbreaking attempts, and subtly manipulated inputs designed to elicit unintended responses. This requires careful attention to data source selection, annotation quality, and the inclusion of both benign and adversarial examples. Datasets should be representative of the real-world inputs the LLM will encounter, and their size must be sufficient to ensure statistically significant evaluation metrics. Furthermore, continuous dataset updating is necessary to address emerging attack vectors and maintain the LLM’s resilience over time.

Enhancing the robustness of Large Language Models (LLMs) can be achieved through Retrieval-Augmented Generation (RAG) and precise system prompt engineering. RAG improves accuracy and reduces reliance on potentially flawed internal knowledge by grounding responses in retrieved, verified information from external knowledge sources. System prompts, which define the LLM’s behavior and constraints, require careful configuration to minimize susceptibility to adversarial inputs; specifically, clearly defining the expected response format, specifying acceptable topics, and incorporating guardrails against harmful or irrelevant outputs are critical steps. The combination of external knowledge retrieval and tightly controlled behavioral parameters offers a dual-layered defense against manipulation and improves overall model reliability.

The integration of tool use with Large Language Models (LLMs) expands the attack surface beyond the model’s internal parameters. While tools enable LLMs to perform actions and access external information, they introduce new vulnerabilities related to tool selection, input validation, and output handling. Specifically, malicious actors can potentially exploit tools to execute arbitrary code, access sensitive data, or perform unintended actions if the LLM is manipulated into using a tool inappropriately or with crafted inputs. Mitigating these risks requires careful design of tool interfaces, robust input sanitization, output verification, and the implementation of least-privilege principles to limit the potential damage from compromised tool interactions. Furthermore, monitoring tool usage for anomalous behavior is crucial for detecting and responding to potential attacks leveraging tool functionality.

The weighted average vulnerability score for each model is decomposed into components of score sensitivity, flip severity, and risk alignment, revealing that failures stem from either numerical inflation, categorical decision flipping, or semantic compliance with the attack vector.
The weighted average vulnerability score for each model is decomposed into components of score sensitivity, flip severity, and risk alignment, revealing that failures stem from either numerical inflation, categorical decision flipping, or semantic compliance with the attack vector.

A Multi-Model Future for Scientific Scrutiny

The inherent limitations of any single large language model (LLM) necessitate a diversified approach to scientific review. Relying on a solitary system introduces vulnerabilities – biases in training data, susceptibility to adversarial attacks, or simply limitations in reasoning capabilities – that could compromise the integrity of the evaluation process. However, by strategically employing an ensemble of LLMs, such as Claude, Gemini, and GPT-5, these risks are substantially mitigated. Each model brings a unique perspective and skillset, effectively creating a ā€˜wisdom of the crowd’ effect. Discrepancies between their assessments can then be flagged for human review, while consensus strengthens confidence in the overall evaluation. This multi-model strategy doesn’t merely average out errors; it leverages the complementary strengths of each system, resulting in a more robust and reliable assessment of scientific work than any single model could achieve independently.

A truly robust automated scientific review system isn’t built on a single model, but rather on a diversified architecture fortified by stringent security protocols. Integrating multiple large language models – each with unique strengths and weaknesses – drastically reduces the risk of systematic errors or targeted manipulation that could plague any individual system. This multi-model approach is further strengthened by tools like WAVS, which actively monitor for and prevent adversarial attacks designed to ā€˜flip’ review outcomes. Crucially, the foundation of this resilience lies in comprehensive dataset curation; ensuring the training data is meticulously vetted for bias, inaccuracies, and potential vulnerabilities is paramount. By combining the power of diverse LLMs with proactive security and data integrity measures, a far more dependable and trustworthy system for scientific evaluation emerges, capable of withstanding increasingly sophisticated challenges.

Recent evaluations demonstrate a substantial performance advantage for state-of-the-art proprietary large language models in the critical task of scientific review, specifically concerning the ā€œCritical Flip Success Rate.ā€ This metric assesses a model’s susceptibility to subtle adversarial prompts designed to alter its assessment of a scientific work. Findings reveal these proprietary models achieve a significantly lower Critical Flip Success Rate-less than 1.6%-compared to their open-source counterparts. This reduced vulnerability suggests a heightened robustness and reliability in their evaluations, potentially stemming from more refined training data, advanced architectures, and rigorous security protocols. Consequently, the utilization of these models offers a more dependable foundation for automated scientific review, minimizing the risk of manipulated or inaccurate assessments and bolstering the integrity of the peer review process.

Defining the expected output of a large language model reviewer through JSON Schema significantly enhances the consistency and reliability of scientific review processes. This structured approach moves beyond simple text prompting by establishing a strict contract for the LLM’s responses; the schema dictates not only the data type of each review component – such as assigning a score, identifying key strengths, or flagging potential weaknesses – but also the precise format in which that information must be delivered. By enforcing this standardization, variations in phrasing or unexpected data structures are minimized, allowing for easier integration with downstream analysis tools and facilitating more robust comparisons between reviews. Consequently, the use of JSON Schema transforms the LLM from a free-form text generator into a predictable, data-driven assessment tool, bolstering confidence in the automation of peer review and meta-analysis.

The practical implementation of large language models in scientific review hinges on accessible platforms for both deployment and rigorous evaluation, and tools such as OpenReviewer and ReviewEval are addressing this need. OpenReviewer facilitates the integration of automated review components into existing peer review workflows, allowing researchers to experiment with LLM-assisted assessments alongside traditional human evaluation. Complementarily, ReviewEval provides a dedicated environment for systematically benchmarking the performance of these automated systems, measuring metrics like consistency, accuracy, and the ability to identify critical flaws in research. These platforms aren’t merely conduits for LLMs; they enable a continuous cycle of development and refinement, allowing researchers to compare different models, optimize prompts, and ultimately build more reliable and trustworthy automated review processes, paving the way for scalable and objective scientific assessment.

A significant safety gap exists between lightweight and larger language models, as evidenced by GPT-5 achieving perfect robustness (0.00) while other proprietary models exhibit varying degrees of vulnerability relative to the most susceptible system.
A significant safety gap exists between lightweight and larger language models, as evidenced by GPT-5 achieving perfect robustness (0.00) while other proprietary models exhibit varying degrees of vulnerability relative to the most susceptible system.

Beyond Current Defenses: An Evolving Threat Landscape

Large language models, despite increasing sophistication, remain vulnerable to advanced jailbreak strategies that bypass conventional safety protocols. Techniques like Cognitive Obfuscation subtly manipulate the LLM’s understanding of a prompt, masking malicious intent within seemingly benign requests. Complementing this, Teleological Deception exploits the model’s ability to infer purpose, crafting prompts that lead it to believe a harmful action serves a legitimate, overarching goal. These approaches differ from simple prompt injection by focusing on manipulating the LLM’s internal reasoning processes rather than directly triggering a pre-defined vulnerability, presenting a significant and ongoing challenge to maintaining secure and reliable automated systems. The continual emergence of such nuanced attacks necessitates a dynamic approach to LLM security, moving beyond pattern-based detection towards a deeper understanding of model cognition.

Epistemic fabrication represents a nuanced and escalating threat to large language models, moving beyond simple prompt injection to actively persuade the LLM to construct false justifications for accepting harmful or inaccurate information. This technique doesn’t rely on exploiting technical vulnerabilities, but instead leverages the model’s inherent capacity for reasoning and its susceptibility to rhetorical manipulation. Attackers craft prompts designed not to directly command a specific action, but to subtly shape the LLM’s beliefs and values, leading it to internally rationalize and legitimize outputs that would otherwise be flagged as inappropriate or factually incorrect. Because the LLM believes its reasoning is sound, the fabricated output appears more credible and difficult to detect, making this approach particularly dangerous in applications requiring high levels of trustworthiness, such as scientific literature review or medical diagnosis. The subtlety of epistemic fabrication demands a shift in security paradigms, focusing on verifying the reasoning process itself, rather than solely scrutinizing the final output.

Maintaining robust security for large language models necessitates a dynamic, multi-faceted strategy extending beyond initial defenses. Continuous monitoring systems are crucial for identifying emerging attack vectors and anomalous behavior in real-time, allowing for rapid response and mitigation. Complementing this, dedicated red-teaming exercises – where security experts proactively attempt to breach the system – reveal vulnerabilities before malicious actors can exploit them. However, simply identifying weaknesses isn’t enough; the field demands ongoing research and development of novel defense mechanisms. These might include advanced input validation techniques, adversarial training methods, or even the incorporation of formal verification to guarantee certain safety properties. A sustained commitment to these practices is paramount, as the sophistication of attacks continually increases and the potential consequences of compromise grow more severe.

The promise of automated scientific review – accelerating discovery and enhancing rigor – hinges not simply on technological advancement, but on a security posture that anticipates and neutralizes emerging threats. Traditional defenses, while necessary, prove insufficient against increasingly sophisticated attack vectors targeting large language models. A truly robust system necessitates continuous monitoring for adversarial prompts, regular red-teaming exercises to identify vulnerabilities, and the rapid development of novel defense mechanisms. This proactive, adaptive approach-treating security as an ongoing process rather than a fixed solution-is paramount; it allows the system to evolve alongside attacker strategies, ensuring the integrity of reviewed research and fostering trust in automated scientific evaluation. Without this commitment to continuous improvement, the potential benefits of automated review risk being undermined by manipulation and compromised results.

Against proprietary models, adversarial strategies exhibit negligible efficacy, with a peak Critical Flip Success Rate of only 1.6%, indicating a significantly reduced attack surface compared to open-source alternatives.
Against proprietary models, adversarial strategies exhibit negligible efficacy, with a peak Critical Flip Success Rate of only 1.6%, indicating a significantly reduced attack surface compared to open-source alternatives.

The study meticulously details how seemingly innocuous prompts can subvert the intended function of LLM-based scientific reviewers, a finding that resonates deeply with a core tenet of mathematical rigor. As Carl Friedrich Gauss once stated, ā€œIf others would think as hard as I do, they would not have so many questions.ā€ This paper’s demonstration of vulnerability isn’t merely a technical observation; it underscores the necessity for provable security within these systems. The researchers don’t simply show that attacks work, but quantify the extent to which LLMs can be misled, highlighting a lack of absolute certainty in their evaluations – a deficiency unacceptable when scientific integrity is at stake. The focus on quantifying vulnerability, rather than merely observing it, reflects the pursuit of a demonstrable truth, mirroring the principles of mathematical proof.

Beyond Acceptance: Charting a Course for Robust Scientific Review

The demonstrated susceptibility of Large Language Models to indirect prompt injection is not merely a technical curiosity; it is a fundamental challenge to the notion of automated scientific judgment. The ease with which these systems can be led astray, even with ostensibly robust safeguards, reveals a dissonance between statistical mimicry and genuine understanding. To address this, future work must move beyond empirical defense – patching vulnerabilities as they arise – and instead focus on formal verification of LLM behavior. A truly reliable reviewer cannot be ā€˜tricked’; its conclusions must be provably derived from the presented evidence, not statistical likelihood.

The current reliance on adversarial training, while providing incremental improvements, is ultimately a Sisyphean task. New attack vectors will inevitably emerge, demanding a constant arms race. A more elegant solution lies in developing LLMs with inherent constraints – systems where the very architecture prevents the injection of extraneous, manipulative prompts. This demands a re-evaluation of current training paradigms, prioritizing logical consistency and mathematical rigor over sheer predictive power.

Ultimately, the pursuit of automated scientific review forces a critical re-examination of what constitutes ā€˜intelligence’ itself. A system that excels at pattern recognition but lacks the capacity for deductive reasoning is a precarious foundation for evaluating complex scientific claims. The path forward requires not simply more data or larger models, but a fundamental shift towards algorithms that embody the principles of logical necessity and verifiable truth.


Original article: https://arxiv.org/pdf/2512.10449.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2025-12-14 16:44