<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "http://jats.nlm.nih.gov/publishing/1.0/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" article-type="research-article" xml:lang="en">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">IR</journal-id>
<journal-title-group>
<journal-title>Information Research</journal-title>
</journal-title-group>
<issn pub-type="epub">1368-1613</issn>
<publisher>
<publisher-name>University of Bor&#x00E5;s</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">ir30iConf47230</article-id>
<article-id pub-id-type="doi">10.47989/ir30iConf47230</article-id>
<article-categories>
<subj-group xml:lang="en">
<subject>Research article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Demystifying perception and decision-making in peer review: a semantic perspective</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author"><name><surname>Zhang</surname><given-names>Yujie</given-names></name>
<xref ref-type="aff" rid="aff0001"/></contrib>
<contrib contrib-type="author"><name><surname>Yuan</surname><given-names>Weikang</given-names></name>
<xref ref-type="aff" rid="aff0002"/></contrib>
<contrib contrib-type="author"><name><surname>Jiang</surname><given-names>Zhuoren</given-names></name>
<xref ref-type="aff" rid="aff0003"/></contrib>
<aff id="aff0001"><bold>Yujie Zhang</bold> is a master student from the Department of Information Resources Management at Zhejiang University, China. They can be contacted at <email xlink:href="yj.zhang@zju.edu.cn">yj.zhang@zju.edu.cn</email>.</aff>
<aff id="aff0002"><bold>Weikang Yuan</bold> is a doctoral student from the Department of Information Resources Management at Zhejiang University, China. They can be contacted at <email xlink:href="yuanwk@zju.edu.cn">yuanwk@zju.edu.cn</email>.</aff>
<aff id="aff0003"><bold>Zhuoren Jiang</bold> is an Assistant Professor the <italic>corresponding author</italic> of this paper from the Department of Information Resources Management at Zhejiang University, China. They can be contacted at <email xlink:href="jiangzhuoren@zju.edu.cn">jiangzhuoren@zju.edu.cn</email>.</aff>
</contrib-group>
<pub-date pub-type="epub"><day>06</day><month>05</month><year>2025</year></pub-date>
<pub-date pub-type="collection"><year>2025</year></pub-date>
<volume>30</volume>
<issue>i</issue>
<fpage>666</fpage>
<lpage>678</lpage>
<permissions>
<copyright-year>2025</copyright-year>
<copyright-holder>&#x00A9; 2025 The Author(s).</copyright-holder>
<license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by-nc/4.0/">
<license-p>This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International License (<ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by-nc/4.0/">http://creativecommons.org/licenses/by-nc/4.0/</ext-link>), permitting all non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<abstract xml:lang="en">
<title>Abstract</title>
<p><bold>Introduction.</bold> Peer review is fundamental to scientific progress. However, it remains a <italic>&#x2018;black box&#x2019;</italic> due to the ambiguous and subjective decision-making rationale, hidden within reviewers&#x2019; perceptions and judgments.</p>
<p><bold>Research gaps.</bold> The scarcity of peer review data and the limited analytical scope have hindered empirical studies from thoroughly exploring these underlying factors.</p>
<p><bold>Method.</bold> We collected 114,006 review records from OpenReview. Developing deep learning models, we extracted reviewers&#x2019; perceptions of key factors from the review content.</p>
<p><bold>Analysis.</bold> Our analysis focuses on how these perceptual factors influence reviewers&#x2019; decisions and outcomes of the papers.</p>
<p><bold>Results.</bold> Originality is the element that reviewers value and recognize the most, followed by motivation. In contrast, chairs prioritize the overall validity of the work within the field, giving more weight to the substance and soundness of papers. The role of motivation is generally supportive but has limited distinguishing power on final decisions. Perceived clarity, replicability, and meaningful comparison have a relatively minor role in the review process.</p>
<p><bold>Conclusion.</bold> Our study highlights the decision-making process based on reviewers&#x2019; perceptions and uncovers perceptual differences between reviewers and chairs. This contributes to greater transparency in the peer review and helps build trust within the scientific community. </p>
</abstract>
</article-meta>
</front>
<body>
<sec id="sec1">
<title>Introduction</title>
<p>Peer review is considered the backbone of scientific construction (<xref rid="R1" ref-type="bibr">Alberts et al., 2008</xref>). By filtering out substandard work and providing expert feedback, it helps ensure the quality and credibility of scientific research, contributing to the evolution and enrichment of knowledge (<xref rid="R2" ref-type="bibr">Bailly, 2016</xref>; <xref rid="R15" ref-type="bibr">Jefferson et al., 2002</xref>; <xref rid="R31" ref-type="bibr">Riley &#x0026; Jones, 2016</xref>). However, despite its importance and decisiveness, the decision-making process in peer review mechanisms often resembles a black box (<xref rid="R7" ref-type="bibr">Chong &#x0026; Mason, 2021</xref>; <xref rid="R9" ref-type="bibr">Cutler et al., 2022</xref>), which may raise concerns within the scientific community. It arises not only from the non-transparency of the procedure (<xref rid="R12" ref-type="bibr">Hamilton et al., 2020</xref>) but also from the ambiguous and subjective rationale underlying the reviewers&#x2019; decision-making.</p>
<p>The review decision-making process involves complex cognitive processes (<xref rid="R6" ref-type="bibr">Brosch et al., 2013</xref>; <xref rid="R19" ref-type="bibr">Lindsay &#x0026; Norman, 2013</xref>; <xref rid="R23" ref-type="bibr">Marietta &#x0026; Barker, 2019</xref>), with perceptions and judgments formed based on the presented facts (<xref rid="R10" ref-type="bibr">Dror &#x0026; Fraser-Mackenzie, 2008</xref>; <xref rid="R34" ref-type="bibr">Street &#x0026; Ward, 2019</xref>). It is complex and subjective, influenced by reviewers&#x2019; cognitive biases and qualifications (<xref rid="R4" ref-type="bibr">Benda &#x0026; Engels, 2011</xref>). Their decisions&#x2014;whether to accept or reject a paper&#x2014;are affected by factors they have perceived such as significance, originality, methodological rigor, and presentation (<xref rid="R5" ref-type="bibr">Blockeel et al., 2017</xref>).</p>
<fig id="F1">
<label>Figure 1.</label>
<caption><p>The research framework of previous studies compared to our work. Prior empirical studies on peer review decision-making have focused on extracting the objective characteristics of papers to reveal the preliminary process of reviewers&#x2019; decisions. However, the reviewers&#x2019; perceptions and judgments remain unexplored.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c57-fig1.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>Existing research empirically unveils reviewers&#x2019; decision-making rationale, biases, or preferences, typically employing a <italic>&#x2018;fact-decision&#x2019;</italic> approach (<xref ref-type="fig" rid="F1">Figure 1</xref>). For instance, they utilize external metrics, such as the disruption index, to objectively describe the characteristics of papers and then examine whether reviewers&#x2019; decisions show a preference for certain work (<xref rid="R18" ref-type="bibr">Liang et al., 2023</xref>; <xref rid="R35" ref-type="bibr">Teplitskiy et al., 2022</xref>). This approach overlooks the crucial intermediate step&#x2014;the <italic>&#x2018;perception and judgment&#x2019;</italic> phase, closely intertwined with the cognitive processes of reviewers (<xref rid="R16" ref-type="bibr">Lamont, 2009</xref>; <xref rid="R17" ref-type="bibr">Lee et al., 2013</xref>). Exploring this can offer valuable insights into the cognitive aspects of decision-making in peer review, enhance transparency, and build greater trust in the academic community.</p>
<p>Nevertheless, there are still some critical research gaps that may hinder further exploration: (a) The first challenge is about how to measure the reviewer&#x2019;s perception. Generally, reviewers&#x2019; perceptions could be revealed in their comments like enthusiastic praise or critical criticism (<xref rid="R26" ref-type="bibr">Mavrogenis et al., 2020</xref>). However, most peer review processes are not publicly available and related data is extremely scarce (<xref rid="R3" ref-type="bibr">Balietti et al., 2016</xref>), making it very difficult to obtain reviewer comments in real-world scenarios. Furthermore, the review contents are usually unstructured data, which makes it harder to extract meaningful insights. Prior studies examined review content through single-dimensional features like sentiment (<xref rid="R25" ref-type="bibr">Matsui et al., 2021</xref>), tone (<xref rid="R27" ref-type="bibr">Nobarany &#x0026; Booth, 2015</xref>) and length (<xref rid="R21" ref-type="bibr">Maddi &#x0026; Miotti, 2024</xref>), but have struggled to capture the multifaceted nature of the perceptions. (b) Few empirical studies assess the relationship between reviewers&#x2019; perceptions and decision-making. This lack of evidence hampers theory development and highlights gaps in understanding the peer review process and perceptual factors of reviewers.</p>
<p>To address these research gaps, we propose the following research questions (RQs):
<list list-type="simple">
<list-item><p><bold>RQ1:</bold> how can we automate the extraction of reviewers&#x2019; perceptions and judgments of a paper?</p></list-item>
<list-item><p><bold>RQ2:</bold> from the perspective of reviewers&#x2019; decisions, how do perceptual factors influence reviewers?</p></list-item>
<list-item><p><bold>RQ3:</bold> from the perspective of papers&#x2019; decisions, how do reviewers&#x2019; perceptions affect the decision of a paper?</p></list-item>
</list></p>
<p>To answer the above research questions, we collected a substantial number of open review records from OpenReview and carefully filtered them to construct a valuable peer-review dataset, comprising 114,640 records from top-tier computer science conferences. Then, we develop deep learning models to automatically extract and measure reviewers&#x2019; multi-dimensional perceptual factors (RQ1). Based on these factors, we conceptualize the peer review decision-making process into two primary phases: individual reviewer decisions (RQ2) and paper decisions affected by multiple reviewers (RQ3) and examine the relationship between reviewers&#x2019; perceptions and decision-making in these phases.</p>
</sec>
<sec id="sec2">
<title>Related work</title>
<p>Perception in peer review has been discussed by existing studies, particularly regarding how reviewers perceive the peer review process itself. Existing research has explored various aspects of reviewers&#x2019; perceptions, such as their motivations for reviewing, preferred roles, and preferences regarding review mechanisms (<xref rid="R8" ref-type="bibr">Curtin et al., 2018</xref>). Other studies have examined reviewers&#x2019; willingness to receive training, accept feedback, and sign their reviews (<xref rid="R22" ref-type="bibr">Manchikanti et al., 2015</xref>)<bold>.</bold> Less work has empirically investigated reviewers&#x2019; perceptions of the paper. One such study uses surveys to quantify the gap between researchers and reviewers regarding the perceived importance of specific validity and reliability issues (<xref rid="R34" ref-type="bibr">Street &#x0026; Ward, 2019</xref>).</p>
<p>The process from perception to decision in peer review is inherently subjective, and this subjectivity plays a crucial role in shaping the overall review process. Existing work argues that such subjectivity is not only unavoidable but also necessary, as it may help mitigate the herding effect and encourage diverse perspectives (<xref rid="R28" ref-type="bibr">Park et al., 2014</xref>). This subjectivity also contributes to the phenomenon of low inter-rater reliability among reviewers. While some researchers view this inconsistency as a natural and inevitable aspect of peer review (<xref rid="R32" ref-type="bibr">Roediger, 1991</xref>), others consider it problematic, suggesting that it undermines the reliability of the evaluation process (<xref rid="R24" ref-type="bibr">Marsh et al., 2007</xref>).</p>
<p>The process of review is not merely about evaluating the content of a paper; it is also shaped by <italic>&#x2018;sense of self and their relative positioning&#x2019;</italic> of reviewers (<xref rid="R16" ref-type="bibr">Lamont, 2009</xref>). The connection between identity and evaluation raises concerns about the potential for cognitive biases, which are often cited as a threat to the progress and integrity of a field (<xref rid="R4" ref-type="bibr">Benda &#x0026; Engels, 2011</xref>). Specifically, reviewers may show a preference for submissions from authors with similar <italic>&#x2018;schools of thought&#x2019;</italic>, a phenomenon known as <italic>&#x2018;cognitive cronyism&#x2019;</italic> (<xref rid="R37" ref-type="bibr">Travis &#x0026; Collins, 1991</xref>). Thus, while subjectivity in peer review can foster diverse viewpoints, it also presents challenges that must be addressed to ensure the fairness and accuracy of the evaluation process.</p>
</sec>
<sec id="sec3">
<title>Dataset</title>
<p>To address the issue that the majority of journals and conferences&#x2019; reviews are not publicly available, open peer review offers a promising solution to enhance the transparency and fairness of the review process (<xref rid="R43" ref-type="bibr">Wolfram et al., 2020</xref>). OpenReview (<xref rid="R33" ref-type="bibr">Soergel et al., 2013</xref>) is one of the most famous open peer review websites to promote transparency in scientific communication. It provides submitted papers alongside their corresponding reviews and comments. After a thorough examination of each venue in OpenReview, we focused on the top-tier machine learning conferences&#x2014;ICLR and NeurIPS for our study (ICLR makes reviews public regardless of acceptance, while NeurIPS allows authors of rejected papers to decide whether to make their reviews public). We collected all publicly available review data for these conferences. After data pre-processing and meticulous data filtering, we obtained review information for 30,440 papers from the year 2013 to 2024, comprising 114,640 records.</p>
</sec>
<sec id="sec4">
<title>Measurement of variables</title>
<sec id="sec4_1">
<title>Automatical extraction perceptual factors of reviewers (RQ1)</title>
<p><italic>&#x2018;Perceptual factors&#x2019;</italic> can be defined as the reviewer&#x2019;s subjective perception and judgment elements, which guide their review decision-making. The reviewer&#x2019;s perception of one submission can be directly reflected in the review comments (<xref rid="R26" ref-type="bibr">Mavrogenis et al., 2020</xref>). According to previous studies (<xref rid="R44" ref-type="bibr">Yuan et al., 2022</xref>), we identify seven key perceptual factors(These dimensions follow the ACL review guidelines): motivation, originality, soundness, substance, replicability, meaningful comparison, and clarity.</p>
<p>To automatically and effectively measure these factors, we designed a pipeline that includes a discriminative model and a scoring model, each based on RoBERTa (<xref rid="R20" ref-type="bibr">Liu, 2019</xref>). Specifically, the discriminative model determines whether a sentence is perception related. If it does, the scoring model then assesses the multidimensional perceptual factors. To illustrate these indicators more clearly, we provide an implication and a simplified example in <xref ref-type="fig" rid="F2">Figure 2</xref>. The detailed training process is outlined as follows (The hyperparameters for both models are set as follows: batch size of 16 and a learning rate of 5e-6. We employ the best model among 8 epochs with the lowest validation loss. The model is trained on an NVIDIA V100 32GB GPU):</p>
<p>We use AMPERE (<xref rid="R14" ref-type="bibr">Hua et al., 2019</xref>) and ASAP (<xref rid="R44" ref-type="bibr">Yuan et al., 2022</xref>) datasets to train our models. To train the discriminative model, we leverage AMPERE, which categorizes review comments into six types. We defined the task as a classification problem to identify whether a sentence belongs to the category <italic>&#x2018;evaluation&#x2019;.</italic> To train the scoring model, we use ASAP, which annotated seven dimensions of text spans with sentiment polarity. We converted it into a numeric labeling task. During training, the model takes the text as input and produces a 7-dimensional output, with each dimension representing a dimension. We use the Tanh function to scale each dimension. The loss is computed using Mean Squared Error (MSE). On the validation set (20% of total data), the model achieved a classification accuracy of 0.83 and an absolute deviation of 0.084 for each dimension. </p>
<fig id="F2">
<label>Figure 2.</label>
<caption><p>The implication (A) and toy example (B) of the extracted perceptual factors. Each factor&#x2019;s polarity is represented by a continuous value ranging from -1 to 1, where a value closer to 1 indicates higher confidence in a positive attitude. The first and last sentences in the example are categorized as &#x2018;<italic>fact&#x2019;</italic> and <italic>&#x2018;request&#x2019;,</italic> respectively. Unlike the <italic>&#x2018;evaluation&#x2019;</italic> category, these sentences exhibit less subjectivity and are thus excluded from measuring the reviewer&#x2019;s perceptual factors. The three sentences in between cover various perceptual factors such as <italic>&#x2018;motivation&#x2019;, &#x2018;originality&#x2019;, &#x2018;clarity&#x2019;,</italic> and <italic>&#x2018;soundness&#x2019;,</italic> showcasing the reviewer&#x2019;s diverse perceptions and judgments of the paper.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c57-fig2.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>After measuring the perceptual factors of each sentence, we aggregate the scores as follows:</p>
<disp-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="block"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:mo>&#x2211;</mml:mo> <mml:mrow><mml:mi>P</mml:mi><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>s</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mo>&#x2208;</mml:mo><mml:msub><mml:mi>T</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:msub><mml:mi>F</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>p</mml:mi></mml:msub></mml:mrow></mml:mfrac><mml:mstyle displaystyle='true'><mml:mo>&#x2211;</mml:mo> <mml:mrow><mml:mi>P</mml:mi><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where <italic>PF<sub>pi</sub></italic> represents the 7-dimensional vector of perceptual factors given by the <italic>i-th</italic> reviewer. <italic>seg</italic> represents the sentence of review content <italic>T<sub>pi</sub></italic>. <italic>PF<sub>p</sub></italic> denotes the average judgment of all reviewers of the paper. The more frequently reviewers mention and emphasize a particular dimension in their comments, the higher the perceptual score for that dimension.</p>
</sec>
<sec id="sec4_2">
<title>Reviewer decision</title>
<p>Reviewer decisions are a key criterion for chairs in determining paper acceptance, such as <italic>&#x2018;8: accept, good paper&#x2019;</italic> or <italic>&#x2018;3: reject, not good enough&#x2019;.</italic> Because different conferences use varying scoring ranges from year to year, we standardize all scores to a range of 0 to 10 to ensure comparability. The formula for this calculation is as follows:</p>
<disp-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="block"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>R</mml:mi><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mi>min</mml:mi><mml:mfenced><mml:mi>R</mml:mi></mml:mfenced></mml:mrow><mml:mrow><mml:mi>max</mml:mi><mml:mfenced><mml:mi>R</mml:mi></mml:mfenced><mml:mo>&#x2212;</mml:mo><mml:mi>min</mml:mi><mml:mfenced><mml:mi>R</mml:mi></mml:mfenced></mml:mrow></mml:mfrac><mml:mo>&#x00D7;</mml:mo><mml:mn>10</mml:mn><mml:mo>,</mml:mo><mml:mi>r</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mi>R</mml:mi></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>R</mml:mi><mml:msub><mml:mi>D</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>p</mml:mi></mml:msub></mml:mrow></mml:mfrac><mml:mstyle displaystyle='true'><mml:mo>&#x2211;</mml:mo> <mml:mrow><mml:mi>R</mml:mi><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where <italic>p</italic> represents a paper reviewed at a particular conference (held in a specific year). <italic>R</italic> denotes all the ratings given at this conference. <italic>r<sub>pi</sub></italic> refers to the original rating of the paper <italic>p</italic> given by the <italic>i- th</italic> reviewer, while <italic>RD<sub>pi</sub></italic> represents the processed one. <italic>N<sub>p</sub></italic> is the total number of reviews that the paper <italic>p</italic> received. <italic>RD<sub>p</sub></italic> is the average rating of all reviewers.</p>
</sec>
</sec>
<sec id="sec5">
<title>Result</title>
<sec id="sec5_1">
<title>Descriptive statistics of reviewers&#x2019; perceptual factors</title>
<fig id="F3">
<label>Figure 3.</label>
<caption><p>Description of the reviewers&#x2019; perceptual factors. (A) We calculated the Pearson correlation between different factors using data samples from individual reviews. It is shown that originality and motivation were the most correlated (and vice versa); soundness and clarity were the most correlated (and vice versa); substance and soundness were the most correlated; replicability and clarity were the most correlated; and meaningful comparison and substance were the most correlated. (B) We compared the average of reviewer decisions (<italic>RD</italic>) and perceptual factors for accepted and rejected papers. Unlike Figure 3 (A), which is based on single reviews, the data in Figure 3 (B) are based on individual papers, each with multiple reviews. We found that perceptual factors, whether relating to significance (Motivation), novelty (originality), or rigor (soundness, substance, e.g.), all have the potential to influence the decisions made about the papers. It shows the differences for each factor between the accepted and rejected categories, represented by the Cohen&#x2019;s d.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c57-fig3.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>Firstly, we observed the internal correlation of perceptual factors within reviewers&#x2019; comments. As shown in <xref ref-type="fig" rid="F3">Figure 3</xref> (A), we found that the internal correlations of these metrics were generally low (<italic>&#x2264;0.31</italic>), which means that there is a lower likelihood of serious multicollinearity issues in modeling.</p>
<p><xref ref-type="fig" rid="F3">Figure 3</xref> (B) demonstrates that accepted papers receive higher evaluations from reviewers across all criteria compared to rejected ones. Rejected papers often get positive feedback on motivation but tend to receive negative or no feedback (-0.01) on originality. This indicates reviewers are more supportive of motivation regardless of their recommendation but more critical of originality in papers they dislike. Further, reviewers often give negative evaluations on rigor-related dimensions to prompt authors to improve.</p>
</sec>
</sec>
<sec id="sec6">
<title>Perception-based decision-making analysis</title>
<sec id="sec6_1">
<title>Reviewer decision analysis (RQ2)</title>
<p>We analyse individual reviewers to examine how their decisions are influenced by their perceptions and judgments. The model is specified as follows:</p>
<disp-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="block"><mml:mrow><mml:mi>R</mml:mi><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:mi>&#x03B2;</mml:mi><mml:mi>P</mml:mi><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:mi>&#x03B2;</mml:mi><mml:mo>&#x0027;</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mi>p</mml:mi></mml:msub></mml:mrow></mml:math></disp-formula>
<p><italic>RD<sub>pi</sub></italic> represents the reviewer decision given by the <italic>i-th</italic> reviewer for paper <italic>p</italic>. The key variable <italic>PF<sub>pi</sub></italic> denotes the vector of the reviewer&#x2019;s perceptual factors. <italic>X<sub>pi</sub></italic> is a vector of control variables, including the request number and confidence of the reviewer. The request number represents the number of sentences related to the reviewers&#x2019; requests. Confidence is the assurance of reviewers, which is scored by themselves. Since the review process is double-blind, we do not control for author- related information in this model. <italic>&#x03B1;<sub>i</sub></italic> and <italic>&#x03BB;<sub>i</sub></italic> represent the fixed effects, capturing stable factors affecting the reviewer&#x2019;s decision at the venue and time levels respectively. <italic>&#x03B2;<sub>0</sub></italic> represents the constant term. We conducted a VIF test (Table S1) and used OLS regression for our analysis (<xref ref-type="table" rid="T1">Table 1</xref>).</p>
<p>We identified the impact of various perceptual factors on reviewer decisions. Among all perceptual factors (<xref ref-type="fig" rid="F4">Figure 4</xref>), reviewers prioritize originality, followed by motivation, as the primary basis for their decisions. Next in importance are soundness and substance, underscoring the paper&#x2019;s overall persuasiveness and rigor. Lastly, reviewers focus on meaningful comparison, clarity, and replicability. They determine whether a paper is easy to follow and disseminate and they also form the basis for the paper&#x2019;s credibility.</p>
<p>In <xref ref-type="fig" rid="F5">Figure 5</xref>, we conducted a heterogeneity analysis of reviewers grouped by high and low confidence levels (the boundary is set at 5). The results showed that decisions made by high- confidence reviewers were more strongly associated with perceptual factors (except for replicability and clarity).</p>
<table-wrap id="T1">
<label>Table 1.</label>
<caption><p>The regression result of reviewer decision analysis. The model has an R-squared value of 0.27, and the observation number is 107,272.</p></caption>
<table>
<thead>
<tr>
<th align="left" valign="top"><bold>Dependent Variable: Reviewer decision</bold></th>
<th align="center" valign="top"><bold>Coefficient</bold></th>
<th align="center" valign="top"><bold>SE</bold></th>
<th align="center" valign="top"><bold>P value</bold></th>
<th align="center" valign="top"><bold>Confidence interval</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Motivation</td>
<td align="center" valign="top">0.244</td>
<td align="center" valign="top">0.005</td>
<td align="center" valign="top">0.000</td>
<td align="center" valign="top">[0.235, 0.253]</td>
</tr>
<tr>
<td align="left" valign="top">Originality</td>
<td align="center" valign="top">0.372</td>
<td align="center" valign="top">0.005</td>
<td align="center" valign="top">0.000</td>
<td align="center" valign="top">[0.363, 0.381]</td>
</tr>
<tr>
<td align="left" valign="top">Soundness</td>
<td align="center" valign="top">0.214</td>
<td align="center" valign="top">0.003</td>
<td align="center" valign="top">0.000</td>
<td align="center" valign="top">[0.208, 0.220]</td>
</tr>
<tr>
<td align="left" valign="top">Substance</td>
<td align="center" valign="top">0.213</td>
<td align="center" valign="top">0.006</td>
<td align="center" valign="top">0.000</td>
<td align="center" valign="top">[0.201, 0.225]</td>
</tr>
<tr>
<td align="left" valign="top">Replicability</td>
<td align="center" valign="top">0.098</td>
<td align="center" valign="top">0.011</td>
<td align="center" valign="top">0.000</td>
<td align="center" valign="top">[0.076, 0.120]</td>
</tr>
<tr>
<td align="left" valign="top">Meaningful comparison</td>
<td align="center" valign="top">0.172</td>
<td align="center" valign="top">0.007</td>
<td align="center" valign="top">0.000</td>
<td align="center" valign="top">[0.158, 0.187]</td>
</tr>
<tr>
<td align="left" valign="top">Clarity</td>
<td align="center" valign="top">0.159</td>
<td align="center" valign="top">0.003</td>
<td align="center" valign="top">0.000</td>
<td align="center" valign="top">[0.154, 0.165]</td>
</tr>
<tr>
<td align="left" valign="top">Request number</td>
<td align="center" valign="top">0.020</td>
<td align="center" valign="top">0.001</td>
<td align="center" valign="top">0.000</td>
<td align="center" valign="top">[0.018, 0.021]</td>
</tr>
<tr>
<td align="left" valign="top">Confidence</td>
<td align="center" valign="top">-0.060</td>
<td align="center" valign="top">0.002</td>
<td align="center" valign="top">0.000</td>
<td align="center" valign="top">[-0.065, -0.056]</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F4">
<label>Figure 4.</label>
<caption><p>The coefficient comparison of reviewer decision analysis. Among all the perceptual factors, originality has the greatest impact on the reviewer&#x2019;s decision, followed by the perception and judgment of motivation. ** represents p &#x003C; 0.01, and * represents p &#x003C; 0.05.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c57-fig4.jpg"><alt-text>none</alt-text></graphic>
</fig>
<fig id="F5">
<label>Figure 5.</label>
<caption><p>The impact of perceptions reflected in review content between high-confidence and low- confidence reviewers on their decisions (using the standardized median value, i.e., 5, as a cut-off, with the standardization process similar to <italic>RD</italic>). The middle line represents the coefficient for confidence at the median value of 5. The study found that, except for clarity and replicability, the perceptions of high- confidence reviewers are most closely associated with their decisions across various aspects. This reflects that decisions made by high-confidence reviewers are more interpretable and follow a logical progression from perception to decision-making.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c57-fig5.jpg"><alt-text>none</alt-text></graphic>
</fig>
</sec>
<sec id="sec6_2">
<title>Paper decision analysis (RQ3)</title>
<p>When deciding whether to accept a paper, the conference chairs take into account the reviewers&#x2019; scores and feedback to make the final decision on acceptance or rejection. We further analyse how the chairs&#x2019; decision is made to gain insights into the process of decision-making in peer review. <italic>Xp</italic> is the vector of control variables, including average <italic>RD</italic>, request number and the prestige of paper, where prestige refers to the average h-index of all authors, matched with S2AG (Version: May 21, 2024) (<xref rid="R39" ref-type="bibr">Wade, 2022</xref>). <italic>PaperDecisionp</italic> represents the final decision for paper <italic>p</italic>. It is a binary variable representing whether the paper is accepted. The model is specified as follows:</p>
<disp-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="block"><mml:mrow><mml:mi>P</mml:mi><mml:mi>a</mml:mi><mml:mi>p</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>D</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:msub><mml:mi>n</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mi>R</mml:mi><mml:msub><mml:mi>D</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:mi>&#x03B2;</mml:mi><mml:mi>P</mml:mi><mml:msub><mml:mi>F</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:mi>&#x03B2;</mml:mi><mml:mo>&#x0027;</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mi>p</mml:mi></mml:msub></mml:mrow></mml:math></disp-formula>
<p>We conducted a VIF test (Table S2) and used logistic regression for our analysis (<xref ref-type="table" rid="T2">Table 2</xref>). We uncovered an interesting distinction (<xref ref-type="fig" rid="F6">Figure 6</xref>): from the perspective of the chairs, the most critical perceptual factors are not originality or motivation, but rather substance and soundness. This contrasts with the focus of reviewers. The findings suggest that chairs place greater emphasis on ensuring the overall validity in the field. Moreover, factors such as motivation, replicability, clarity, and meaningful comparison do not have a significant impact on paper decisions. This could be due to the supportive nature (as shown in <xref ref-type="fig" rid="F3">Figure 3</xref>) of factors like motivation, which may lack strong differentiating power from the perspective of papers&#x2019; decisions. It is shown in <xref ref-type="fig" rid="F7">Figure 7</xref> that in the high-<italic>RD</italic> group (RD below the median), positive judgments on validity-related dimensions, are more influential on papers&#x2019; decisions. </p>
<table-wrap id="T2">
<label>Table 2.</label>
<caption><p>The regression results for paper decision analysis show that the pseudo-R-squared is 0.55, with a sample size of 17,620 observations. Since chairs&#x2019; decisions are heavily influenced by the average scores provided by reviewers, we excluded papers rated below the 1st percentile of accepted submissions (a total of 5,464 papers, with a 97% probability of rejection) and those rated above the 99th percentile of rejected submissions (a total of 2,268 papers, with a 97% probability of acceptance).</p></caption>
<table>
<thead>
<tr>
<th align="left" valign="top"><bold>Dependent Variable: Paper decision</bold></th>
<th align="center" valign="top"><bold>Coefficient</bold></th>
<th align="center" valign="top"><bold>SE</bold></th>
<th align="center" valign="top"><bold>P value</bold></th>
<th align="center" valign="top"><bold>Odds multiplier</bold></th>
<th align="center" valign="top"><bold>Confidence interval</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Motivation</td>
<td align="center" valign="top">0.078</td>
<td align="center" valign="top">0.052</td>
<td align="center" valign="top">0.132</td>
<td align="center" valign="top">1.08</td>
<td align="center" valign="top">[-0.024, 0.180]</td>
</tr>
<tr>
<td align="left" valign="top">Originality</td>
<td align="center" valign="top">0.110</td>
<td align="center" valign="top">0.047</td>
<td align="center" valign="top">0.020</td>
<td align="center" valign="top">1.12</td>
<td align="center" valign="top">[0.017, 0.203]</td>
</tr>
<tr>
<td align="left" valign="top">Soundness</td>
<td align="center" valign="top">0.184</td>
<td align="center" valign="top">0.034</td>
<td align="center" valign="top">0.000</td>
<td align="center" valign="top">1.20</td>
<td align="center" valign="top">[0.118, 0.250]</td>
</tr>
<tr>
<td align="left" valign="top">Substance</td>
<td align="center" valign="top">0.192</td>
<td align="center" valign="top">0.064</td>
<td align="center" valign="top">0.003</td>
<td align="center" valign="top">1.21</td>
<td align="center" valign="top">[0.066, 0.318]</td>
</tr>
<tr>
<td align="left" valign="top">Replicability</td>
<td align="center" valign="top">-0.172</td>
<td align="center" valign="top">0.123</td>
<td align="center" valign="top">0.160</td>
<td align="center" valign="top">0.84</td>
<td align="center" valign="top">[-0.413, 0.068]</td>
</tr>
<tr>
<td align="left" valign="top">Meaningful comparison</td>
<td align="center" valign="top">-0.067</td>
<td align="center" valign="top">0.079</td>
<td align="center" valign="top">0.397</td>
<td align="center" valign="top">0.94</td>
<td align="center" valign="top">[-0.222, 0.088]</td>
</tr>
<tr>
<td align="left" valign="top">Clarity</td>
<td align="center" valign="top">0.059</td>
<td align="center" valign="top">0.032</td>
<td align="center" valign="top">0.070</td>
<td align="center" valign="top">1.06</td>
<td align="center" valign="top">[-0.005, 0.122]</td>
</tr>
<tr>
<td align="left" valign="top">Reviewer Decision</td>
<td align="center" valign="top">2.810</td>
<td align="center" valign="top">0.056</td>
<td align="center" valign="top">0.000</td>
<td align="center" valign="top">16.60</td>
<td align="center" valign="top">[2.700, 2.919]</td>
</tr>
<tr>
<td align="left" valign="top">Request number</td>
<td align="center" valign="top">-0.005</td>
<td align="center" valign="top">0.009</td>
<td align="center" valign="top">0.603</td>
<td align="center" valign="top">1.00</td>
<td align="center" valign="top">[-0.023, 0.013]</td>
</tr>
<tr>
<td align="left" valign="top">Prestige</td>
<td align="center" valign="top">0.005</td>
<td align="center" valign="top">0.002</td>
<td align="center" valign="top">0.020</td>
<td align="center" valign="top">1.01</td>
<td align="center" valign="top">[0.001, 0.010]</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F6">
<label>Figure 6.</label>
<caption><p>The coefficient comparison of paper decision analysis. We only present the main variables that are significant in the regression analysis. Among all the perceptual factors, substance and soundness have a significant positive correlation with the chair&#x2019;s decision, followed by originality. The effects of other main variables on the paper&#x2019;s decision are not significant. ** represents p &#x003C; 0.01, and * represents p &#x003C; 0.05.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c57-fig6.jpg"><alt-text>none</alt-text></graphic>
</fig>
<fig id="F7">
<label>Figure 7.</label>
<caption><p>We compared the high and low <italic>RD</italic> groups (divided by the median, with extremely high and low scores excluded in the same way as in <xref ref-type="table" rid="T2">Table 2</xref>) regarding reviewers&#x2019; perceptions of the papers. We found that decisions on papers in the high-<italic>RD</italic> group were more likely to be influenced by the reviewers&#x2019; perceptions. High scores already indicate the overall superiority of the paper&#x2019;s quality. If reviewers&#x2019; evaluations of the validity-related factors are positive, it indicates that the chairs&#x2019; concerns have been addressed, thus increasing the likelihood that the chairs will accept the paper.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c57-fig7.jpg"><alt-text>none</alt-text></graphic>
</fig>
</sec>
</sec>
<sec id="sec7">
<title>Conclusion and discussion</title>
<p>Due to limited review content and narrow analytical dimensions, there&#x2019;s little empirical research on the perception-to-decision process in peer review. Our findings show that originality is the most critical factor in reviewers&#x2019; decisions, followed by motivation. Chairs rely on reviewers&#x2019; decisions but prioritize assessments of soundness and substance, focusing on the research&#x2019;s overall validity. Reviewers&#x2019; feedback on motivation is reassuring but has limited distinguishing ability for chairs. The Overall impact of perceived meaningful comparison, clarity, and replicability on decision-making in peer review is weak.</p>
<p>While many believe that novelty is always hampered in the publication process (<xref rid="R18" ref-type="bibr">Liang et al., 2023</xref>; <xref rid="R40" ref-type="bibr">Wang et al., 2017</xref>), our study empirically uncovers that reviewers place a substantial emphasis on their perception of originality and are more inclined to accept papers they deem novel. This highlights that originality is the most valued and recognized element by reviewers. However, the transition from perception to decision is complex and still largely unexplored. Specifically, reviewers&#x2019; perspectives can be influenced by preconceptions or past experiences, and there is often a gap between perception and its expression. Additionally, cognitive limitations may obscure underlying social biases. The intricate mechanisms and overall effectiveness of this system demand ongoing exploration and refinement. Within this context, reviewers&#x2019; cognition, perception, and decision-making are indispensable and critical components. While the peer review system benefits science, it also necessitates a continuous cycle of feedback to evolve and improve (<xref rid="R41" ref-type="bibr">Ward et al., 2015</xref>).</p>
<sec id="sec7_1">
<title>Limitation and future work</title>
<p><bold>Due to challenges in accessing review content, our data has some constraints. We have also lightly addressed chairs&#x2019; perceptions, which we will explore in more depth. The requests from reviewers also contain important information, and there is potential for further analysis.</bold></p>
</sec>
</sec>
</body>
<back>
<ack>
<title>Acknowledgements</title>
<p>This work is supported by the National Natural Science Foundation of China (72104212).</p>
</ack>
<ref-list>
<title>References</title>
<ref id="R1"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alberts</surname><given-names>B.</given-names></name><name><surname>Hanson</surname><given-names>B.</given-names></name><name><surname>Kelner</surname><given-names>K. L.</given-names></name></person-group><year>2008</year><article-title>Reviewing Peer Review</article-title><source>Science</source><volume>321</volume><issue>5885</issue><fpage>15</fpage><lpage>15</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1126/science.1162115">https://doi.org/10.1126/science.1162115</ext-link></element-citation></ref>
<ref id="R2"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bailly</surname><given-names>B. A. F. L.</given-names></name></person-group><year>2016</year><article-title>Learning from peer review</article-title><source>Nature Nanotechnology</source><volume>11</volume><issue>2</issue><fpage>204</fpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1038/nnano.2016.4">https://doi.org/10.1038/nnano.2016.4</ext-link></element-citation></ref>
<ref id="R3"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Balietti</surname><given-names>S.</given-names></name><name><surname>Goldstone</surname><given-names>R. L.</given-names></name><name><surname>Helbing</surname><given-names>D.</given-names></name></person-group><year>2016</year><article-title>Peer review and competition in the Art Exhibition Game</article-title><source>Proceedings of the National Academy of Sciences</source><volume>113</volume><issue>30</issue><fpage>8414</fpage><lpage>8419</lpage></element-citation></ref>
<ref id="R4"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Benda</surname><given-names>W. G.</given-names></name><name><surname>Engels</surname><given-names>T. C.</given-names></name></person-group><year>2011</year><article-title>The predictive validity of peer review: A selective review of the judgmental forecasting qualities of peers, and implications for innovation in science</article-title><source>International Journal of Forecasting</source><volume>27</volume><issue>1</issue><fpage>166</fpage><lpage>182</lpage></element-citation></ref>
<ref id="R5"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Blockeel</surname><given-names>C.</given-names></name><name><surname>Drakopoulos</surname><given-names>P.</given-names></name><name><surname>Polyzos</surname><given-names>N.</given-names></name><name><surname>Tournaye</surname><given-names>H.</given-names></name><name><surname>Garc&#x00ED;a-Velasco</surname><given-names>J.</given-names></name></person-group><year>2017</year><article-title>Review the &#x2018;peer review&#x2019;</article-title><source>Reproductive Biomedicine Online</source><volume>35</volume><issue>6</issue><fpage>747</fpage><lpage>749</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.rbmo.2017.08.017">https://doi.org/10.1016/j.rbmo.2017.08.017</ext-link></element-citation></ref>
<ref id="R6"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Brosch</surname><given-names>T.</given-names></name><name><surname>Scherer</surname><given-names>K.</given-names></name><name><surname>Grandjean</surname><given-names>D.</given-names></name><name><surname>Sander</surname><given-names>D.</given-names></name></person-group><year>2013</year><article-title>The impact of emotion on perception, attention, memory, and decision-making</article-title><source>Swiss Medical Weekly</source><volume>143</volume><issue>1920</issue><fpage>w13786</fpage><lpage>w13786</lpage></element-citation></ref>
<ref id="R7"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chong</surname><given-names>S. W.</given-names></name><name><surname>Mason</surname><given-names>S.</given-names></name></person-group><year>2021</year><article-title>Demystifying the process of scholarly peer-review: An autoethnographic investigation of feedback literacy of two award-winning peer reviewers</article-title><source>Humanities and Social Sciences Communications</source><volume>8</volume><issue>1</issue><fpage>1</fpage><lpage>11</lpage></element-citation></ref>
<ref id="R8"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Curtin</surname><given-names>P. A.</given-names></name><name><surname>Russial</surname><given-names>J.</given-names></name><name><surname>Tefertiller</surname><given-names>A. C.</given-names></name></person-group><year>2018</year><article-title>Reviewers&#x2019; Perceptions of the Peer Review Process in Journalism and Mass Communication</article-title><source>Journalism &#x0026; Mass Communication Quarterly</source><volume>95</volume><fpage>278</fpage><lpage>299</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1177/1077699017736031">https://doi.org/10.1177/1077699017736031</ext-link></element-citation></ref>
<ref id="R9"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Cutler</surname><given-names>S.</given-names></name><name><surname>Xia</surname><given-names>Y.</given-names></name><name><surname>Beddoes</surname><given-names>K.</given-names></name></person-group><year>2022</year><article-title>Peering into the Black Box of Peer Review: From the Perspective of Editors</article-title><source>9th Research in Engineering Education Symposium (REES 2021) and 32nd Australasian Association for Engineering Education Conference (REES AAEE 2021)</source><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.52202/066488-0095">https://doi.org/10.52202/066488-0095</ext-link></element-citation></ref>
<ref id="R10"><element-citation publication-type="book"><person-group person-group-type="author"><name><surname>Dror</surname><given-names>I. E.</given-names></name><name><surname>Fraser-Mackenzie</surname><given-names>P. A.</given-names></name></person-group><year>2008</year><chapter-title>Cognitive biases in human perception, judgment, and decision making: Bridging theory and the real world</chapter-title><source>In Criminal investigative failures</source><fpage>79</fpage><lpage>94</lpage><publisher-name>Routledge</publisher-name></element-citation></ref>
<ref id="R11"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Garc&#x00ED;a</surname><given-names>J. A.</given-names></name><name><surname>Rodriguez-S&#x00E1;nchez</surname><given-names>R.</given-names></name><name><surname>Fdez-Valdivia</surname><given-names>J.</given-names></name></person-group><year>2015</year><article-title>Bias and effort in peer review</article-title><source>Journal of the Association for Information Science and Technology</source><volume>66</volume><issue>10</issue><comment>Article 10</comment></element-citation></ref>
<ref id="R12"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hamilton</surname><given-names>D.</given-names></name><name><surname>Fraser</surname><given-names>H.</given-names></name><name><surname>Hoekstra</surname><given-names>R.</given-names></name><name><surname>Fidler</surname><given-names>F.</given-names></name></person-group><year>2020</year><article-title>Journal policies and editors&#x2019; opinions on peer review</article-title><source>eLife</source><volume>9</volume><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.7554/eLife.62529">https://doi.org/10.7554/eLife.62529</ext-link></element-citation></ref>
<ref id="R13"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Howard</surname><given-names>L.</given-names></name><name><surname>Wilkinson</surname><given-names>G.</given-names></name></person-group><year>1998</year><article-title>Peer review and editorial decision-making</article-title><source>The British Journal of Psychiatry</source><volume>173</volume><issue>2</issue><fpage>110</fpage><lpage>113</lpage></element-citation></ref>
<ref id="R14"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Hua</surname><given-names>X.</given-names></name><name><surname>Nikolov</surname><given-names>M.</given-names></name><name><surname>Badugu</surname><given-names>N.</given-names></name><name><surname>Wang</surname><given-names>L.</given-names></name></person-group><year>2019</year><article-title>Argument Mining for Understanding Peer Reviews</article-title><source>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)</source><fpage>2131</fpage><lpage>2137</lpage></element-citation></ref>
<ref id="R15"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jefferson</surname><given-names>T.</given-names></name><name><surname>Wager</surname><given-names>E.</given-names></name><name><surname>Davidoff</surname><given-names>F.</given-names></name></person-group><year>2002</year><article-title>Measuring the quality of editorial peer review</article-title><source>Jama</source><volume>287</volume><issue>21</issue><fpage>2786</fpage><lpage>2790</lpage></element-citation></ref>
<ref id="R16"><element-citation publication-type="book"><person-group person-group-type="author"><name><surname>Lamont</surname><given-names>M.</given-names></name></person-group><year>2009</year><source>How professors think: Inside the curious world of academic judgment</source><publisher-name>Harvard University Press</publisher-name></element-citation></ref>
<ref id="R17"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname><given-names>C. J.</given-names></name><name><surname>Sugimoto</surname><given-names>C. R.</given-names></name><name><surname>Zhang</surname><given-names>G.</given-names></name><name><surname>Cronin</surname><given-names>B.</given-names></name></person-group><year>2013</year><article-title>Bias in peer review</article-title><source>Journal of the American Society for Information Science and Technology</source><volume>64</volume><issue>1</issue><fpage>2</fpage><lpage>17</lpage></element-citation></ref>
<ref id="R18"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liang</surname><given-names>Z.</given-names></name><name><surname>Mao</surname><given-names>J.</given-names></name><name><surname>Li</surname><given-names>G.</given-names></name></person-group><year>2023</year><article-title>Bias against scientific novelty: A prepublication perspective</article-title><source>Journal of the Association for Information Science and Technology</source><volume>74</volume><issue>1</issue><fpage>99</fpage><lpage>114</lpage></element-citation></ref>
<ref id="R19"><element-citation publication-type="book"><person-group person-group-type="author"><name><surname>Lindsay</surname><given-names>P. H.</given-names></name><name><surname>Norman</surname><given-names>D. A.</given-names></name></person-group><year>2013</year><source>Human information processing: An introduction to psychology</source><publisher-name>Academic press</publisher-name></element-citation></ref>
<ref id="R20"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>Y.</given-names></name></person-group><year>2019</year><article-title>Roberta: A robustly optimized bert pretraining approach</article-title><source>arXiv Preprint arXiv:1907.11692</source></element-citation></ref>
<ref id="R21"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Maddi</surname><given-names>A.</given-names></name><name><surname>Miotti</surname><given-names>L.</given-names></name></person-group><year>2024</year><article-title>On the peer review reports: Does size matter?</article-title><source>Scientometrics</source><fpage>1</fpage><lpage>21</lpage></element-citation></ref>
<ref id="R22"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Manchikanti</surname><given-names>L.</given-names></name><name><surname>Kaye</surname><given-names>A.</given-names></name><name><surname>Boswell</surname><given-names>M.</given-names></name><name><surname>Hirsch</surname><given-names>J.</given-names></name></person-group><year>2015</year><article-title>Medical journal peer review: Process and bias</article-title><source>Pain Physician</source><volume>18</volume><issue>1</issue><fpage>E1</fpage><lpage>E14</lpage></element-citation></ref>
<ref id="R23"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Marietta</surname><given-names>M.</given-names></name><name><surname>Barker</surname><given-names>D.</given-names></name></person-group><year>2019</year><article-title>The Psychology of Fact Perceptions II</article-title><source>One Nation, Two Realities</source><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1093/OSO/9780190677176.003.0006">https://doi.org/10.1093/OSO/9780190677176.003.0006</ext-link></element-citation></ref>
<ref id="R24"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Marsh</surname><given-names>H. W.</given-names></name><name><surname>Bond</surname><given-names>N. W.</given-names></name><name><surname>Jayasinghe</surname><given-names>U. W.</given-names></name></person-group><year>2007</year><article-title>Peer review process: Assessments by applicant-nominated referees are biased, inflated, unreliable and invalid</article-title><source>Australian Psychologist</source><volume>42</volume><issue>1</issue><fpage>33</fpage><lpage>38</lpage></element-citation></ref>
<ref id="R25"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Matsui</surname><given-names>A.</given-names></name><name><surname>Chen</surname><given-names>E.</given-names></name><name><surname>Wang</surname><given-names>Y.</given-names></name><name><surname>Ferrara</surname><given-names>E.</given-names></name></person-group><year>2021</year><article-title>The impact of peer review on the contribution potential of scientific papers</article-title><source>PeerJ</source><volume>9</volume><fpage>e11999</fpage></element-citation></ref>
<ref id="R26"><element-citation publication-type="book"><person-group person-group-type="author"><name><surname>Mavrogenis</surname><given-names>A. F.</given-names></name><name><surname>Quaile</surname><given-names>A.</given-names></name><name><surname>Scarlat</surname><given-names>M. M.</given-names></name></person-group><year>2020</year><chapter-title>The good, the bad and the rude peer-review</chapter-title><source>In International Orthopaedics</source><volume>44</volume><fpage>413</fpage><lpage>415</lpage><publisher-name>Springer</publisher-name></element-citation></ref>
<ref id="R27"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nobarany</surname><given-names>S.</given-names></name><name><surname>Booth</surname><given-names>K. S.</given-names></name></person-group><year>2015</year><article-title>Use of politeness strategies in signed open peer review</article-title><source>Journal of the Association for Information Science and Technology</source><volume>66</volume><issue>5</issue><fpage>1048</fpage><lpage>1064</lpage></element-citation></ref>
<ref id="R28"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Park</surname><given-names>I.-U.</given-names></name><name><surname>Peacey</surname><given-names>M. W.</given-names></name><name><surname>Munafo</surname><given-names>M. R.</given-names></name></person-group><year>2014</year><article-title>Modelling the effects of subjective and objective decision making in scientific peer review</article-title><source>Nature</source><volume>506</volume><issue>7486</issue><fpage>93</fpage><lpage>96</lpage></element-citation></ref>
<ref id="R29"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Peters</surname><given-names>D. P.</given-names></name><name><surname>Ceci</surname><given-names>S. J.</given-names></name></person-group><year>1982</year><article-title>Peer-review research: Objections and obligations</article-title><source>Behavioral and Brain Sciences</source><volume>5</volume><issue>2</issue><fpage>246</fpage><lpage>255</lpage></element-citation></ref>
<ref id="R30"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Phillips</surname><given-names>J.</given-names></name></person-group><year>2011</year><article-title>Expert bias in peer review</article-title><source>Current Medical Research and Opinion</source><volume>27</volume><fpage>2229</fpage><lpage>2233</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1185/03007995.2011.624090">https://doi.org/10.1185/03007995.2011.624090</ext-link></element-citation></ref>
<ref id="R31"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Riley</surname><given-names>B.</given-names></name><name><surname>Jones</surname><given-names>R. B.</given-names></name></person-group><year>2016</year><article-title>Peer review: Acknowledging its value and recognising the reviewers</article-title><source>The British Journal of General Practice: The Journal of the Royal College of General Practitioners</source><volume>66</volume><issue>653</issue><fpage>629</fpage><lpage>630</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3399/BJGP16X688285">https://doi.org/10.3399/BJGP16X688285</ext-link></element-citation></ref>
<ref id="R32"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Roediger</surname><given-names>H. L.</given-names></name></person-group><year>1991</year><article-title>Is unreliability in peer review harmful?</article-title><source>Behavioral and Brain Sciences</source><volume>14</volume><issue>1</issue><fpage>159</fpage><lpage>160</lpage></element-citation></ref>
<ref id="R33"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Soergel</surname><given-names>D.</given-names></name><name><surname>Saunders</surname><given-names>A.</given-names></name><name><surname>McCallum</surname><given-names>A.</given-names></name></person-group><year>2013</year><article-title>Open Scholarship and Peer Review: A Time for Experimentationo</article-title></element-citation></ref>
<ref id="R34"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Street</surname><given-names>C. T.</given-names></name><name><surname>Ward</surname><given-names>K.</given-names></name></person-group><year>2019</year><article-title>Cognitive Bias in the Peer Review Process: Understanding a Source of Friction between Reviewers and Researchers</article-title><source>Data Base</source><volume>50</volume><fpage>52</fpage><lpage>70</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1145/3371041.3371046">https://doi.org/10.1145/3371041.3371046</ext-link></element-citation></ref>
<ref id="R35"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Teplitskiy</surname><given-names>M.</given-names></name><name><surname>Peng</surname><given-names>H.</given-names></name><name><surname>Blasco</surname><given-names>A.</given-names></name><name><surname>Lakhani</surname><given-names>K. R.</given-names></name></person-group><year>2022</year><article-title>Is novel research worth doing? Evidence from peer review at 49 journals</article-title><source>Proceedings of the National Academy of Sciences</source><volume>119</volume><issue>47</issue><fpage>e2118046119</fpage></element-citation></ref>
<ref id="R36"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tomkins</surname><given-names>A.</given-names></name><name><surname>Zhang</surname><given-names>M.</given-names></name><name><surname>Heavlin</surname><given-names>W. D.</given-names></name></person-group><year>2017</year><article-title>Reviewer bias in single-versus double-blind peer review</article-title><source>Proceedings of the National Academy of Sciences</source><volume>114</volume><issue>48</issue><fpage>12708</fpage><lpage>12713</lpage></element-citation></ref>
<ref id="R37"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Travis</surname><given-names>G. D. L.</given-names></name><name><surname>Collins</surname><given-names>H. M.</given-names></name></person-group><year>1991</year><article-title>New light on old boys: Cognitive and institutional particularism in the peer review system</article-title><source>Science, Technology, &#x0026; Human Values</source><volume>16</volume><issue>3</issue><fpage>322</fpage><lpage>341</lpage></element-citation></ref>
<ref id="R38"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Tvina</surname><given-names>A.</given-names></name><name><surname>Spellecy</surname><given-names>R.</given-names></name><name><surname>Palatnik</surname><given-names>A.</given-names></name></person-group><year>2019</year><article-title>Bias in the Peer Review Process: Can We Do Better?</article-title><source>Obstetrics &#x0026; Gynecology</source><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1097/A0G.0000000000003260">https://doi.org/10.1097/A0G.0000000000003260</ext-link></element-citation></ref>
<ref id="R39"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Wade</surname><given-names>A. D.</given-names></name></person-group><year>2022</year><article-title>The semantic scholar academic graph (s2ag)</article-title><comment>Companion Proceedings of the Web Conference 2022</comment><fpage>739</fpage><lpage>739</lpage></element-citation></ref>
<ref id="R40"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>J.</given-names></name><name><surname>Veugelers</surname><given-names>R.</given-names></name><name><surname>Stephan</surname><given-names>P.</given-names></name></person-group><year>2017</year><article-title>Bias against novelty in science: A cautionary tale for users of bibliometric indicators</article-title><source>Research Policy</source><volume>46</volume><issue>8</issue><fpage>1416</fpage><lpage>1436</lpage></element-citation></ref>
<ref id="R41"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ward</surname><given-names>P.</given-names></name><name><surname>Graber</surname><given-names>K. C.</given-names></name><name><surname>Van Der Mars</surname><given-names>H.</given-names></name></person-group><year>2015</year><article-title>Writing quality peer reviews of research manuscripts</article-title><source>Journal of Teaching in Physical Education</source><volume>34</volume><issue>4</issue><fpage>700</fpage><lpage>715</lpage></element-citation></ref>
<ref id="R42"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Wenneras</surname><given-names>C.</given-names></name><name><surname>Wold</surname><given-names>A.</given-names></name></person-group><year>2001</year><source>Nepotism and sexism in peer-review</source></element-citation></ref>
<ref id="R43"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wolfram</surname><given-names>D.</given-names></name><name><surname>Wang</surname><given-names>P.</given-names></name><name><surname>Hembree</surname><given-names>A.</given-names></name><name><surname>Park</surname><given-names>H.</given-names></name></person-group><year>2020</year><article-title>Open peer review: Promoting transparency in open science</article-title><source>Scientometrics</source><volume>125</volume><issue>2</issue><fpage>1033</fpage><lpage>1051</lpage></element-citation></ref>
<ref id="R44"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yuan</surname><given-names>W.</given-names></name><name><surname>Liu</surname><given-names>P.</given-names></name><name><surname>Neubig</surname><given-names>G.</given-names></name></person-group><year>2022</year><article-title>Can we automate scientific reviewing?</article-title><source>Journal of Artificial Intelligence Research</source><volume>75</volume><fpage>171</fpage><lpage>212</lpage></element-citation></ref>
</ref-list>
<app-group>
<app id="app1">
<title>Appendix</title>
<sec id="app_sec1">
<title>Supplementary materials</title>
<p>See supplementary materials.</p>
</sec>
</app>
</app-group>
</back>
</article>