This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Game-Theoretical Analysis of Reviewer Rewards in Peer-Review Journal Systems: Analysis and Experimental Evaluation using Deep Reinforcement Learning

Minhyeok Lee1
Abstract

In this paper, we navigate the intricate domain of reviewer rewards in open-access academic publishing, leveraging the precision of mathematics and the strategic acumen of game theory. We conceptualize the prevailing voucher-based reviewer reward system as a two-player game, subsequently identifying potential shortcomings that may incline reviewers towards binary decisions. To address this issue, we propose and mathematically formalize an alternative reward system with the objective of mitigating this bias and promoting more comprehensive reviews. We engage in a detailed investigation of the properties and outcomes of both systems, employing rigorous game-theoretical analysis and deep reinforcement learning simulations. Our results underscore a noteworthy divergence between the two systems, with our proposed system demonstrating a more balanced decision distribution and enhanced stability. This research not only augments the mathematical understanding of reviewer reward systems, but it also provides valuable insights for the formulation of policies within journal review system. Our contribution to the mathematical community lies in providing a game-theoretical perspective to a real-world problem and in the application of deep reinforcement learning to simulate and understand this complex system.

Keywords: game theory; open-access journals; reviewer rewards; voucher-based rewards; decision-making; Nash equilibrium; mathematical modeling; peer review process; academic publishing

1 Introduction

The world of academia has witnessed a profound shift with the rise and growth of open-access journals, enabling scientific research to be shared publicly and widely at a scale never seen before [1, 2, 3]. Nevertheless, the critical need for thorough and impactful peer review continues to be a central challenge in this setting. To inspire reviewers, some open-access journals have introduced a reward system in the form of fee waivers for publishing, offered as vouchers [4]. This paper delves into a thorough mathematical and game-theoretical examination of this voucher-based reviewer reward system.

Under the current system, a tendency exists for reviewers to gravitate towards black-and-white outcomes, that is, either accepting or rejecting a manuscript, instead of proposing amendments or enhancements. This is largely because rewards are typically granted more quickly when a paper is either wholly accepted or rejected, as opposed to when revisions are suggested. Although this method can speed up the review process, it risks compromising the quality and depth of peer reviews, thus promoting an all-or-nothing decision-making approach over a more refined process that includes revisions. This paper proposes and examines a new reward system designed to balance these incentives and promote more thorough reviews.

Our study begins by framing the peer review process as a strategic interaction, or game, between two reviewers. Using the principles of game theory, we carefully analyze the strategies and behaviors displayed by the reviewers when influenced by both the current and proposed reward systems. In our newly proposed system, the reward structure is adjusted to not only promote agreement between reviewers but also to discourage hasty or binary decisions.

To investigate the features of both reward systems, we use deep reinforcement learning techniques to simulate this game [5, 6, 7]. Two agents, represented as neural networks [8, 9, 10], are trained to mimic the roles of reviewers, aiming to maximize their anticipated gain in the game. The training of these agents is steered by a reward function that reflects the voucher reward system common in open-access journals, and a loss function that promotes decisions in line with the true quality of the paper under review.

This paper offers a mathematical explanation of the current reviewer reward system, interpreted as a game. It outlines a model of the current system, followed by a game-theoretical analysis that includes defining dominant strategies and a thorough discussion of reviewer behavior under this system. Furthermore, we propose a novel reviewer reward system based on game-theoretical methods, followed by analyzing the strategies and behaviors it encourages in reviewers. This includes defining optimal strategies and an in-depth look at how these strategies are shaped by the changes we suggest to the reward system.

Throughout this research, we emphasize the game-theoretical foundations of our work. By breaking down the interaction between reviewers as a strategic game, we can meticulously analyze and contrast the two systems in a mathematical framework. This approach allows us to pinpoint potential weaknesses in the existing system and provide solid mathematical evidence supporting the effectiveness of our proposed system.

This paper intends to contribute significantly to the existing body of mathematical literature, offering a fresh game-theoretic perspective on the reviewer reward system in open-access journals. We trust that our findings will encourage further mathematical exploration into this vital aspect of the academic publishing process. Our work may also inform the policies of open-access journals, equipping them with a sound theoretical foundation for crafting reviewer reward systems that endorse thorough and unbiased reviews.

2 Preliminaries

2.1 Game Theory

Game theory is a mathematical framework designed for understanding the interaction among rational agents [11, 12, 13]. A game in this context is a situation where a number of agents interact, each trying to optimize their own objectives.

Definition 2.1.

A game 𝒢\mathcal{G} is defined by a triplet (𝒫,𝒮,𝒰)(\mathcal{P},\mathcal{S},\mathcal{U}), where:

  • 𝒫={1,2,,n}\mathcal{P}=\{1,2,...,n\} is a set of nn players.

  • 𝒮=S1×S2××Sn\mathcal{S}=S_{1}\times S_{2}\times\cdots\times S_{n} is the strategy space, with SiS_{i} being the strategy set of player ii.

  • 𝒰=(U1,U2,,Un)\mathcal{U}=(U_{1},U_{2},...,U_{n}) is the set of utility functions, with Ui:𝒮U_{i}:\mathcal{S}\rightarrow\mathbb{R} being the utility function of player ii.

In this setting, a strategy is a complete plan of action that defines what a player would do in every situation. A strategy profile s𝒮s\in\mathcal{S} is a vector that specifies a strategy for every player. Each player ii’s utility function UiU_{i} maps from the set of strategy profiles to the real numbers, representing the player’s preferences over the set of possible outcomes.

Definition 2.2.

A strategy profile s𝒮s^{*}\in\mathcal{S} is a Nash equilibrium if, for each player i𝒫i\in\mathcal{P}, we have

Ui(si,si)Ui(si,si),siSi,U_{i}(s^{*}_{i},s^{*}_{-i})\geq U_{i}(s_{i},s^{*}_{-i}),\quad\forall s_{i}\in S_{i}, (1)

where sis^{*}_{-i} denotes the strategies of all players except player ii in strategy profile ss^{*}.

The Nash equilibrium, as defined in Definition 2.2, is a state of the game where no player can unilaterally deviate from their current strategy without decreasing their utility, given that all other players stick to their strategies. This concept is central in game theory and provides a solution concept for strategic interactions.

Remark 2.1.

In some games, players may not have pure strategy Nash equilibriums, i.e., equilibriums where players use deterministic strategies. In such cases, players may resort to mixed strategies, where they randomize over their pure strategies. A mixed strategy Nash equilibrium is a profile of mixed strategies from which no player wishes to deviate.

Definition 2.3.

A payoff matrix \mathcal{M} of a two-player game is a matrix where the (i,j)(i,j)-th entry represents the utility of player 1 when player 1 chooses strategy ii and player 2 chooses strategy jj.

The payoff matrix is a straightforward representation of a two-player game where each player has a finite number of strategies [14, 15, 16]. This matrix offers an efficient method to examine the utilities for different strategy combinations.

Example 2.1.

Consider a two-player game where each player has two strategies. The payoff matrix for player 1 is

=[U1(s1,s1)U1(s1,s2)U1(s2,s1)U1(s2,s2)].\mathcal{M}=\begin{bmatrix}U_{1}(s_{1},s_{1})&U_{1}(s_{1},s_{2})\\ U_{1}(s_{2},s_{1})&U_{1}(s_{2},s_{2})\end{bmatrix}.
Proposition 2.1.

In a two-player game, if there exists a strategy sis_{i} for player ii such that Ui(si,sj)Ui(sk,sj)U_{i}(s_{i},s_{j})\geq U_{i}(s_{k},s_{j}) for all sjSjs_{j}\in S_{j} and skSis_{k}\in S_{i}, sksis_{k}\neq s_{i}, then sis_{i} is a dominant strategy for player ii.

This proposition defines a dominant strategy, which yields the highest utility for a player, regardless of the strategies chosen by other players.

Lemma 2.1.

In a two-player game, if there exists a strategy profile (s1,s2)(s^{*}_{1},s^{*}_{2}) such that s1s^{*}_{1} is a dominant strategy for player 1 and s2s^{*}_{2} is a dominant strategy for player 2, then (s1,s2)(s^{*}_{1},s^{*}_{2}) is a Nash equilibrium.

This lemma establishes a connection between the dominant strategy and Nash equilibrium in a two-player game. If both players have dominant strategies, the strategy profile consisting of these dominant strategies is a Nash equilibrium.

Assumption 2.1.

We assume that all players are rational, i.e., each player seeks to maximize their own utility.

This assumption is a basic tenet of game theory and is essential for our subsequent analysis. It posits that players will always act in a way that they believe will yield them the greatest utility.

2.2 Open-Access Journal Review System

Open-access (OA) journals have emerged as a significant development in the scholarly publishing industry [17, 18, 19]. These journals provide unrestricted online access to peer-reviewed scholarly research. In the OA model, rather than charging readers to access the content, the publication costs are usually covered by author fees or institutional funding.

Definition 2.4.

Open-access publishing is a model for scholarly journal publishing that makes research articles available to readers at no cost, as opposed to the traditional subscription model in which readers have access to scholarly research by paying a subscription fee.

The open-access publishing model has revolutionized the way knowledge is disseminated, allowing for a much wider spread of scholarly research. It is predicated on the philosophy of communal sharing of knowledge, making the fruits of academic endeavor available to all, free of charge. The financial burden is thus transferred from the reader to the author, ensuring that the paywalls, which may hinder the proliferation of knowledge, are eliminated.

An integral part of the academic publishing process, irrespective of the publishing model, is the peer-review process. This process is designed to maintain the quality and integrity of scholarly research. In the OA model, the reviewers play a crucial role in assessing the quality of papers.

Definition 2.5.

A peer review is an evaluation of scientific, academic, or professional work by others working in the same field to maintain quality and improve performance. In the context of OA publishing, the peer reviewers assess the quality of the submissions and make a recommendation regarding their publication.

Peer review acts as a gatekeeper, ensuring that only rigorous and high-quality research is disseminated [20, 21, 22]. The process involves the evaluation of an academic manuscript by experts in the same field as the paper. This stringent process helps to ensure that the research presented is of the highest standard and adds value to the field. The rigorous nature of this process is a critical factor in upholding the reputation and credibility of scientific research.

There are usually three options for the reviewers when assessing a paper: Accept, Revision, and Reject. An "Accept" decision means that the paper is ready for publication in its current form. A "Revision" decision indicates that the paper has potential but requires some changes before it is ready for publication. A "Reject" decision means that the paper is not suitable for publication in the journal.

2.3 Assumptions in Review System

Definition 2.6.

Let 𝒟={Accept,Revision,Reject}\mathcal{D}=\{\text{Accept},\text{Revision},\text{Reject}\} represent the set of possible decisions a reviewer can make for a paper. A decision d𝒟d\in\mathcal{D} is a recommendation made by a reviewer regarding the publication of a paper.

The tripartite system of Accept, Revision, and Reject decisions provides a comprehensive framework for evaluating academic papers. An acceptance indicates a work of high quality, requiring no further alterations. A recommendation for revision, on the other hand, indicates that while the work holds promise, there are aspects that require refinement and improvement before it meets the standards of the journal. A rejection signifies that the work, in its current form, does not meet the necessary standards of quality and contribution to the field.

Assumption 2.2.

We assume that each paper is reviewed by two independent reviewers, and each reviewer makes a decision from the set 𝒟\mathcal{D} independently based on their assessment of the paper’s quality.

Remark 2.2.

In the event that the two reviewers provide differing decisions, a third reviewer may be called upon to break the tie. The final decision is then based on the majority opinion.

The introduction of a third reviewer in case of a disagreement between the initial reviewers ensures that the decision-making process is fair and unbiased. This mechanism ensures that a single reviewer cannot unduly influence the outcome, thus promoting the integrity of the review process. This procedure also underscores the importance of consensus in academic publishing.

In some OA journals, as a part of the incentive system, reviewers are rewarded with vouchers upon completion of their review tasks. These vouchers can be used for discounts on the open access publishing fees when the reviewers submit their own papers.

Definition 2.7.

A reviewer reward voucher is a form of remuneration given to reviewers upon completion of a review task. This voucher can be used as a discount on the open-access publishing fees when the reviewers submit their own papers.

The voucher system serves as an incentive for reviewers, providing them with tangible recognition for their work. It is a critical aspect of the open-access model, as it helps to ensure a constant supply of reviewers. This system also encourages the reviewers to submit their own work to the journal, creating a cycle of submission and review that sustains the journal.

Assumption 2.3.

We assume that the time at which the voucher is received by a reviewer depends on the decision they make. If the decision is either "Accept" or "Reject", the voucher is received immediately after the decision. If the decision is "Revision", the voucher is received after the revised paper has been reviewed.

Remark 2.3.

Given Assumption 2.3, the voucher value is effectively discounted over time in the case of a "Revision" decision due to the delay in receipt. Consequently, even though the ostensible amount of the voucher is the same for all three decisions, the present value of the voucher for a "Revision" decision is less than that for an "Accept" or "Reject" decision.

The temporal discounting of the voucher in case of a "Revision" decision introduces an interesting dynamic into the reviewer’s decision-making process. While the face value of the voucher remains the same, the delay in receiving it effectively reduces its value.

The quality of a paper is an inherent characteristic that can be assessed to varying degrees by different reviewers. We denote the true quality of a paper "p" as SpS_{p}, and we assume that each reviewer has a certain ability to assess this quality with some bias.

Definition 2.8.

The quality score of a paper pp, denoted as SpS_{p}, is an inherent attribute of the paper that indicates its suitability for publication. Higher scores represent higher quality.

The quality score of a paper, though a subjective measure, provides an essential yardstick for evaluation. It encapsulates the perceived merit of the paper, including aspects such as originality, rigor, and contribution to the field. This score, though subjective, is a critical component in the decision-making process of the reviewers.

Assumption 2.4.

The estimated quality score of a paper pp by reviewer ii, denoted as S^pi\hat{S}^{i}_{p}, is given by Sp+bpiS_{p}+b^{i}_{p}, where bpib^{i}_{p} is the bias of reviewer ii.

Assumption 2.5.

We assume that the estimated quality score S^pi\hat{S}^{i}_{p} of a paper by reviewer ii is private information, not shared among the reviewers.

This assumption is critical to our model and reflects the real-world situation in the peer review process. Typically, reviewers do not have access to each other’s evaluations until after they have submitted their own. This ensures the independence and integrity of the review process.

The estimated quality score S^pi\hat{S}^{i}_{p} encapsulates both the inherent quality of the paper SpS_{p} and the personal bias of the reviewer bpib^{i}_{p}. The inherent quality SpS_{p} is assumed to be a fixed attribute of the paper, known to all reviewers in an ideal world. However, the bias bpib^{i}_{p} reflects the individual judgment of the reviewer and can vary widely among different reviewers. This variation introduces uncertainty into the review process, as reviewers may disagree about the value of bpib^{i}_{p} and thus the quality of the paper S^pi\hat{S}^{i}_{p}.

Given the lack of common knowledge about S^pi\hat{S}^{i}_{p}, each reviewer must make their decision based on their individual perception of the paper’s quality, taking into consideration the potential decisions of the other reviewers. This forms the basis of the strategic interaction between reviewers in our model, as captured by the game 𝒢c\mathcal{G}_{c}. The individualistic nature of the S^pi\hat{S}^{i}_{p} estimation is what drives the need for the Nash equilibrium analysis. Reviewers must not only consider their perception of the paper’s quality but also anticipate the reactions of the other reviewers to maximize their payoff.

Definition 2.9.

We define S^pi\hat{S}^{i}_{p} as the estimated quality score of the paper pp by reviewer ii, and bpib^{i}_{p} as the bias of reviewer ii. The bias represents the reviewer’s personal judgment, which may deviate from the true quality score of the paper.

Remark 2.4.

The bias term bpib^{i}_{p} is a reflection of the inherent subjectivity in the review process. It may depend on factors such as the reviewer’s expertise, personal preferences, or even their mood at the time of review. In real-world scenarios, this bias term can never be fully eliminated. However, by making the reward system more efficient, we can motivate reviewers to minimize the impact of their bias, which in turn improves the overall quality of the review process.

It is generally accepted that there are certain thresholds in the quality score that determine whether a paper should be accepted, requires revision, or should be rejected. We denote these thresholds as ThaTh_{a} and ThrTh_{r}, respectively.

Definition 2.10.

Let ThaTh_{a} and ThrTh_{r} be the quality score thresholds for acceptance and revision, respectively. A paper is deemed suitable for acceptance if its quality score is greater than ThaTh_{a}; it requires revision if its quality score is between ThaTh_{a} and ThrTh_{r}; and it is deemed suitable for rejection if its quality score is less than ThrTh_{r}.

The quality score thresholds provide a quantitative benchmark against which the merit of a paper can be gauged. The thresholds effectively segregate papers into distinct categories, allowing for a more streamlined review process. These thresholds, though arbitrary, serve to uphold the standards of the journal.

Assumption 2.6.

We assume that the thresholds ThaTh_{a} and ThrTh_{r} are such that Tha>ThrTh_{a}>Th_{r}. This assumption is made to ensure that papers of higher quality are more likely to be accepted, while papers of lower quality are more likely to be rejected.

Assumption 2.7.

We assume that the thresholds ThaTh_{a} and ThrTh_{r} are common knowledge among reviewers.

Remark 2.5.

Given Assumption 2.7, each reviewer makes their decision based on S^pi\hat{S}^{i}_{p}, their own biased estimate of the paper’s quality. This is their best strategy to maximize their expected voucher value, as they are trying to estimate the other reviewer’s decision.

Assumption 2.8.

Although reviewers’ decisions are influenced by several factors, in this paper, we assume that each reviewer aims to maximize their voucher value.

In the ensuing examination, we delve deeper into the implications of the postulated Assumption 2.8, which asserts that a reviewer’s primary objective is to maximize their voucher value. This assumption is made to provide a framework for analyzing the effect of the voucher system on the decision-making process of reviewers. It is pivotal to underscore that this assumption is a simplification of reality, as it abstracts away from other considerations that might influence a reviewer’s decision.

Remark 2.6.

The assumption elucidated in Assumption 2.8, that a reviewer’s decision-making is predominantly driven by the intent to augment their voucher value, serves primarily as a tool of simplification for the complexity inherent in the system under scrutiny. It is essential to interpret this assumption within the context of the theoretical framework we adopt in this paper, and not as a blanket statement on the motives of every reviewer in the open-access journal landscape.

We take a moment to emphasize the role of our assumption, as specified in Assumption 2.8 and Remark 2.6, within the broader scope of our study. Our focus is on unveiling the strategic implications of the reward system, a task for which the assumption of reward maximization provides a tractable analytical lens. This assumption is not intended to be a definitive or comprehensive characterization of reviewers’ motivations, which, in reality, are multifaceted and may encompass a commitment to scholarly rigor, the advancement of their field, and upholding the integrity of the scientific method.

Remark 2.7.

It is incumbent upon us, in our quest for mathematical rigor, to acknowledge the limits of the assumptions we make. In Assumption 2.8, we posit a certain degree of economic rationality in reviewers’ behaviors. This does not, however, negate the potential for other non-economic factors to influence their decisions. Indeed, many reviewers are guided by a sense of duty to their academic community, and their decisions reflect a careful evaluation of the manuscript’s quality and contribution to knowledge. Our assumption, therefore, is an abstraction that aids in the exploration of certain strategic dynamics within the reviewer reward system, rather than a comprehensive model of reviewer behavior.

The underpinning assumption, Assumption 2.8, can be perceived as a behavioral model, an economic perspective on the decision-making process of reviewers. The assumption postulates that each reviewer aims to maximize their voucher value, i.e., reviewers are essentially trying to maximize the expected present value of the voucher they receive. It is worth pointing out that this assumption does not necessarily signify that reviewers prioritize their self-interest over the integrity of the review process. Instead, it provides an economic rationale for the reviewer’s decision-making process, which complements the existing understanding that the reviewer’s decision is primarily driven by their assessment of the paper’s quality.

In the light of this assumption, the decision-making process of a reviewer can be modeled as a form of strategic interaction where each reviewer makes their decision considering the potential decisions of the other reviewer, the quality of the paper, and the voucher system’s time-dependent nature. The reviewer’s decision-making process thus becomes an intricate dance between maintaining the integrity of the review process, assessing the quality of the paper, and maximizing the expected voucher value.

In the context of our study, the time-dependent nature of the voucher system introduces a dynamic that further complicates the reviewer’s decision-making process. It is important to note that, under Assumption 2.3, the "Revision" decision effectively discounts the value of the voucher due to the time delay in its receipt, even though the face value of the voucher is the same for all three decisions. Hence, when a reviewer is trying to maximize their voucher value, they might be incentivized to choose either "Accept" or "Reject" over "Revision".

3 Current Reviewer Reward System: A Game-Theoretic Analysis

3.1 Model Formulation

We begin the analysis by modeling the current reviewer reward system as a game. Recall from Section 2 that a game is defined by a triplet (𝒫,𝒮,𝒰)(\mathcal{P},\mathcal{S},\mathcal{U}), where 𝒫\mathcal{P} is the set of players, 𝒮\mathcal{S} is the strategy space, and 𝒰\mathcal{U} is the set of utility functions.

Definition 3.1.

The reviewer reward system game under the current system, denoted 𝒢c\mathcal{G}_{c}, is defined as:

  • 𝒫=1,2\mathcal{P}={1,2}, the set of two reviewers.

  • 𝒮=𝒟×𝒟\mathcal{S}=\mathcal{D}\times\mathcal{D}, where 𝒟={Accept,Revision,Reject}\mathcal{D}=\{\text{Accept},\text{Revision},\text{Reject}\} is the set of decisions a reviewer can make (refer to Definition 2.6).

  • 𝒰=(U1,U2)\mathcal{U}=(U_{1},U_{2}), where Ui:𝒮U_{i}:\mathcal{S}\rightarrow\mathbb{R} is the utility function of reviewer ii.

In this game, each reviewer independently chooses a decision d𝒟d\in\mathcal{D} based on their evaluation of the paper’s quality. The utility of a reviewer is determined by the voucher value they receive, which depends on the decision they make and the time at which they receive the voucher (refer to Assumption 2.3 and Remark 2.3).

An example of the payoff matrix of this game, denoted c\mathcal{M}_{c}, is as follows:

c=Reviewer 1AcceptRevisionRejectAccept100,100100dt,100dt100dt,100dtReviewer 2Revision100dt,100dt100dt,100dt100dt,100dtReject100dt,100dt100dt,100dt100,100\mathcal{M}_{c}=\begin{array}[]{cc|ccc}&&&\textbf{Reviewer 1}&\\ &&\text{Accept}&\text{Revision}&\text{Reject}\\ \hline\cr&\text{Accept}&100,100&100-d_{t},100-d_{t}&100-d_{t},100-d_{t}\\ \textbf{Reviewer 2}&\text{Revision}&100-d_{t},100-d_{t}&100-d_{t},100-d_{t}&100-d_{t},100-d_{t}\\ &\text{Reject}&100-d_{t},100-d_{t}&100-d_{t},100-d_{t}&100,100\end{array} (2)

where the first number in each cell is the utility of reviewer 1, the second number is the utility of reviewer 2, and dtd_{t} represents the time discount due to the delay in voucher receipt in the case of a "Revision" decision.

For instance, to calculate dtd_{t}, we consider the daily interest rate rdr_{d} and the number of days of revision ndn_{d}. Specifically, dtd_{t} can be computed as dt=(1+rd)ndd_{t}=(1+r_{d})^{n_{d}}. This formulation captures the impact of delayed voucher receipt on the utilities of both reviewers in the "Revision" decision scenario.

The reviewers aim to maximize their utility, which, in this game, is equivalent to receiving the highest possible voucher value in the shortest possible time. They do this by estimating the other reviewer’s decision based on their own assessment of the paper’s quality, and choosing a decision that they believe is most likely to be in agreement with the other reviewer’s decision. This is formalized in the following proposition:

Proposition 3.1.

In the game 𝒢c\mathcal{G}_{c}, the dominant strategy for reviewer ii is to choose a decision d𝒟d\in\mathcal{D} such that d=argmaxdUi(d,d^i)d=\arg\max_{d^{\prime}}U_{i}(d^{\prime},\hat{d}^{-i}), where d^i\hat{d}^{-i} is reviewer ii’s estimation of the other reviewer’s decision.

Proof.

Given the payoff matrix c\mathcal{M}_{c} (Equation 2), the utility for reviewer ii when they choose decision dd and the other reviewer chooses d^i\hat{d}^{-i} is Ui(d,d^i)U_{i}(d,\hat{d}^{-i}). By choosing d=argmaxdUi(d,d^i)d=\arg\max_{d^{\prime}}U_{i}(d^{\prime},\hat{d}^{-i}), reviewer ii is maximizing their utility, given their estimation of the other reviewer’s decision. This proves that choosing such a dd is a dominant strategy for reviewer ii. ∎

3.2 Nash Equilibrium Analysis

We now turn our attention to the analysis of Nash equilibriums for the game 𝒢c\mathcal{G}_{c} defined in Definition 3.1. Nash equilibriums, as defined in Definition 2.2, are strategy profiles from which no player has an incentive to unilaterally deviate, given the strategies of the other players.

Assumption 3.1.

Each reviewer ii’s strategy is to choose the decision d𝒟d\in\mathcal{D} that maximizes their expected utility, given their estimated quality score S^pi\hat{S}^{i}_{p} and the thresholds ThaTh_{a} and ThrTh_{r} from Assumption 2.6.

Let’s denote the set of strategies that maximize the expected utility of reviewer ii given their estimated quality score S^pi\hat{S}^{i}_{p} as Si(S^pi)S^{i}(\hat{S}^{i}_{p}). Reviewer ii’s optimal strategy sis^{*}_{i} is then a member of this set.

Lemma 3.1.

Given the thresholds ThaTh_{a} and ThrTh_{r}, a strategy profile (s1,s2)(s^{*}_{1},s^{*}_{2}) is a Nash equilibrium of the game 𝒢c\mathcal{G}_{c} if and only if siSi(S^pi)s^{*}_{i}\in S^{i}(\hat{S}^{i}_{p}) for i{1,2}i\in\{1,2\}.

Proof of this lemma follows directly from Assumption 3.1 and the definition of a Nash equilibrium (Definition 2.2). In particular, if siSi(S^pi)s^{*}_{i}\in S^{i}(\hat{S}^{i}_{p}), then by definition, there does not exist a strategy siSis_{i}\in S_{i} such that Ui(si,si)>Ui(si,si)U_{i}(s_{i},s^{*}_{-i})>U_{i}(s^{*}_{i},s^{*}_{-i}), which means that (s1,s2)(s^{*}_{1},s^{*}_{2}) satisfies the condition for a Nash equilibrium.

Theorem 3.1.

For all SpS_{p}\in\mathbb{R}, there exists a Nash equilibrium (s1,s2)(s^{*}_{1},s^{*}_{2}) only in (Accept, Accept) and (Reject, Reject) in the game 𝒢c\mathcal{G}_{c}.

Proof.

The proof of Theorem 3.1 can be completed through an exhaustive analysis of the strategies available to each player under the conditions given by the assumptions and definitions. We aim to show that, for all SpS_{p}\in\mathbb{R}, the only Nash equilibrium strategies are (Accept, Accept) and (Reject, Reject).

Recall the strategy sets Si(S^pi)S^{i}(\hat{S}^{i}_{p}) and the optimal strategies siSi(S^pi)s^{*}_{i}\in S^{i}(\hat{S}^{i}_{p}) as defined in Lemma 3.1. These strategy sets are determined by the thresholds ThaTh_{a} and ThrTh_{r} as well as the estimated quality scores S^pi\hat{S}^{i}_{p} of the paper pp by reviewer ii.

Assume that the thresholds ThaTh_{a} and ThrTh_{r} are fixed. Then, the estimated quality score S^pi\hat{S}^{i}_{p} of the paper pp by reviewer ii uniquely determines sis^{*}i. This is because S^pi\hat{S}^{i}_{p} is an increasing function of Sp+bpiS_{p}+b^{i}_{p}, and the reviewers will choose the decision that maximizes their expected utility.

Consider the three possible decisions for each player: Accept, Revision, and Reject. We will discuss each possibility in turn.

  • If S^pi>Tha\hat{S}^{i}_{p}>Th_{a}, the reviewers will choose to accept the paper. In this case, both reviewers are in agreement, and the Nash equilibrium is (Accept, Accept).

  • If Tha>S^pi>ThrTh_{a}>\hat{S}^{i}_{p}>Th_{r}, the reviewers may choose to revise the paper. However, considering the payoff matrix c\mathcal{M}_{c} (Equation 2), choosing to revise does not maximize the reviewers’ utility, as they face a time delay in receiving their vouchers, represented by dtd_{t} in the payoff matrix. Instead, the reviewers will be motivated to unilaterally deviate from this decision to either accept or reject the paper. Therefore, (Revision, Revision) cannot be a Nash equilibrium.

  • If Thr>S^piTh_{r}>\hat{S}^{i}_{p}, the reviewers will choose to reject the paper. Both reviewers are again in agreement, and the Nash equilibrium is (Reject, Reject).

In conclusion, for any inherent quality score SpS_{p}\in\mathbb{R}, the only Nash equilibrium strategies in the game 𝒢c\mathcal{G}_{c} are (Accept, Accept) and (Reject, Reject). ∎

Example 3.1.

Consider a scenario where Sp=0.7S_{p}=0.7, bp1=0.5b^{1}_{p}=-0.5, and bp2=0.1b^{2}_{p}=0.1. Then, S^p1=0.65\hat{S}^{1}_{p}=0.65 and S^p2=0.80\hat{S}^{2}_{p}=0.80. Assuming Tha=0.75Th_{a}=0.75 and Thr=0.50Th_{r}=0.50, reviewer 1 would likely choose to revise the paper, while reviewer 2 would likely choose to accept it. However, to maximize their utility, they would try to align their decisions. Depending on their risk tolerance and other factors, they might both choose to revise or accept. Both of these strategy profiles, (Revision, Revision) and (Accept, Accept), could be Nash equilibriums of this game. Note that the choice between these equilibriums could significantly impact the outcome for the paper.

Remark 3.1.

In the current reviewer reward system, the Nash equilibrium could incentivize reviewers to align their decisions, potentially leading to less accurate evaluations of papers. This could be particularly problematic if the inherent quality score of a paper is close to the threshold for acceptance or rejection, as in the case of Example 3.1. This highlights the potential need for a more effective reviewer reward system, which we will explore in the next section.

Thus, our analysis shows that under the current reviewer reward system, there exists a Nash equilibrium for each paper. However, the alignment of reviewer decisions in these equilibriums could potentially compromise the quality of the review process.

4 Proposed Reviewer Reward System: A Game-Theoretic Analysis

4.1 Model Formulation

Having analyzed the current system, we now turn our attention to the proposed reviewer reward system, which we model as a strategic game.

Definition 4.1.

The reviewer reward system game under the proposed system, denoted as 𝒢p\mathcal{G}_{p}, is defined as follows:

  • 𝒫=1,2\mathcal{P}={1,2}, the set of two reviewers.

  • 𝒮=𝒟×𝒟\mathcal{S}=\mathcal{D}\times\mathcal{D}, where 𝒟=Accept,Revision,Reject\mathcal{D}={\text{Accept},\text{Revision},\text{Reject}} is the set of decisions a reviewer can make.

  • 𝒰=(U1,U2)\mathcal{U}=(U_{1},U_{2}), where Ui:𝒮U_{i}:\mathcal{S}\rightarrow\mathbb{R} is the utility function of reviewer ii.

Similar to the current system, each reviewer independently chooses a decision d𝒟d\in\mathcal{D} based on their evaluation of the paper’s quality. The main difference in the proposed system lies in the utility functions: the voucher value, which is the reviewers’ utility, varies not only with their decision and the timing of voucher receipt but also with the agreement between the reviewers’ decisions.

An example of the payoff matrix for this game, denoted p\mathcal{M}_{p}, is given by:

p=Reviewer 1AcceptRevisionRejectAccept200,200100,10050,50Reviewer 2Revision100,100200,200100,100Reject50,50100,100200,200\mathcal{M}_{p}=\begin{array}[]{cc|ccc}&&&\textbf{Reviewer 1}&\\ &&\text{Accept}&\text{Revision}&\text{Reject}\\ \hline\cr&\text{Accept}&200,200&100,100&50,50\\ \textbf{Reviewer 2}&\text{Revision}&100,100&200,200&100,100\\ &\text{Reject}&50,50&100,100&200,200\end{array} (3)

where the first number in each cell is the utility of reviewer 1 and the second number is the utility of reviewer 2.

Given the structure of the payoff matrix p\mathcal{M}_{p}, the reviewers are incentivized to make their decisions not only based on their own assessments but also with an eye towards agreement with the other reviewer. This leads to the following proposition:

Proposition 4.1.

In the game 𝒢p\mathcal{G}_{p}, the optimal strategy for reviewer ii is to choose a decision d𝒟d\in\mathcal{D} such that d=argmaxdUi(d,d^i)d=\arg\max_{d^{\prime}}U_{i}(d^{\prime},\hat{d}^{-i}), where d^i\hat{d}^{-i} is reviewer ii’s estimation of the other reviewer’s decision.

Proof.

The proof of Proposition 4.1 is similar to that of Proposition 3.1 in Section 3.1. Given the payoff matrix p\mathcal{M}_{p} (Equation 3), the utility for reviewer ii when they choose decision dd and the other reviewer is the utility of reviewer 2. ∎

The underlying principles in evaluating the quality of the paper remain the same as in the previous game, as detailed in Assumption 2.6 and Definition 2.9. However, given the modified payoff matrix p\mathcal{M}_{p}, the strategy for each reviewer changes.

Proposition 4.2.

In the game 𝒢p\mathcal{G}_{p}, the strategy for reviewer ii is to choose a decision d𝒟d\in\mathcal{D} such that d=argmaxdUi(d,d^i)d=\arg\max_{d^{\prime}}U_{i}(d^{\prime},\hat{d}^{-i}), where d^i\hat{d}^{-i} is reviewer ii’s estimation of the other reviewer’s decision.

Proof.

Given the payoff matrix p\mathcal{M}_{p} (Equation 3), the utility for reviewer ii when they choose decision dd and the other reviewer chooses d^i\hat{d}^{-i} is Ui(d,d^i)U_{i}(d,\hat{d}^{-i}). By choosing d=argmaxdUi(d,d^i)d=\arg\max_{d^{\prime}}U_{i}(d^{\prime},\hat{d}^{-i}), reviewer ii is maximizing their utility, given their estimation of the other reviewer’s decision. This proves that choosing such a dd is a rational strategy for reviewer ii in the proposed system. ∎

Remark 4.1.

We observe that the change in the payoff matrix affects the decision-making process of the reviewers. Reviewers are now incentivized to more accurately assess the paper and coordinate their decision with the other reviewer, as the highest reward is received when both make the same decision.

Definition 4.2.

We define the reward function R:𝒮R:\mathcal{S}\rightarrow\mathbb{R} for the proposed system as R(d1,d2)=p[d1,d2]R(d_{1},d_{2})=\mathcal{M}_{p}[d_{1},d_{2}], where d1,d2𝒟d_{1},d_{2}\in\mathcal{D} are the decisions of the reviewers.

In the proposed system, the reward function is symmetric, i.e., R(d1,d2)=R(d2,d1)R(d_{1},d_{2})=R(d_{2},d_{1}) for all d1,d2𝒟d_{1},d_{2}\in\mathcal{D}. This symmetry in rewards encourages cooperation between reviewers, driving them to reach a consensus on the quality of the paper. The proposed system thus fosters a more collaborative and quality-focused review process.

Assumption 4.1.

We assume that reviewer ii believes that the other reviewer, ii^{\prime}, makes their decision based on their estimated quality score, S^pi\hat{S}^{i^{\prime}}_{p}.

This assumption adds another layer of complexity to the reviewers’ decision-making process. Under this assumption, reviewer ii must not only consider their own evaluation of the paper, but also estimate the decision of the other reviewer, ii^{\prime}, based on their own quality assessment.

Given that reviewer ii only has the paper as a source of information, the most reasonable estimation they can make about S^ip\hat{S}^{i^{\prime}}{p} is that it equals their own estimated quality score, S^pi\hat{S}^{i}_{p}. This leads to the assumption that S^piS^pi\hat{S}^{i^{\prime}}_{p}\simeq\hat{S}^{i}_{p}. In other words, in the absence of any other information, reviewer ii assumes that the other reviewer’s evaluation of the paper aligns with their own.

This assumption holds some basis in reality. When a reviewer is evaluating a paper, they may naturally assume that another reviewer, with a similar level of expertise and understanding, would evaluate the paper similarly. Indeed, this may not always be the case due to differing biases and perspectives, but in the absence of other information, this serves as a reasonable assumption.

Under this assumption, each reviewer tries to estimate the other’s decision based on their own assessment of the paper. Consequently, this encourages reviewers to more closely scrutinize the paper, as their reward is now dependent on not only their own decision but also on accurately predicting the other reviewer’s decision.

Therefore, in the proposed system, we see that the incentives are designed in such a way as to encourage both accurate evaluation of the paper and coordination between reviewers. This could potentially lead to an increase in the quality of the reviews, and ultimately, of the papers accepted by the journal.

4.2 Nash Equilibrium Analysis

We start by considering the Nash Equilibrium of the proposed game 𝒢p\mathcal{G}_{p}.

Definition 4.3.

A strategy profile (s1,s2)(s^{*}_{1},s^{*}_{2}) is a Nash equilibrium of the proposed game 𝒢p\mathcal{G}_{p} if and only if for every player i{1,2}i\in\{1,2\}, we have Ui(si,si)Ui(si,si)U_{i}(s^{*}_{i},s^{*}_{-i})\geq U_{i}(s_{i},s^{*}_{-i}) for all siSis_{i}\in S_{i}.

This definition follows directly from the standard definition of Nash equilibrium. In particular, a strategy profile is a Nash equilibrium if no player can unilaterally improve their utility by deviating from their current strategy.

Proposition 4.3.

In the proposed game 𝒢p\mathcal{G}_{p}, a strategy profile (s1,s2)(s^{*}_{1},s^{*}_{2}) is a Nash equilibrium if and only if siSi(S^pi)s^{*}_{i}\in S^{i}(\hat{S}^{i}_{p}) for i{1,2}i\in\{1,2\}, where Si(S^pi)S^{i}(\hat{S}^{i}_{p}) is the set of optimal strategies for player ii given their estimated quality score S^pi\hat{S}^{i}_{p}.

Proof.

The proof of Proposition 4.3 is similar to that of Lemma 3.1 for the current system. Specifically, if siSi(S^pi)s^{i}\in S^{i}(\hat{S}^{i}_{p}), then by definition of Si(S^pi)S^{i}(\hat{S}^{i}_{p}), there does not exist a strategy siSis_{i}\in S_{i} such that Ui(si,si)>Ui(si,si)U_{i}(s_{i},s^{*}_{-i})>U_{i}(s^{*}_{i},s^{*}_{-i}), which means that (s1,s2)(s^{*}_{1},s^{*}_{2}) satisfies the condition for a Nash equilibrium. ∎

Theorem 4.1.

For all SpS_{p}\in\mathbb{R}, there exists a Nash equilibrium (s1,s2)(s^{*}_{1},s^{*}_{2}) in the game 𝒢p\mathcal{G}_{p}.

Proof.

Given SpS_{p}, the estimated quality score by each reviewer ii is S^pi=Sp+bpi\hat{S}^{i}_{p}=S_{p}+b^{i}_{p}. From Proposition 4.2, we know that each reviewer ii chooses a decision d𝒟d\in\mathcal{D} that maximizes their expected utility, which results in a strategy siSi(S^pi)s^{i}\in S^{i}(\hat{S}^{i}_{p}). From Proposition 4.3, we know that if siSi(S^pi)s^{i}\in S^{i}(\hat{S}^{i}_{p}) for i{1,2}i\in\{1,2\}, then (s1,s2)(s^{*}_{1},s^{*}_{2}) is a Nash equilibrium of the game 𝒢p\mathcal{G}_{p}. Therefore, we can conclude that for all SpS_{p}\in\mathbb{R}, there exists a Nash equilibrium (s1,s2)(s^{*}_{1},s^{*}_{2}) in the game 𝒢p\mathcal{G}_{p}. ∎

Remark 4.2.

In the proposed game 𝒢p\mathcal{G}_{p}, the strategy space SiS_{i} of a player ii is the set of all possible decisions, Si={Accept,Revision,Reject}S_{i}=\{Accept,Revision,Reject\}. However, the effective strategy space Si(S^pi)S^{i}(\hat{S}^{i}_{p}) can be much smaller depending on the estimated quality score S^pi\hat{S}^{i}_{p}. In particular, if S^pi\hat{S}^{i}_{p} is close to either end of the spectrum, the effective strategy space can be reduced to a singleton set, making the player’s decision trivial.

Example 4.1.

Consider a paper with an inherent quality score Sp=0.7S_{p}=0.7. Let’s assume that reviewer 1 and reviewer 2 have biases bp1=0.1b^{1}_{p}=0.1 and bp2=0.1b^{2}_{p}=-0.1, respectively. Therefore, the estimated quality scores by the two reviewers are S^p1=0.8\hat{S}^{1}_{p}=0.8 and S^p2=0.6\hat{S}^{2}_{p}=0.6. If the thresholds are Tha=0.75Th_{a}=0.75 and Thr=0.5Th_{r}=0.5, then the optimal strategy for reviewer 1 is to accept the paper, while the optimal strategy for reviewer 2 is to suggest a revision. In this case, the Nash equilibrium of the game is (Accept,Revision)(Accept,Revision), which results in a reward of 100 for each reviewer according to the proposed reward system.

Definition 4.4.

A strategy profile (s1,s2)(s^{*}_{1},s^{*}_{2}) is Pareto efficient in the proposed game 𝒢p\mathcal{G}_{p} if and only if there does not exist another strategy profile (s1,s2)(s_{1},s_{2}) such that Ui(si,si)Ui(si,si)U_{i}(s_{i},s_{-i})\geq U_{i}(s^{*}_{i},s^{*}_{-i}) for all i{1,2}i\in\{1,2\}, and Ui(si,si)>Ui(si,si)U_{i}(s_{i},s_{-i})>U_{i}(s^{*}_{i},s^{*}_{-i}) for at least one ii.

Proposition 4.4.

In the proposed game 𝒢p\mathcal{G}_{p}, a strategy profile (s1,s2)(s^{*}_{1},s^{*}_{2}) is Pareto efficient if and only if si=argmaxsiSiUi(si,si)s^{*}_{i}=\arg\max_{s_{i}\in S_{i}}U_{i}(s_{i},s^{*}_{-i}) for i{1,2}i\in\{1,2\}.

The analysis of the Nash equilibria in the proposed game 𝒢p\mathcal{G}_{p} reveals that the game encourages reviewers to make accurate estimations of the quality of a paper and to choose their decisions accordingly.

5 Simulations and Results

In this section, we explore the characteristics and attributes of the existing reviewer reward system by conducting simulations. The simulations are performed using a straightforward deep reinforcement learning approach, in which two agents are employed to represent the reviewers. The primary objective of these agents is to maximize their anticipated utility within the game 𝒢c\mathcal{G}_{c}, as defined in Definition 3.1. It is important to note that the focus of this paper does not lie in the intricate details of the deep reinforcement model architecture; hence, we employ a simplified model for our simulations.

5.1 Simulation Settings of Current Reviewer Reward System

The two agents are modelled as neural networks with a single input, three outputs, and a hidden layer of size 256. The input to each agent is the estimated quality score S^pi\hat{S}^{i}_{p} of a paper, while the output represents the probability distribution over the set of possible decisions 𝒟={Accept,Revision,Reject}\mathcal{D}=\{\text{Accept},\text{Revision},\text{Reject}\}. The structure of the agents’ neural networks is depicted in Figure 1.

Refer to caption
Figure 1: Deep reinforcement learning structure of the agents.

In addition, each neural network employs a series of ten residual modules [23] to facilitate the training process. These modules are designed to learn identity functions that allow the input signals to bypass some layers and reach deeper parts of the network directly, effectively addressing the issue of vanishing gradients commonly observed in deep neural networks [24].

The agents’ neural networks are trained using the Adam optimizer [25] with a learning rate of 0.000010.00001, and the training process lasts for 100,000100,000 epochs. The reward at each epoch is calculated based on whether the decisions made by the agents are in agreement and whether these decisions are correct. Specifically, the agents receive a reward of 1.01.0 if they both make the correct decision (either accepting or rejecting the paper), and a reward of 1.0dt1.0-d_{t} if they make different decisions or both decide to request a revision. The quantity dtd_{t} is a time discount factor set to 0.20.2, which represents the delay in voucher receipt for a "Revision" decision.

In order to ensure that the agents’ decisions are not solely based on the maximization of immediate reward, an auxiliary loss term is included in the loss function. This term is calculated as the cross-entropy between the agents’ decisions and the true decision, and serves to guide the agents towards making decisions that are consistent with the actual quality of the papers.

The quality score SpS_{p} and the biases bpib^{i}_{p} are randomly generated for each epoch. The quality score is drawn from a uniform distribution between 0 and 11, while the biases are drawn from a uniform distribution between 0.05-0.05 and 0.050.05. The thresholds ThaTh_{a} and ThrTh_{r} for making "Accept" and "Revision" decisions are set to 0.660.66 and 0.330.33, respectively, in order to have the same probabilities of "Accept", "Revision", and "Reject" for a random paper, in accordance with Assumptions 2.6 and 2.7.

In order to observe the evolution of the agents’ decision-making process during the training, we record the moving average of the rewards received by the agents, as well as the distribution of their decisions. The moving average reward is calculated over a sliding window of 100100 epochs, and the decision distribution is computed by counting the number of times each pair of decisions is made by the agents.

5.2 Definition of Reward and Loss Functions

Reward Function

The reinforcement learning framework for our simulation necessitates a reward function, which facilitates the process of decision-making. As the agents interact with the environment, they receive rewards that depend on the actions they take. The reward function we define here is constructed to encourage a pair of decisions that are both accurate and concordant.

We denote the reward at time tt as RtR_{t}. Let dti𝒟d^{i}_{t}\in\mathcal{D} represent the decision made by agent ii at time tt, where 𝒟={Accept,Revision,Reject}\mathcal{D}=\{\text{Accept},\text{Revision},\text{Reject}\}. The reward is defined as follows:

Rt={1.0if dt1=dt2 and (dti=Accept or dti=Reject),1.0dtotherwise.R_{t}=\begin{cases}1.0&\text{if }d^{1}_{t}=d^{2}_{t}\text{ and }(d^{i}_{t}=\text{Accept}\text{ or }d^{i}_{t}=\text{Reject}),\\ 1.0-d_{t}&\text{otherwise}.\end{cases} (4)

The quantity dtd_{t} is a time discount factor set to 0.20.2, which represents the delay in voucher receipt for a "Revision" decision. This reward structure encourages the agents to agree with each other and to make correct decisions.

Loss Function

Training our agents involves minimizing a loss function, which is composed of two parts: the expected reward and an auxiliary term. The expected reward, 𝔼[Rt]-\mathbb{E}[R_{t}], encourages the agents to make decisions that maximize their reward. The auxiliary term, LtauxL^{aux}_{t}, is introduced to ensure that the agents’ decisions align with the assessed quality of the papers. This auxiliary loss term is calculated as the cross-entropy between the agents’ decisions and the true decision.

For agent ii, the loss function at time tt is defined as follows:

Lti=𝔼[Rt]+λLtaux,L^{i}_{t}=-\mathbb{E}[R_{t}]+\lambda L^{aux}_{t}, (5)

where λ\lambda is a hyperparameter balancing the contribution of the auxiliary term. In our implementation, we set λ=0.05\lambda=0.05.

The auxiliary loss for agent ii is calculated as:

Ltaux=logpti(dti),L^{aux}_{t}=-\log p^{i}_{t}(d^{i}_{t}), (6)

where pti(dti)p^{i}_{t}(d^{i}_{t}) is the probability assigned by agent ii to its decision dtid^{i}_{t} at time tt.

Remark 5.1.

We use the cross-entropy loss as our auxiliary term because it measures the dissimilarity between the predicted probability distribution and the target distribution. By minimizing this term, we encourage our agents to make decisions that align with the actual quality of the papers.

5.3 Simulation Settings of Proposed Reviewer Reward System

The agents’ architectures and the simulation framework remain the same for the proposed reviewer reward system as in the current system, as detailed in Subsection 5.1. However, the crucial distinction lies in the reward function, which we redefine in this context to better incentivize the reviewers’ decision-making process.

Definition 5.1 (Proposed Reward Function).

Given that dti𝒟d^{i}_{t}\in\mathcal{D} represents the decision made by agent ii at time tt, where 𝒟=Accept,Revision,Reject\mathcal{D}={\text{Accept},\text{Revision},\text{Reject}}, we define the new reward at time tt as RtR^{\prime}_{t}. This reward is characterized as follows:

Rt={1.0if dt1=dt2,0.5if (dti,dti)=(Accept, Revision) or (Revision, Reject),0.2otherwise.R^{\prime}_{t}=\begin{cases}1.0&\text{if }d^{1}_{t}=d^{2}_{t},\\ 0.5&\text{if }(d^{i}_{t},d^{i^{\prime}}_{t})=(\text{Accept, Revision})\text{ or }(\text{Revision, Reject}),\\ 0.2&\text{otherwise}.\end{cases} (7)
Remark 5.2.

In the proposed reward function, the agents get a reward of 1.01.0 if their decisions agree and are correct, a reward of 0.50.5 if their decisions are off by one category, and a reward of 0.20.2 otherwise. This grading system is designed to encourage agreement among the agents, while at the same time accommodating minor discrepancies.

We employ this newly formulated reward function in our simulation, replacing the former reward function detailed in Subsection 5.2. The loss function defined in Equation 5 remains unaltered for this experiment, as it provides a stable and consistent means of optimizing the agents’ actions.

Refer to caption
Figure 2: Results of the current reviewer reward system. (A) Rewards over epoch. (B) Decision distributions of Reviewer 1 and 2.

5.4 Outcome Analysis and Interpretation

The results of the simulations, as showcased in Figures 2 and 3, offer valuable insights into the performance of the current and proposed reviewer reward systems, respectively. In this subsection, we delve into these outcomes to distill the key findings and implications.

After a sequence of 20,000 epochs, the agents stabilize towards a Nash equilibrium, as discerned from the fact that the rewards settle around the value of 0.950.95 and do not exhibit significant variations henceforth. The existence of Nash equilibria is predominantly evident in the pairings of (Accept, Accept) and (Reject, Reject), as illustrated in Figure 2. In particular, the (Accept, Accept) combination is realized with a frequency of 31.40%31.40\%, and the (Reject, Reject) pairing manifests itself with a frequency of 37.20%37.20\%. Conversely, the (Revision, Revision) pairing is less common, with a likelihood of 10.60%10.60\%. This pattern indicates that under the current reward system, reviewers have a propensity to adopt a binary decision-making approach, favoring either Accept or Reject verdicts.

In stark contrast, the proposed reward system demonstrates a more balanced distribution of decision pairings. As evidenced in Figure 3, Nash equilibria can be found in all decision pairings, including (Accept, Accept), (Revision, Revision), and (Reject, Reject). In terms of percentages, these pairings are realized with probabilities of 31.00%31.00\%, 28.20%28.20\%, and 25.60%25.60\%, respectively. Given the thresholds ThaTh_{a} and ThrTh_{r} set at 0.660.66 and 0.330.33 for "Accept" and "Revision" decisions, the ideal probabilities for the respective judgments ought to be around 33.33%33.33\%. Consequently, the proposed reward system is more conducive to achieving this desirable state, demonstrating a more balanced approach to the review process compared to the current system.

We begin by observing the deviation of decision pairing frequencies from the idealized probabilities under both systems. For the current system, the deviation for (Accept, Accept) pairing, denoted as ΔAAc\Delta_{AA}^{c}, can be calculated as |31.40%33.33%|=1.93%|31.40\%-33.33\%|=1.93\%, and similarly, ΔRRc=3.87%\Delta_{RR}^{c}=3.87\%, and ΔRevRevc=22.73%\Delta_{RevRev}^{c}=22.73\%. On the other hand, the proposed system exhibits the deviations as ΔAAp=2.33%\Delta_{AA}^{p}=2.33\%, ΔRRp=7.73%\Delta_{RR}^{p}=7.73\%, and ΔRevRevp=5.13%\Delta_{RevRev}^{p}=5.13\%. The significant reduction in deviation for the (Revision, Revision) pairing from 22.73%22.73\% to 5.13%5.13\% suggests a more balanced distribution of decision pairings under the proposed system.

Refer to caption
Figure 3: Results of the proposed reviewer reward system. (A) Rewards over epoch. (B) Decision distributions of Reviewer 1 and 2.

A deeper analysis of these deviations reveals a noteworthy trend. Let Δmaxc=max{ΔAAc,ΔRRc,ΔRevRevc}\Delta_{max}^{c}=\max\{\Delta_{AA}^{c},\Delta_{RR}^{c},\Delta_{RevRev}^{c}\} and Δmaxp=max{ΔAAp,ΔRRp,ΔRevRevp}\Delta_{max}^{p}=\max\{\Delta_{AA}^{p},\Delta_{RR}^{p},\Delta_{RevRev}^{p}\} be the maximum deviations under the current and proposed systems, respectively. We observe that Δmaxc>Δmaxp\Delta_{max}^{c}>\Delta_{max}^{p}, indicating that the proposed system brings the decision pairings closer to the idealized probabilities.

Another crucial observation pertains to the stability of the Nash equilibrium. By investigating the standard deviations of the rewards, denoted as σc\sigma^{c} and σp\sigma^{p} for the current and proposed systems respectively, we find that σc>σp\sigma^{c}>\sigma^{p}. This suggests that the rewards are more stable under the proposed system, implying a more stable Nash equilibrium.

6 Conclusion

This study commenced with a comprehensive mathematical and game-theoretical examination of the reviewer reward systems prevalent in open-access journals. The existing system, construed as a strategic interaction between two reviewers, was discovered to engender a bias towards binary decisions. This bias was evidenced by the preponderance of Nash equilibria in (Accept, Accept) and (Reject, Reject) pairings, a trend that might imperil the thoroughness of peer reviews.

To counter this, we designed and examined an alternative reward system, intended to foster more balanced and comprehensive reviews. Through a combination of rigorous game-theoretical analysis and deep reinforcement learning simulations, we demonstrated the merits of our proposed system in fostering a more balanced and stable distribution of decisions. Our findings revealed that our proposed system nudges the distribution of decision pairings closer to the ideal probabilities, thereby promoting a more detailed review process.

However, it is crucial to acknowledge the inherent limitations of our model, which is predicated on Assumption 2.8. The presumption that reviewers primarily act to maximize their voucher value offers a simplified, yet insightful perspective, enabling us to isolate and scrutinize the influence of the reward system on decision-making. Yet, it overlooks a myriad of other factors that could potentially sway a reviewer’s decision-making process. We concede that our model may not fully encapsulate the complexity and diversity of motivations inherent in the real-world peer review process. Consequently, our study should be viewed as an initial step rather than an endpoint. It paves the way for further research, encouraging subsequent studies to develop more sophisticated models of reviewer behavior that take into account the balance between voucher value maximization and other influencing factors. Such an endeavor could possibly furnish a more holistic understanding of the dynamics in the peer review process, enhancing the models we propose here and improving their practical relevance.

Our research offers a novel viewpoint on the topic of reviewer rewards in open-access journals, supplying a solid mathematical foundation to steer future research and policy deliberations. Although our findings hold promise, additional efforts should be dedicated to testing our proposed system in real-world contexts and assessing its practicality and effectiveness. Moreover, our methodology could be broadened to scrutinize other facets of the peer review process, thereby enriching the mathematical comprehension of this indispensable aspect of scholarly publishing.

References

  • [1] Heather Piwowar, Jason Priem, Vincent Larivière, Juan Pablo Alperin, Lisa Matthias, Bree Norlander, Ashley Farley, Jevin West, and Stefanie Haustein. The state of oa: a large-scale analysis of the prevalence and impact of open access articles. PeerJ, 6:e4375, 2018.
  • [2] Alberto Martín-Martín, Rodrigo Costas, Thed Van Leeuwen, and Emilio Delgado López-Cózar. Evidence of open access of scientific publications in google scholar: A large-scale analysis. Journal of informetrics, 12(3):819–841, 2018.
  • [3] John Willinsky. Scholarly associations and the economic viability of open access publishing. Health Sciences: An OJS Theme Demo, 1(2), 2018.
  • [4] Multidisciplinary digital publishing institute. https://www.mdpi.com/. Accessed: [May 14, 2023].
  • [5] Pawel Ladosz, Lilian Weng, Minwoo Kim, and Hyondong Oh. Exploration in deep reinforcement learning: A survey. Information Fusion, 2022.
  • [6] Ngan Le, Vidhiwar Singh Rathour, Kashu Yamazaki, Khoa Luu, and Marios Savvides. Deep reinforcement learning in computer vision: a comprehensive survey. Artificial Intelligence Review, pages 1–87, 2022.
  • [7] Marcel Panzer and Benedict Bender. Deep reinforcement learning in production systems: a systematic literature review. International Journal of Production Research, 60(13):4316–4341, 2022.
  • [8] Zonghan Wu, Shirui Pan, Fengwen Chen, Guodong Long, Chengqi Zhang, and S Yu Philip. A comprehensive survey on graph neural networks. IEEE transactions on neural networks and learning systems, 32(1):4–24, 2020.
  • [9] Changliu Liu, Tomer Arnon, Christopher Lazarus, Christopher Strong, Clark Barrett, Mykel J Kochenderfer, et al. Algorithms for verifying deep neural networks. Foundations and Trends® in Optimization, 4(3-4):244–404, 2021.
  • [10] Yu Zhang, Peter Tiňo, Aleš Leonardis, and Ke Tang. A survey on neural network interpretability. IEEE Transactions on Emerging Topics in Computational Intelligence, 5(5):726–742, 2021.
  • [11] Patrick Bajari, Han Hong, and Denis Nekipelov. Game theory and econometrics: A survey of some recent research. In Advances in economics and econometrics, 10th world congress, volume 3, pages 3–52, 2013.
  • [12] Hai-Yan Shi, Wan-Liang Wang, Ngai-Ming Kwok, and Sheng-Yong Chen. Game theory for wireless sensor networks: a survey. Sensors, 12(7):9055–9097, 2012.
  • [13] Yan Zhou, Murat Kantarcioglu, and Bowei Xi. A survey of game theoretic approach for adversarial machine learning. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 9(3):e1259, 2019.
  • [14] Thomas C Schelling. Game theory: a practitioner’s approach. Economics & Philosophy, 26(1):27–46, 2010.
  • [15] The Anh Han, Arne Traulsen, and Chaitanya S Gokhale. On equilibrium properties of evolutionary multi-player games with random payoff matrices. Theoretical Population Biology, 81(4):264–272, 2012.
  • [16] Erol Akçay and Joan Roughgarden. The evolution of payoff matrices: providing incentives to cooperate. Proceedings of the Royal Society B: Biological Sciences, 278(1715):2198–2206, 2011.
  • [17] Karen M Albert. Open access: implications for scholarly publishing and medical libraries. Journal of the Medical Library Association, 94(3):253, 2006.
  • [18] David J Solomon and Bo-Christer Björk. A study of open access journals using article processing charges. Journal of the American Society for Information Science and Technology, 63(8):1485–1495, 2012.
  • [19] Bo-Christer Björk and David Solomon. Open access versus subscription journals: a comparison of scientific impact. BMC medicine, 10(1):1–10, 2012.
  • [20] Parveen Azam Ali and Roger Watson. Peer review and the publication process. Nursing open, 3(4):193–202, 2016.
  • [21] Carole J Lee, Cassidy R Sugimoto, Guo Zhang, and Blaise Cronin. Bias in peer review. Journal of the American Society for Information Science and Technology, 64(1):2–17, 2013.
  • [22] Tom Jefferson, Philip Alderson, Elizabeth Wager, and Frank Davidoff. Effects of editorial peer review: a systematic review. Jama, 287(21):2784–2786, 2002.
  • [23] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  • [24] Sandeep Paul, Lotika Singh, et al. A review on advances in deep learning. In 2015 IEEE Workshop on Computational Intelligence: Theories, Applications and Future Directions (WCI), pages 1–6. IEEE, 2015.
  • [25] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.