Thermodynamic optimization of finite-time feedback protocols for Markov jump systems
Abstract
In recent advances in finite-time thermodynamics, optimization of entropy production required for finite-time information processing is an important issue. In this work, we consider finite-time feedback processes in classical discrete systems described by Markov jump processes, and derive achievable bounds on entropy production for feedback processes controlled by Maxwell’s demons. The key ingredients of our approach is optimal transport theory and an achievable Fano’s inequality, by which we optimize the Wasserstein distance over final distributions under fixed consumed information. Our study reveals the minimum entropy production for consuming a certain amount of information, and moreover, the optimal feedback protocol to achieve it. These results are expected to lead to design principles for information processing in various stochastic systems with discrete states.
I Introduction
One of the key concepts in stochastic thermodynamics is entropy production [1, 2, 3], which quantifies entropy production accompanying irreversibility of dynamics. Its fundamental bound is imposed by the second law of thermodynamics, which can be achieved in the quasi-static limit requiring infinite time. Beyond the quasi-static regime, recent advancements in stochastic thermodynamics have established various bounds on entropy production by incorporating finite-time effects, represented by thermodynamic speed limits [4, 5, 6, 7, 8, 9, 10, 11] and thermodynamic uncertainty relations [12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22]. In particular, the speed limits based on optimal transport theory [23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36] provide achievable bounds for any finite operation times, by explicitly identifying the optimal protocols.
Recently, there has been progress in applying these finite-time frameworks to thermodynamics of information [37]. In the finite-time regime, the speed limits and the thermodynamic uncertainty relations have been extended to situations incorporating information processing [38, 39, 40, 41, 42, 43, 44, 45, 46, 47]. In particular, optimal transport theory provides achievable bounds for the finite-time entropy production for various information processing including measurement, feedback and information erasure, when the initial and final probability distributions are fixed [38, 41, 44, 46, 47]. However, optimizing the finite-time entropy productions for a given amount of processed information has been addressed only for measurement processes [45], and the optimal entropy productions for consuming a given amount of information through feedback processes have not been elucidated.
In this study, we consider optimization of feedback processes in classical discrete systems obeying Markov jump processes. Specifically, we determine the optimal entropy production for fixed consuming mutual information . The optimization consists of two stages. The first stage is based on optimal transport theory, where we optimize entropy production over time-dependent protocols under fixed initial and identify final distributions [36]. In the second stage, we further optimize entropy production over final distributions under fixed consumed information , and obtain the achievable lower bounds on entropy production and the optimal final distributions. These are derived by exploiting an achievable Fano’s inequality [48], which bounds the conditional Shannon entropy by a function of error rate. Our results are relevant to experimental platforms such as single electron systems [49, 50, 51].
The organization of this paper is as follows. In Sec. II, we describe our setup along with a brief review of thermodynamics of information, optimal transport theory and thermodynamical speed limits. In Sec. III, we introduce an achievable Fano’s inequality as a mathematical tool, and describe our main mathematical Theorem, and provide its proof. In Sec. IV, we present two bounds on entropy production of feedback process as physical consequences of the main Theorem. We also identify the optimal protocols to achieve these bounds. In Sec. V, we numerically demonstrate our bounds and the optimal protocols by a coupled two-level system. In Sec. VI, we summarize the results of this paper and discuss future prospects.
II Setup
II.1 Thermodynamics of information
We consider a bipartite classical stochastic system consisting of subsystems and , which take discrete states. The entire system is in contact with a heat bath at inverse temperature , and its dynamics is described by a Markov jump process. System is the target of feedback control, while plays the role of Maxwell’s demon applying feedback operations to based on the measurement results on ’s state.
The set of possible states for is , and for is . Here corresponds to the measurement result on . The joint state of the entire system is represented by the pair , and the probability to find the system in state at time is denoted as . The marginal probabilities for and are defined as and , respectively.
Since we focus on the feedback process, we assume that and are correlated at the initial time , as a consequence of the measurement performed beforehand. Here, we suppose that is a controller that stores and uses the information of the target system . Moreover, we make a simple assumption that the measurement is error-free. That is, the initial distribution is given of the form
(1) |
with being the Kronecker delta, which guarantees that holds with unit probability. During the feedback, the state of is assumed to remain unchanged, i.e., no transitions occur between different states in . That is, transitions from to are prohibited if . This assumption is reasonable for our classical stochastic processes, where the feedback operation on does not influence ’s state itself [37]. Under this assumption, the marginal distribution is fixed to a certain distribution throughout the process, and the initial joint distribution is given by .
The time evolution of the entire system during the feedback process, from to , is described by the master equation:
(2) |
where is the transition rate from to at time . This rate describes the feedback operation that applies to when the measurement result is . The set denotes the pairs of different ’s states that are allowed to transition when the measurement result is . These transitions are assumed to be bidirectional, meaning implies . The stochastic heat absorbed by during a transition from to satisfies the local detailed balance condition:
(3) |
We next introduce mutual information, which represents the amount of information shared between and . The mutual information between and at time is defined as
(4) |
where is the Shannon entropy of the probability distribution . If and are uncorrelated, , while otherwise. In our setup, with the assumptiuon (1), , yielding , which is the maximum mutual information for fixed . This means that has fully acquired the information of at time .
We also introduce entropy production, which represents thermodynamic cost required for the feedback process. The entropy production from time to , denoted as , is defined using the total heat absorbed by up to time , , as
(5) |
Here, the first two terms represent the entropy change of system , and the third term represents the entropy change of the heat bath. Since does not undergo transitions, equals the total heat absorbed by the system . Thus, represents the entropy change of the entire system including the heat bath, and describes dissipation due to irreversibility.
We here introduce the probability of a transition from state to denoted as , and the probability current from state to defined as . We also define thermodynamic force for the transition from to as
(6) |
Then, the entropy production rate can be expressed as
(7) |
which satisfies the second law of thermodynamics .
entropy production for subsystem , , is defined as
(8) |
Using this and the change in the mutual information , the entropy production can be decomposed as [52]. In the present setup, since the maximum mutual information is stored at the initial time, we have . Therefore, the decomposition can be rewritten as
(9) |
By substituting this decomposition into the second law, we obtain
(10) |
This indicates that by consuming mutual information, it is possible to achieve a lower entropy production than that determined by the second law for the case where is absent, . We emphasize that the equality in (10) is achieved in the quasi-static limit, which requires infinite time [37].
II.2 Optimal transport theory and the speed limits
We next briefly overview optimal transport theory. A key insight of optimal transport theory is that the minimized transport cost, known as the Wasserstein distance, serves as a metric between distributions. Optimal transport theory finds applications in fields such as image processing [53], machine learning [54], and biology [55]. Applying it to thermodynamics reveals that the achievable speed limits for entropy productions can be expressed in terms of the Wasserstein distance between initial and final distributions, enabling the identification of optimal protocols that achieve the equality.
The Wasserstein distance between two probability distributions and is defined as
(11) |
where represents the probability transported from state to state . The collection of these probabilities for all states, , is referred to as the transport plan. We denote by the set of all transport plans that transform the initial distribution into the final distribution . This set satisfies the probability conservation conditions: , . The coefficient is the minimum number of transitions required for moving from state to state under the Markov jump process defined in Eq. (2), which determines the cost of transport per unit probability. Thus, the definition in Eq. (11) indicates that by fixing the initial distribution and the final distribution , the minimum total cost over all possible transport plans can be defined as the distance between the distributions. The transport plan that achieves this minimum cost is referred to as the optimal transport plan.
When probability distribution evolves into following Eq. (2) from time to , the entropy production is bounded as [36]
(12) |
where is a function determined by the choice of fixed timescale .
We introduce two quantities representing the timescale: activity and mobility. We define activity as
(13) |
Here, the local activity between states and is defined as , which corresponds to the number of jumps per unit time at time . Additionally, the time integral representing the total number of jumps from time to is expressed as , and its time average is given by . We define mobility as
(14) |
This represents the probability currents induced by the thermodynamic forces applied to the system. As well as activity, the time integral is expressed as , and the time average is given by .
Depending on whether corresponds to either of these two quantities, the function in the lower bound (12) changes. If represents the time-averaged mobility , then ; if represents the time-averaged activity , then .
For either choice of , the optimal protocol that achieves the equality in inequality (12) can be constructed by the condition that the probability is transported from to along the optimal transport plan with a uniform and constant thermodynamic force.
By substituting the decomposition (9) into inequality (12), we obtain the fundamental bound on energy for the fixed initial probability distribution and the final distribution
(15) |
The equality is achieved by the same protocol that achieves inequality (9).
The minimum entropy production for consuming a certain amount of information via feedback processes is not determined by inequality (15) alone. This is because, for a fixed consumed information , there exist infinitely many possible final distributions , which have different Wasserstein distance . Therefore, we can further minimize the right-hand side of (15) which depends on over under fixed , and determine the truly minimum entropy production for consuming .
III Fundamental bounds on the consumed information
III.1 Fano’s inequality
To address the above problem, we first introduce the mathematical tool called Fano’s inequality. It enables us to evaluate the error that arises when a sender sends a message to a receiver through a classical communication channel. Let the set of messages to be transmitted and received be . Due to the presence of errors in the communication channel, when the sender sends a message , the receiver may not always receive with probability 1. The performance of the communication channel is characterized by the conditional probability distribution , which indicates the ambiguity about the original message given the received message .
Consider the case where the sender stochastically transmits messages according to a distribution . Denote the joint probability that sends and receives as . Now we aim to evaluate the error probability ) based on the channel performance . This evaluation is provided by Fano’s inequality, and the traditional form is given by
(16) |
Here, the left-hand side represents the conditional Shannon entropy , which quantifies the randomness of the communication channel’s performance . The right-hand side is an increasing function of the error probability . Thus, Fano’s inequality indicates that the randomness of the channel’s performance leads to a higher error probability. The equality holds when the conditional probabilities satisfies
(17) |
However, depending on the probability distribution , it might not be possible to construct in accordance with Eq. (17). Therefore, the traditional Fano inequality (16) is not generally achievable.
On this matter, a previous research [48] has introduced a tighter version of Fano’s inequality that can be achieved for any . Applying it to the above setup of communication channel, we can summarize the achievable Fano’s inequality as follows. For fixed and , there exists a probability distribution (whose construction will be explained below) determined by and , such that
(18) |
which is achievable for any .
Here, the construction of is as follows:

- (i)
-
(ii)
Define a function that specifies a state of corresponding to (the construction of the function will be provided in step (iii)). Then, is defined as
(19) The corresponding histogram is shown in Fig. 1(c).
-
(iii)
is defined as the largest that satisfies
(20) When , such is uniquely determined in the range . This can be interpreted as the largest that ensures the probability distribution [defined in Eq. (19)] is in descending order.
The achievable Fano’s inequality (18) can be regarded as an inequality that evaluates the upper bound of the conditional Shannon entropy when the error probability and the marginal distribution are fixed. Therefore, this inequality can be applied not only to the communication channel setting described above, but also to general joint systems characterized by any joint probability distribution .
III.2 Main theorem
To obtain the fundamental bounds on entropy production required to consume information by the feedback processes, we consider the minimization problem of the second term on the right-hand side of inequality (15) for a fixed . Since this term is a monotonically increasing function of for both and , it suffices to minimize .
Our main theorem is now stated as follows, which will be used to solve the above problem in the next section.
Theorem.
For a fixed and ,
(21) |
holds regardless of the structure of the entire system . is determined by identifying in Eq. (19). The equality can be achieved by setting the final distribution to a certain distribution (the explicit form will be described in the following section) if the entire system satisfies the following condition (C):
-
(C)
For all states , the direct transition in system from the state to any other state is possible. In other words,the direct transition from state to is possible.
In our setup, we assumed that at , holds with probability 1 through an error-free measurement. Therefore, the condition (C) indicates that regardless of the measurement result , the direct transition from the initial state to any other state is allowed in the feedback control. Inequality (21) provides an upper bound on the mutual information that can be consumed in a finite-time feedback process when the Wasserstein distance is fixed. Since the right-hand side, , is an increasing function of , this implies that the larger the Wasserstein distance between the initial and final distributions, the more mutual information can be consumed by the feedback prosess.
This Theorem comprises two statements: first, that inequality (21) holds for any system which satisfies our setup; second, that if the system meets the condition (C), the equality in (21) can be achieved by setting to an optimal distribution. We give the proof of these two statements in the two following subsections respectively.
III.3 Proof of inequality (21)
We here prove inequality (21). By transforming Eq. (4), the mutual information can be rewritten as
where denotes the conditional Shannon entropy of given at time . In the present setting, is fixed, and initially, with probability 1. Therefore, . Thus, we obtain
(22) |
Let the probability that holds at time be . From achievable Fano’s inequality (18), we obtain
(23) |
Here, is an increasing function of , and holds (the proofs are provided in the Appendix). Substituting these into inequality (23) yields the main Theorem (21).
III.4 Proof of the optimality
We prove that the equality in (21) can be achieved by setting to an optimal distribution if the entire system satisfies the condition (C). First, we will describe the method for constructing the optimal final distribution . Next, we will prove that this final distribution satisfies the constraints and . Finally, we will show that this final distribution achieves the equality in (21).
The optimal final distribution can be constructed as follows. Since the joint probability distribution of the entire system can be expressed as , it suffices to define and . Here, let denote the permutation that rearranges the states in descending order of the probability distribution . In other words, means that corresponds to the -th largest value of . If there are states with equal probabilities, their ordering in can be chosen arbitrarily; once is defined, it must be fixed throughout the following discussion.
The optimal is then defined as
(24) |
From the definition of , we have and . The optimal conditional probability distribution is given as
(25) |

This procedure for constructing the optimal probability distribution can also be visualized graphically as follows: begin with (Fig. 2 (a)), which by definition is a probability distribution arranged in descending order along the states . Apply the permutation to (Fig. 2 (b)), which rearranges the states such that the descending order of probabilities matches that of . Then, swap the probability of the state with the probability of the state with the largest probability (i.e., the state satisfying ) (Fig. 2 (c)). From the second operation, the optimal conditional probability distribution satisfies .
We next show that its marginal distribution matches the fixed . For such that holds, from Eq. (25), holds independently of , which yields . For such that holds, from Eq. (25) and the definition of , takes the value for and for . Therefore,
(26) | ||||
(27) | ||||
(28) |
Here, we used Eq. (24) for the transformation from Eq. (27) to Eq. (28).
We next show that under the condition (C), the Wasserstein distance between the constructed and equals the fixed value . When the condition (C) is satisfied, holds (the proof is provided in the Appendix). Therefore, we obtain
(29) | ||||
(30) | ||||
(31) | ||||
(32) | ||||
(33) |
We finally show that the constructed is optimal, i.e., it achieves the equality in (21). From Eq. (25), since holds regardless of ,
(34) | ||||
(35) | ||||
(36) |
holds. Therefore, the equality in (21) can be achieved.
IV Fundamental bound on entropy production
IV.1 Speed limit for fixed mobility
In the previous section, we considered the maximization problem for the consumed information for fixed , whose solution is given by the main Theorem (21). Since the original problem of minimizing for fixed corresponds to the dual problem, its solution is directly derived by the main Theorem. From this dual solution, we can obtain the fundamental bound on entropy production required to consume .
First, the right-hand side of inequality (21), , is a strictly increasing function of in the range , and takes a constant value for (the proof is provided in the Appendix). Therefore, given a fixed , the inverse function can be defined as . This function provides the minimum Wasserstein distance for the fixed amount of consumed information , .
Here, considering Eq. (15) with chosen as the time-averaged mobility , the right-hand side becomes an increasing function of . Therefore, minimizing is equivalent to minimizing entropy production . This allows us to derive the speed limit in the feedback process as
(37) |
which provides the minimum entropy production required to consume information through feedback processes keeping time-averaged mobility fixed.
The equality in inequality (37) is achieved by simultaneously satisfying the equalities in inequalities (12) and (21). First, the equality in (21) is achieved by setting the final distribution to the form determined by Eqs. (24) and (25), while satisfying condition (C). Next, the equality in (12) is achieved by constructing a protocol that transports the probability between and along the optimal transport plan, with uniform and constant thermodynamic force. Consequently, inequality (37) is satisfied. This highlights the thermodynamic implication of our main Theorem.
IV.2 Speed limit for fixed activity
When we choose time-averaged activity as , the second term on the right-hand side of inequality (12) becomes . Similarly to the case where mobility is fixed, this term is an increasing function of for any fixed and . Therefore, by applying the main Theorem, we can derive another speed limit
(38) |
which provides the minimum entropy production required to consume the information in feedback control when time-averaged activity is fixed.
The equality in inequality (38) is achieved by simultaneously achieving the equalities in inequalities (12) and (21). First, the equality in (21) is realized by setting the final distribution to the form determined by Eqs. (24) and (25), while satisfying condition (C). Next, the equality in (12) is achieved by constructing a protocol that transports the probability between the specified initial and final distributions and along the optimal transport plan, with uniform and constant thermodynamic force. Consequently, inequality (38) is achieved. This again highlights the thermodynamic implication of the main Theorem.
V Example: Coupled two level system
We demonstrate our main inequalities (37) and (38) by numerical simulation of a coupled system which can take two states, and . Such a setup may be realized, for instance, by confining colloidal particles to two positions corresponding to and using an optical potential, with coupling mediated by Coulomb interactions. In this case, if the time evolution of the entire system is governed by Eq. (2), the controllable transition rates are given by the set . Throughout this section, we represent the probability distribution as a matrix .
First, when fixing , the initial distribution is given as
(41) |
The equality in the bound (37) is achieved when the equalities in both (12) and (21) are simultaneously satisfied. By setting with respect to the consumed mutual information , the optimal final distribution that satisfies the equality in (21) is obtained as
(44) |
For this final distribution, the protocol that achieves the equality in (12) is realized by transporting probabilities along the optimal transport plan from to under a constant and uniform thermodynamic force . Given the protocol (the detailed form of the protocol is provided in the Appendix), we can calculate entropy production for it. A numerical calculation of for each is shown as the orange line in Fig. 3 (a).

Comparing this result with the right-hand side of (37) computed for each (shown as the green dotted line in Fig. 3 (a)), it is evident that this protocol achieves the equality in the bound (38). Notably, the fact that entropy production is negative indicates that the feedback using information enables achieving an entropy production smaller than that imposed by the conventional thermodynamic second law, . In other words, it implies the ability to extract work exceeding the negative of the change in nonequilibrium free energy.
Additionally, we also show examples of non-optimal protocols. Specifically, we consider the case where the equality in (12) is achieved but the equality in (21) is not. For instance, for the same initial distribution, constructing a final distribution of the form
(47) |
by choosing such that the mutual information change remains fixed at yields a non-optimal final distribution for consuming . For this distribution, numerical calculations of entropy production under a protocol that achieves the equality in (12) (the detailed form of the protocol is provided in the Appendix) are shown as the gray line in Fig. 3(a). Since it does not achieve the equality in (37), it is evident that this is not an optimal feedback protocol for consuming .
We next numerically demonstrate the bound (38) for fixed activity. For the initial distribution given by (41), we can take an optimal final distribution given by (44) and a non-optimal final distribution given by (47). For these final distributions, the protocols that achieve the equality in (12) transport probabilities from to along the optimal transport plan under a uniform and constant thermodynamic force. When represents the time-averaged activity fixed as , maintaining a constant thermodynamic force corresponds to setting the probability current to a constant value . Numerical calculations of entropy production for such a protocol, (the detailed form of the protocol is provided in the Appendix), are shown in Fig. 3 (b). The orange line corresponds to the optimal final distribution, while the gray line corresponds to the non-optimal final distribution. Comparing these results with the right-hand side of (38) computed for each (shown as the green dotted line in Fig. 3(a)), it can be seen that the bound (38) is achieved for the optimal final distribution, while it is not achieved for the non-optimal final distribution.
VI Conclusion
In this study, we have derived the main Theorem (21) which provides the upper bound on the change in mutual information under the fixed error-free initial distribution, the fixed marginal distribution of the controller, and the fixed Wasserstein distance. The equality can be achieved under the condition (C) by setting the optimal final distribution. This Theorem gives the maximum change in mutual information that can be induced under a fixed minimum dissipation when the controller performs feedback control based on error-free measurement outcomes. Based on this Theorem, we have derived the speed limits (37) and (38) which give the lower bounds on entropy production for consuming a fixed amount of information through the feedback process. These bounds are achievable under the condition (C), and we have identified the optimal protocols. While a similar optimization problem for measurement processes has been solved in our previous work [45], the present work for feedback processes is based on largely different mathematical techniques including the achievable Fano’s inequality (16) [48], and therefore separate treatments are required for measurement and feedback processes.
Our results theoretically advance thermodynamics of information by specifying the form of contribution of processed information to finite-time entropy production. The optimal feedback protocols identified in this study can be applied to a wide range of feedback processes described by Markov jump processes, which would be experimentally implemented by single electron devices [49, 50, 51], double quantum dots in the classical regime [56, 57], and effective discrete dynamics by Brownian nanoparticles [58]. Future directions also include extending our theoretical framework to continuous or quantum systems.
ACKNOWLEDGEMENTS
We thank Kosuke Kumasaki and Yosuke Mitsuhashi for valuable discussions. R.N. is supported by the World-leading Innovative Graduate Study Program for Materials Research, Industry, and Technology (MERIT-WINGS) of the University of Tokyo. T.S. is supported by Japan Society for the Promotion of Science (JSPS) KAKENHI Grant No. JP19H05796, JST, CREST Grant No. JPMJCR20C1 and JST ERATO-FS Grant No. JPMJER2204. T.S. is also supported by the Institute of AI and Beyond of the University of Tokyo and JST ERATO Grant No. JPMJER2302, Japan.
Appendix A Derivation of the main Theorem and the speed limits
A.1 Majorization
We introduce the concept of majorization, which is used to provide an upper bound on mutual information. For two probability distributions and , we denote that majorizes , or , when they satisfy [60, 59]
(48) |
Here, and are the distributions obtained by rearranging and in descending order, respectively. When , the following relationship holds for any convex function [60, 59]:
(49) |
By setting and taking the negative of the both sides, we obtain the following relation of the Shannon enropy:
(50) |
A.2 Proof of the main Theorem
The proof of the main Theorem is structured into two propositions.
Proposition A.1.
Proof.
Since the Wasserstein distance is greater than or equal to the total variation distance, we obtain
(51) |
Since the marginal distribution of is fixed as , the absolute value on the right-hand side can be calculated as
(52) |
By using , we get
(53) |
∎
Proposition A.2.
is monotonically and strictly increasing in .
Proof.
Recalling the definition of , if there exists a maximum that satisfies , this is defined as . First, we prove that decreases with respect to . Let . For , we have
(54) | ||||
(55) | ||||
(56) | ||||
(57) | ||||
(58) |
Therefore, increases with respect to . Consequently, , which is determined as the maximum satisfying , decreases with respect to .
We next recall the definition of :
(59) |
We note that the permutation , which rearranges in descending order, coincides with , the permutation rearranging in descending order. Now, consider . We aim to show that , i.e.,
(60) |
First, when , also holds. Therefore, we have
(61) |
When ,
(62) | ||||
(63) |
holds, which yields
(64) | |||
(65) | |||
(66) | |||
(67) |
Additionally, when ,
(68) | ||||
(69) |
holds. Therefore,
(71) | |||
(72) | |||
(73) |
From the above, it follows that . Moreover, by definition, when , holds. ∎
A.3 Formalism by Lorenz curves
While was constructed using histograms of probability distributions in Sec.III, this construction can also be explained using Lorenz curves. A Lorenz curve is a polyline uniquely associated with a probability distribution and is constructed as follows [59]. First, arrange the probability distribution in descending order and obtain . Next, plot the cumulative probability as a function of state . The polyline obtained by connecting these points in addition to the origin is the Lorenz curve for the probability distribution .
The Lorenz curve visualizes the concentration of probabilities in the distribution. This can be understood as follows: consider two probability distributions and , and assume that the Lorenz curve of lies above that of . That is, for all , . Then, from the definition of majorization introduced in Appendix A1, holds, which implies . Therefore, a probability distribution with a Lorenz curve higher on the graph has smaller Shannon entropy and is more concentrated.
Using this property of Lorenz curves, the construction of can be intuitively explained. First, if , . For , first draw the Lorenz curve of (the black line in Fig. 4 (a)). Next, extend a half-line upward and to the right from point defined by coordinate (the red point in Fig. 4 (a)), and rotate it around to approach the Lorenz curve of from above (the red dotted line in Fig. 4 (a)). When the half-line touches the Lorenz curve of , stop the rotation, and call the point of contact (Fig. 4 (b)). The state corresponding to point is as defined in Sec. III. Finally, connect the three points (origin, , and ) with polyline, and extend the curve beyond along the Lorenz curve of . The probability distribution corresponding to this modified Lorenz curve (the red line in Fig. 4 (c)) is defined as .

Appendix B Optimal protocol for a two level system
In the example of the two-level system discussed in Sec. V, the optimal transport from the initial distribution defined by Eq. (41) to the optimal final distribution defined by Eq. (44), as well as to the non-optimal final distribution defined by Eq. (47), was constructed. To provide a general method for constructing these protocols, we present the protocol for optimal transport from the initial distribution defined by Eq. (41) to the final distribution defined by
(76) |
B.1 For fixed mobility
The protocol that achieves the fundamental limit (37) for fixed mobility transports probabilities along the optimal transport plan from the initial distribution to the final distribution under a constant and uniform thermodynamic force . The optimal transport plan is to transport a probability from state to and a probability from state to .
To implement this using a constant thermodynamic force by controlling the transition rates in a Markov jump process, the time interval is divided into two intervals: and . In individual intervals, the transports of and are performed.
First, in the interval , only the transport of probability from state to is carried out. An example of the time evolution implemented in this interval is
(77) | ||||
(78) |
while keeping the probabilities and . To implement this under a constant thermodynamic force , the transition rates should be set as
(79) | ||||
(80) |
The remaining transition rates are set to .
In the interval , the probability transport of from state to is performed. The example of the time evolution to be implemented in this case is
(81) | ||||
(82) |
with and kept constant. To implement this under a constant and uniform thermodynamic force , the transition rates should be set as
(83) | ||||
(84) |
The remaining transition rates are set to . Moreover, the entropy production generated by the entire protocol over the two intervals is given by
(85) | ||||
(86) | ||||
(87) |
which achieves the bound (37). The orange solid line and gray solid line representing entropy production in Fig. 3 (a) are plotted using calculated from Eq. (85), with .
B.2 For fixed activity
Next, we present the protocol that achieves the fundamental bound (38) for fixed activity. This protocol transports probabilities along the optimal transport plan from the initial distribution to the final distribution under a constant and uniform activity and probability current . The optimal transport plan, as well as the case of fixed mobility, transports a probability from state to and a probability from state to .
To achieve this under a constant activity and probability current (which is uniquely determined by the fixed time and the initial and final distributions), the time interval is divided into two intervals and . In individual intervals, the transports of and are performed.
First, in the interval , the transport of a probability from state to is performed. The time evolution to be implemented is
(88) |
with and kept constant. To implement this under a constant activity and probability current , the transition rates should be set as
(89) | ||||
(90) |
The remaining transition rates are set to .
In the interval , the transport of a probability from state to is conducted. The time evolution to be implemented in this interval is
(91) | ||||
(92) |
with and kept constant. To implement this under a constant activity and probability current , the transition rates should be set as
(93) | ||||
(94) |
The remaining transition rates are set to . The total entropy production generated by the entire protocol is calculated as
(95) | ||||
(96) | ||||
(97) |
which achieves the bound (38). The orange solid line and gray solid line representing entropy production in Fig. 3 (b) are plotted using calculated from Eq. (95), with .
References
- Peliti and Pigolotti [2021] L. Peliti and S. Pigolotti, Stochastic Thermodynamics: An Introduction (Princeton University Press, 2021).
- Seifert [2012] U. Seifert, Reports on progress in physics 75, 126001 (2012).
- Landi and Paternostro [2021] G. T. Landi and M. Paternostro, Rev. Mod. Phys. 93, 035008 (2021).
- Aurell et al. [2012a] E. Aurell, K. Gawȩdzki, C. Mejía-Monasterio, R. Mohayaee, and P. Muratore-Ginanneschi, J. Stst. Phys. 147, 487 (2012a).
- Shiraishi et al. [2018] N. Shiraishi, K. Funo, and K. Saito, Phys. Rev. Lett. 121, 070601 (2018).
- Ito [2018] S. Ito, Phys. Rev. Lett. 121, 030605 (2018).
- Ito and Dechant [2020] S. Ito and A. Dechant, Phys. Rev. X 10, 021056 (2020).
- Vo et al. [2020] V. T. Vo, T. Van Vu, and Y. Hasegawa, Phys. Rev. E 102, 062132 (2020).
- Plata et al. [2020] C. A. Plata, D. Guéry-Odelin, E. Trizac, and A. Prados, Phys. Rev. E 101, 032129 (2020).
- Falasco and Esposito [2020] G. Falasco and M. Esposito, Phys. Rev. Lett. 125, 120604 (2020).
- Yoshimura and Ito [2021] K. Yoshimura and S. Ito, Phys. Rev. Lett. 127, 160601 (2021).
- Barato and Seifert [2015] A. C. Barato and U. Seifert, Phys. Rev. Lett. 114, 158101 (2015).
- Gingrich et al. [2016] T. R. Gingrich, J. M. Horowitz, N. Perunov, and J. L. England, Phys. Rev. Lett. 116, 120601 (2016).
- Pietzonka et al. [2016] P. Pietzonka, A. C. Barato, and U. Seifert, Journal of Statistical Mechanics: Theory and Experiment 2016, 124004 (2016).
- Shiraishi et al. [2016] N. Shiraishi, K. Saito, and H. Tasaki, Phys. Rev. Lett. 117, 190601 (2016).
- Horowitz and Gingrich [2017] J. M. Horowitz and T. R. Gingrich, Phys. Rev. E 96, 020103 (2017).
- Dechant [2018] A. Dechant, Journal of Physics A: Mathematical and Theoretical 52, 035001 (2018).
- Brandner et al. [2018] K. Brandner, T. Hanazato, and K. Saito, Phys. Rev. Lett. 120, 090601 (2018).
- Hasegawa and Van Vu [2019] Y. Hasegawa and T. Van Vu, Phys. Rev. E 99, 062126 (2019).
- Koyuk and Seifert [2020] T. Koyuk and U. Seifert, Phys. Rev. Lett. 125, 260604 (2020).
- Horowitz and Gingrich [2020] J. M. Horowitz and T. R. Gingrich, Nature Physics 16, 15 (2020).
- Liu et al. [2020] K. Liu, Z. Gong, and M. Ueda, Phys. Rev. Lett. 125, 140602 (2020).
- Jordan et al. [1998] R. Jordan, D. Kinderlehrer, and F. Otto, SIAM Journal on Mathematical Analysis 29, 1 (1998).
- Benamou and Brenier [2000] J.-D. Benamou and Y. Brenier, Numerische Mathematik 84, 375 (2000).
- Maas [2011] J. Maas, Journal of Functional Analysis 261, 2250 (2011).
- Aurell et al. [2011] E. Aurell, C. Mejía-Monasterio, and P. Muratore-Ginanneschi, Phys. Rev. Lett. 106, 250601 (2011).
- Aurell et al. [2012b] E. Aurell, K. Gawȩdzki, C. Mejía-Monasterio, R. Mohayaee, and P. Muratore-Ginanneschi, J. Stat. Phys. 147, 487 (2012b).
- Dechant and Sakurai [2019] A. Dechant and Y. Sakurai, Thermodynamic interpretation of wasserstein distance (2019), arXiv:1912.08405 [cond-mat.stat-mech] .
- Chen et al. [2017] Y. Chen, T. T. Georgiou, L. Ning, and A. Tannenbaum, IEEE control systems Lett. 1, 14 (2017).
- Chen et al. [2019] Y. Chen, T. T. Georgiou, and A. Tannenbaum, IEEE transactions on automatic control 65, 2979 (2019).
- Fu et al. [2021] R. Fu, A. Taghvaei, Y. Chen, and T. T. Georgiou, Automatica 123, 109366 (2021).
- Miangolarra et al. [2022] O. M. Miangolarra, A. Taghvaei, Y. Chen, and T. T. Georgiou, IEEE Control Systems Lett. 6, 3409 (2022).
- Van Vu and Hasegawa [2021] T. Van Vu and Y. Hasegawa, Phys. Rev. Lett. 126, 010601 (2021).
- Dechant [2022] A. Dechant, Journal of Physics A: Mathematical and Theoretical 55, 094001 (2022).
- Chennakesavalu and Rotskoff [2023] S. Chennakesavalu and G. M. Rotskoff, Phys. Rev. Lett. 130, 107101 (2023).
- Van Vu and Saito [2023] T. Van Vu and K. Saito, Phys. Rev. X 13, 011013 (2023).
- Parrondo et al. [2015] J. M. R. Parrondo, J. M. Horowitz, and T. Sagawa, Nat. Phys. 11, 131 (2015).
- Zulkowski and DeWeese [2014] P. R. Zulkowski and M. R. DeWeese, Phys. Rev. E 89, 052140 (2014).
- Proesmans et al. [2020a] K. Proesmans, J. Ehrich, and J. Bechhoefer, Phys. Rev. E 102, 032105 (2020a).
- Proesmans et al. [2020b] K. Proesmans, J. Ehrich, and J. Bechhoefer, Phys. Rev. Lett. 125, 100602 (2020b).
- Nakazato and Ito [2021] M. Nakazato and S. Ito, Phys. Rev. Res. 3, 043093 (2021).
- Tanogami et al. [2023] T. Tanogami, T. Van Vu, and K. Saito, Phys. Rev. Res. 5, 043280 (2023).
- Xia et al. [2024] S. Xia, S. Han, O. Pan, Y. Pan, J. Chen, and S. Su, arXiv preprint arXiv:2407.02848 (2024).
- Fujimoto and Ito [2024] Y. Fujimoto and S. Ito, Phys. Rev. Res. 6, 013023 (2024).
- Nagase and Sagawa [2024] R. Nagase and T. Sagawa, Phys. Rev. Res. 6, 033239 (2024).
- Kamijima et al. [2024a] T. Kamijima, K. Funo, and T. Sagawa, arXiv preprint arXiv:2409.08606 (2024a).
- Kamijima et al. [2024b] T. Kamijima, A. Takatsu, K. Funo, and T. Sagawa, arXiv preprint arXiv:2410.11603 (2024b).
- Sakai [2020] Y. Sakai, Entropy 22 (2020).
- Koski et al. [2014] J. V. Koski, V. F. Maisi, T. Sagawa, and J. P. Pekola, Phys. Rev. Lett. 113, 030601 (2014).
- Koski et al. [2015] J. V. Koski, A. Kutvonen, I. M. Khaymovich, T. Ala-Nissila, and J. P. Pekola, Phys. Rev. Lett. 115, 260602 (2015).
- Chida et al. [2017] K. Chida, S. Desai, K. Nishiguchi, and A. Fujiwara, Nature communications 8, 15301 (2017).
- Sagawa and Ueda [2009] T. Sagawa and M. Ueda, Phys. Rev. Lett. 102, 250602 (2009).
- Haker et al. [2004] S. Haker, L. Zhu, A. Tannenbaum, and S. Angenent, Int. J. Comput. Vis. 60, 225 (2004).
- Kolouri et al. [2017] S. Kolouri, S. R. Park, M. Thorpe, D. Slepcev, and G. K. Rohde, IEEE Signal Processing Magazine 34, 43 (2017).
- Schiebinger et al. [2019] G. Schiebinger, J. Shu, M. Tabaka, B. Cleary, V. Subramanian, A. Solomon, J. Gould, S. Liu, S. Lin, P. Berube, L. Lee, J. Chen, J. Brumbaugh, P. Rigollet, K. Hochedlinger, R. Jaenisch, A. Regev, and E. S. Lander, Cell 176, 928 (2019).
- Schaller et al. [2010] G. Schaller, G. Kießlich, and T. Brandes, Phys. Rev. B 82, 041303 (2010).
- Annby-Andersson et al. [2020] B. Annby-Andersson, P. Samuelsson, V. F. Maisi, and P. P. Potts, Physical Review B 101, 165404 (2020).
- Gingrich et al. [2017] T. R. Gingrich, G. M. Rotskoff, and J. M. Horowitz, Journal of Physics A: Mathematical and Theoretical 50, 184004 (2017).
- Sagawa [2022] T. Sagawa, Classical majorization, in Entropy, Divergence, and Majorization in Classical and Quantum Thermodynamics (Springer Singapore, Singapore, 2022) pp. 21–37.
- Bhatia [2013] R. Bhatia, Matrix analysis, Vol. 169 (Springer Science & Business Media, 2013).