Optimized quantum -divergences
Abstract
The quantum relative entropy is a measure of the distinguishability of two quantum states, and it is a unifying concept in quantum information theory: many information measures such as entropy, conditional entropy, mutual information, and entanglement measures can be realized from it. As such, there has been broad interest in generalizing the notion to further understand its most basic properties, one of which is the data processing inequality. The quantum -divergence of Petz is one generalization of the quantum relative entropy, and it also leads to other relative entropies, such as the Petz–Rényi relative entropies. In this contribution, I introduce the optimized quantum -divergence as a related generalization of quantum relative entropy. I prove that it satisfies the data processing inequality, and the method of proof relies upon the operator Jensen inequality, similar to Petz’s original approach. Interestingly, the sandwiched Rényi relative entropies are particular examples of the optimized -divergence. Thus, one benefit of this approach is that there is now a single, unified approach for establishing the data processing inequality for both the Petz–Rényi and sandwiched Rényi relative entropies, for the full range of parameters for which it is known to hold.
Full version of this paper is accessible at arXiv:1710.10252
I Introduction
The quantum relative entropy [1] is a foundational distinguishability measure in quantum information theory (QIT). It is a function of two quantum states and measures how well one can tell the two states apart by a quantum-mechanical experiment. One important reason for why it has found such widespread application is that it satisfies a data-processing inequality [2, 3]: it does not increase under the action of a quantum channel on the two states. This can be interpreted as saying that two quantum states do not become more distinguishable if the same quantum channel is applied to them, and a precise interpretation of this statement in terms of quantum hypothesis testing is available in [4, 5, 6]. Quantum relative entropy generalizes its classical counterpart [7].
The wide interest in relative entropy sparked various researchers to generalize and study it further, in an attempt to elucidate the fundamental properties that govern its behavior. One notable generalization is Rényi’s relative entropy [8], but this was subsequently generalized even further in the form of the -divergence [9, 10, 11]. For probability distributions and and a convex function , the -divergence is defined as in the case that for all such that . The resulting quantity is then non-increasing under the action of a classical channel that produces the output distributions and . Some years after these developments, a quantum generalization of -divergence appeared in [12, 13] In [12, 13] and a later development [14], the quantum data-processing inequality was proved in full generality for arbitrary quantum channels, whenever the underlying function is operator convex.
Interestingly, when generalizing a notion from classical to QIT, there is often more than one way to do so, and sometimes there could even be an infinite number of ways to do so. This has to do with the non-commutativity of quantum states. For example, there are several different ways that one could generalize the relative entropy to the quantum case, and two prominent formulas were put forward in [1] and [15]. This added complexity for the quantum case could potentially be problematic, but the typical way of determining on which generalizations we should focus is to show that a given formula is the answer to a meaningful operational task. The papers [4, 5] accomplished this for the formula from [1], and since then, researchers have realized more and more just how foundational the formula of [1] is. As a consequence, the formula of [1] is now known as quantum relative entropy.
The situation becomes more intricate when it comes to quantum generalizations of Rényi relative entropy. For many years, the Petz–Rényi relative entropy of [12, 13] has been widely studied and given an operational interpretation [16, 17], again in the context of quantum hypothesis testing. However, in recent years, the sandwiched Rényi relative entropy of [18, 19] has gained prominence, due to its role in establishing strong converses for communication tasks (see, e.g.,[19, 20]). The result of [21] solidified its fundamental meaning in QIT, proving that it has an operational interpretation in the strong converse exponent of quantum hypothesis testing. As such, the situation is that there are two generalizations of Rényi relative entropy that should be considered in QIT, due to their operational role mentioned above.
The same work that introduced the Petz–Rényi relative entropy also introduced a quantum generalization of the notion of -divergence [12, 13] (see also [22]), with the Petz–Rényi relative entropy being a particular example. Since then, other quantum -divergences have appeared [23, 24], now known as minimal and maximal -divergences [25, 24]. However, it has not been known how the sandwiched Rényi relative entropy fits into the paradigm of quantum -divergences.
In this paper, I modify Petz’s definition of quantum -divergence [12, 13, 22], by allowing for a particular optimization (see Definition 1 for details of the modification). As such, I call the resulting quantity the optimized quantum -divergence. I prove that it obeys a quantum data processing inequality, and as such, my perspective is that it deserves to be considered as another variant of the quantum -divergence, in addition to the original, the minimal, and the maximal. Interestingly, the sandwiched Rényi relative entropy is directly related to the optimized quantum -divergence, thus bringing the sandwiched quantity into the -divergence formalism.
One benefit of the results of this paper is that there is now a single, unified approach for establishing the data-processing inequality for both the Petz–Rényi relative entropy and the sandwiched Rényi relative entropy, for the full Rényi parameter ranges for which it is known to hold. This unified approach is based on Petz’s original approach that employed the operator Jensen inequality [26], and it is useful for presenting a succint proof of the data processing inequality for both quantum Rényi relative entropy families.
In the rest of the paper, I begin by defining the optimized quantum -divergence in the next section. In Section III, I prove that the optimized -divergence satisfies the quantum data processing inequality under partial trace whenever the underlying function is operator anti-monotone with domain and range . The core tool underlying this proof is the operator Jensen inequality [26]. In Section IV, I show how the quantum relative entropy and the sandwiched Rényi relative entropies are directly related to the optimized quantum -divergence. Section V then discusses the relation between Petz’s -divergence and the optimized one. I finally conclude in Section VI with a summary.
II Optimized quantum -divergence
Let us begin by defining the optimized quantum -divergence. Here I focus exclusively on the case of positive definite operators, and the full version provides details for the more general case of positive semi-definite operators.
Definition 1 (Optimized quantum -divergence)
Let be a function with domain and range . For positive definite operators and acting on a Hilbert space , we define the optimized quantum -divergence as
(1) |
where is defined for positive definite and acting on as
(2) | ||||
(3) |
In the above, is an auxiliary Hilbert space isomorphic to , for orthonormal bases and , and the superscript indicates transpose with respect to the basis .
The case of greatest interest for us here is when the underlying function is operator anti-monotone; i.e., for Hermitian operators and , the function is such that (see, e.g., [27]). This property is rather strong, but there are several functions of interest in quantum physical applications that obey it (see Section IV). One critical property of an operator anti-monotone function with domain and range is that it is also operator convex and continuous (see, e.g., [28]). In this case, we have the following simple proposition, proved in the full version:
Proposition 2
Let be an operator anti-monotone function with domain and range . For positive definite operators and acting on a Hilbert space ,
(4) |
and the function is concave in .
III Quantum data processing
Our first main objective is to prove that deserves the name “-divergence” or “-relative entropy,” i.e., that it is monotone non-increasing under the action of a completely positive trace-preserving map :
(5) |
Such a map is also called a quantum channel, due to its purpose in quantum physics as modeling the physical evolution of the state of a quantum system. In QIT contexts, the inequality in (5) is known as the quantum data processing inequality. According to the Stinespring dilation theorem [29], to every quantum channel , there exists an isometry such that
(6) |
As such, we can prove the inequality in (5) in two steps. Isometric invariance: First show that
(7) |
for any isometry and any positive semi-definite and . This is done in the full version of this work, using the general definition given there. Monotonicity under partial trace: Then show that
(8) |
for positive semi-definite operators and acting on the tensor-product Hilbert space , with and .
We now discuss the second step toward quantum data processing, mentioned above, and here we focus exclusively on positive definite operators:
Theorem 3 (Monotonicity under partial trace)
Let be an operator anti-monotone function with domain and range . Given positive definite operators and acting on the tensor-product Hilbert space , the optimized quantum -divergence does not increase under the action of a partial trace, in the sense that
(9) |
where and .
Proof:
The quantities of interest are as follows:
(10) |
(11) |
where and are invertible density operators and, by definition,
(12) |
The following map, acting on an operator , is a quantum channel known as the Petz recovery channel [30, 31]:
(13) |
It is completely positive because it consists of the serial concatenation of three completely positive maps: sandwiching by , tensoring in the identity , and sandwiching by . It is also trace preserving. The Petz recovery channel has the property that it perfectly recovers if is input because
(14) |
Every completely positive and trace preserving map has a Kraus decomposition, which is a set of operators such that and A standard construction for an isometric extension of a channel is then to pick an orthonormal basis for an auxiliary Hilbert space and define
(15) |
One can then readily check that and . For the Petz recovery channel, we can figure out a Kraus decomposition by expanding the identity operator , with respect to some orthonormal basis , so that
Thus, Kraus operators for the Petz recovery channel are . According to the standard recipe in (15), we can construct an isometric extension of the Petz recovery channel as
(16) |
We can then extend this isometry to act as an isometry on a larger space by tensoring it with the identity operator , and so we define
(17) |
We can also see that acting on generates : . This can be interpreted as a generalization of (14) in the language of QIT: an isometric extension of the Petz recovery channel perfectly recovers a purification of from a purification of . Since the Petz recovery channel is indeed a channel, we can pick as the output state of the Petz recovery channel acting on an invertible state :
(18) |
Observe that is invertible. Then consider that
(19) | |||
(20) | |||
(21) | |||
(22) |
For the second equality, we used the fact that for the choice of in (18). With this setup, we can now readily establish the desired inequality by employing the operator Jensen inequality [26] and operator convexity of the function :
(23) | |||
(24) | |||
(25) | |||
(26) | |||
(27) |
Taking a supremum over such that and , we conclude that the following inequality holds for all invertible states :
(28) |
After taking a supremum over invertible states , we find that the inequality in (9) holds when is invertible. ∎
IV Examples of optimized quantum -divergences
I now show how several known quantum divergences are particular examples of an optimized quantum -divergence, including the quantum relative entropy [1] and the sandwiched Rényi relative quasi-entropies [18, 19]. The result will be that Theorem 3 recovers quantum data processing for the sandwiched Rényi relative entropies for the full range of parameters for which it is known to hold. Thus, one benefit of Theorem 3 and earlier work of [12, 13, 14] is a single, unified approach, based on the operator Jensen inequality [26], for establishing quantum data processing for all of the Petz– and sandwiched Rényi relative entropies for the full parameter ranges for which data processing is known to hold.
IV-A Quantum relative entropy as optimized quantum -divergence
Let be an invertible state and and positive definite. Let . Pick the function , which is an operator anti-monotone function with domain and range , and we find that
(29) | |||
(30) | |||
(31) | |||
(32) |
The inequality is a consequence of Klein’s inequality [32] (see also [33]), establishing that the optimal is set to . So we find that where the quantum relative entropy is defined as [1] .
IV-B Sandwiched Rényi relative quasi-entropy as optimized quantum -divergence
Take , , and as defined in Section IV-A. For , pick the function which is an operator anti-monotone function with domain and range . Note that this is a reparametrization of for . I now show that
(33) |
which is the known expression for sandwiched quasi-entropy for [18, 19]. To see this, consider that
(34) |
Now optimizing over invertible states and employing Hölder duality, in the form of the reverse Hölder inequality and as observed in [18], we find that
(35) |
where for positive semi-definite , we define . We then get that
(36) | ||||
(37) |
which is the sandwiched Rényi relative quasi-entropy for the range . The sandwiched Rényi relative entropy itself is defined up to a normalization factor as [18, 19]
(38) |
Thus, Theorem 3 implies quantum data processing for the sandwiched Rényi relative entropy for the parameter range , which is a result previously established in [34].
For , pick the function which is an operator anti-monotone function with domain and range . Note that this is a reparametrization of for . I now show that
(39) |
which is the known expression for sandwiched Rényi relative quasi-entropy for [18, 19]. To see this, consider that the same development as above gives that
(40) |
Again employing Hölder duality, as observed in [18], we find
(41) |
We then get that
(42) | ||||
(43) |
where the equalities hold as observed in [18]. The sandwiched Rényi relative entropy itself is defined up to a normalization factor as in (38) [18, 19]. Thus, Theorem 3 implies quantum data processing for the sandwiched Rényi relative entropy for the parameter range , which is a result previously established in full by [34, 35, 21] and for by [18, 19].
V On Petz’s quantum -divergence
I now discuss in more detail the relation between the optimized quantum -divergence and Petz’s -divergence from [12, 13]. In brief, we find that the Petz -divergence can be recovered by replacing in Definition 1 with .
Definition 4 (Petz quantum -divergence)
Let be a continuous function with domain and range . For positive definite operators and acting on a Hilbert space , the Petz quantum -divergence is defined as
(44) |
where the notation is the same as in Definition 1.
One main concern is about quantum data processing with the Petz -divergence. To show this, we take to be an operator anti-monotone function with domain and range . As discussed in Section III, one can establish data processing by showing isometric invariance and monotonicity under partial trace. Isometric invariance of follows from the same proof as given in the full version and was also shown in [14]. Monotonicity of under partial trace for positive definite and follows from the operator Jensen inequality [12, 13].
Special and interesting cases of the Petz -divergence are found by taking for , and for . Each of these functions are operator anti-monotone with domain and range . As shown in [12, 13], all of the following quantities obey the data processing inequality:
(45) | ||||
(46) | ||||
(47) |
By a reparametrization , the latter two quantities are directly related to the Petz Rényi relative entropy . Thus, the data processing inequality holds for for [13, 14].
VI Conclusion
The main contribution of the present work is the definition of the optimized quantum -divergence and the proof that the data processing inequality holds for it whenever the function is operator anti-monotone with domain and range . The proof of the data processing inequality relies on the operator Jensen inequality [26], and it bears some similarities to the original approach from [12, 13, 14]. Furthermore, I showed how the sandwiched Rényi relative entropies are particular examples of the optimized quantum -divergence. As such, one benefit of this paper is that there is now a single, unified approach, based on the operator Jensen inequality [26], for establishing the data processing inequality for the Petz–Rényi and sandwiched Rényi relative entropies, for the full range of parameters for which it is known to hold.
Acknowledgements. I thank Anna Vershynina for discussions related to the topic of this paper, and I acknowledge support from the NSF under grant no. 1714215.
References
- [1] H. Umegaki, “Conditional expectations in an operator algebra IV,” Kodai Math. Sem. Rep., vol. 14, no. 2, pp. 59–85, 1962.
- [2] G. Lindblad, “Completely positive maps and entropy inequalities,” Comm. Math. Phys., vol. 40, no. 2, pp. 147–151, June 1975.
- [3] A. Uhlmann, “Relative entropy and the Wigner-Yanase-Dyson-Lieb concavity in an interpolation theory,” Communications in Mathematical Physics, vol. 54, no. 1, pp. 21–32, 1977.
- [4] F. Hiai and D. Petz, “The proper formula for relative entropy and its asymptotics in quantum probability,” Communications in Mathematical Physics, vol. 143, no. 1, pp. 99–114, December 1991.
- [5] T. Ogawa and H. Nagaoka, “Strong converse and Stein’s lemma in quantum hypothesis testing,” IEEE Transactions on Information Theory, vol. 46, no. 7, pp. 2428–2433, November 2000.
- [6] I. Bjelakovic and R. Siegmund-Schultze, “Quantum Stein’s lemma revisited, inequalities for quantum entropies, and a concavity theorem of Lieb,” July 2012, arXiv:quant-ph/0307170.
- [7] S. Kullback and R. A. Leibler, “On information and sufficiency,” Ann. Math. Stat., vol. 22, no. 1, pp. 79–86, March 1951.
- [8] A. Rényi, “On measures of entropy and information,” Proceedings of the 4th Berkeley Symposium on Mathematics, Statistics and Probability, vol. 1, pp. 547–561, 1961.
- [9] I. Csiszár, “Information type measure of difference of probability distributions and indirect observations,” Studia Scientiarum Mathematicarum Hungarica, vol. 2, pp. 299––318, 1967.
- [10] S. M. Ali and S. D. Silvey, “A general class of coefficients of divergence of one distribution from another,” Journal of the Royal Statistical Society. Series B (Methodological), vol. 28, no. 1, pp. 131–142, 1966.
- [11] T. Morimoto, “Markov processes and the -theorem,” Journal of the Physical Society of Japan, vol. 18, no. 3, pp. 328–331, 1963.
- [12] D. Petz, “Quasi-entropies for states of a von Neumann algebra,” Publ. RIMS, Kyoto University, vol. 21, pp. 787–800, 1985.
- [13] ——, “Quasi-entropies for finite quantum systems,” Reports in Mathematical Physics, vol. 23, pp. 57–65, 1986.
- [14] M. Tomamichel, R. Colbeck, and R. Renner, “A fully quantum asymptotic equipartition property,” IEEE Transactions on Information Theory, vol. 55, no. 12, pp. 5840–5847, December 2009.
- [15] V. P. Belavkin and P. Staszewski, “C*-algebraic generalization of relative entropy and entropy,” Annales de l’I.H.P. Physique théorique, vol. 37, no. 1, pp. 51–58, 1982.
- [16] H. Nagaoka, “The converse part of the theorem for quantum Hoeffding bound,” November 2006, arXiv:quant-ph/0611289.
- [17] M. Hayashi, “Error exponent in asymmetric quantum hypothesis testing and its application to classical-quantum channel coding,” Physical Review A, vol. 76, no. 6, p. 062301, December 2007.
- [18] M. Müller-Lennert, F. Dupuis, O. Szehr, S. Fehr, and M. Tomamichel, “On quantum Rényi entropies: a new definition and some properties,” J. Math. Phys., vol. 54, no. 12, p. 122203, December 2013.
- [19] M. M. Wilde, A. Winter, and D. Yang, “Strong converse for the classical capacity of entanglement-breaking and Hadamard channels via a sandwiched Rényi relative entropy,” Communications in Mathematical Physics, vol. 331, no. 2, pp. 593–622, October 2014.
- [20] M. M. Wilde, M. Tomamichel, and M. Berta, “Converse bounds for private communication over quantum channels,” IEEE Transactions on Information Theory, vol. 63, no. 3, pp. 1792–1817, March 2017.
- [21] M. Mosonyi and T. Ogawa, “Quantum hypothesis testing and the operational interpretation of the quantum Rényi relative entropies,” Comm. Math. Phys., vol. 334, no. 3, pp. 1617–1648, March 2015.
- [22] F. Hiai, M. Mosonyi, D. Petz, and C. Beny, “Quantum -divergences and error correction,” Reviews in Mathematical Physics, vol. 23, no. 7, pp. 691–747, August 2011.
- [23] D. Petz and M. B. Ruskai, “Contraction of generalized relative entropy under stochastic mappings on matrices,” Inf. Dim. Ana., Quantum Prob. and Related Topics, vol. 1, no. 1, pp. 83–89, January 1998.
- [24] F. Hiai and M. Mosonyi, “Different quantum -divergences and the reversibility of quantum operations,” Reviews in Mathematical Physics, vol. 29, no. 7, p. 1750023, August 2017.
- [25] K. Matsumoto, “A new quantum version of -divergence,” 2013, arXiv:1311.4722.
- [26] F. Hansen and G. K. Pedersen, “Jensen’s operator inequality,” Bulletin London Math. Soc., vol. 35, no. 4, pp. 553–564, July 2003.
- [27] R. Bhatia, Matrix Analysis. Springer, 1997.
- [28] F. Hansen, “The fast track to Löwner’s theorem,” Linear Algebra and its Applications, vol. 438, no. 11, pp. 4557–4571, June 2013.
- [29] W. F. Stinespring, “Positive functions on C*-algebras,” Proceedings of the American Mathematical Society, vol. 6, pp. 211–216, 1955.
- [30] D. Petz, “Sufficient subalgebras and the relative entropy of states of a von Neumann algebra,” Communications in Mathematical Physics, vol. 105, no. 1, pp. 123–131, March 1986.
- [31] ——, “Sufficiency of channels over von Neumann algebras,” Quarterly Journal of Mathematics, vol. 39, no. 1, pp. 97–108, 1988.
- [32] O. Klein, “Zur Quantenmechanischen Begründung des zweiten Hauptsatzes der Wärmelehre,” Z. Physik, vol. 72, pp. 767–775, 1931.
- [33] M. B. Ruskai, “Inequalities for quantum entropy: A review with conditions for equality,” Journal of Mathematical Physics, vol. 43, no. 9, pp. 4358–4375, September 2002.
- [34] R. L. Frank and E. H. Lieb, “Monotonicity of a relative Rényi entropy,” J. Math. Phys., vol. 54, no. 12, p. 122201, December 2013.
- [35] S. Beigi, “Sandwiched Rényi divergence satisfies data processing inequality,” J. Math. Phys., vol. 54, no. 12, p. 122202, December 2013.