This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Bayesian Logarithmic Derivative Type Lower Bounds for Quantum Estimation thanks: This work was supported in part by JSPS KAKENHI Grant Numbers JP21K11749, JP24K14816.

1st Jianchao Zhang Graduate School of Informatics and Engineering,
The University of Electro-Communications,
Tokyo, 182-8585 Japan
[email protected]
   2nd Jun Suzuki Graduate School of Informatics and Engineering,
The University of Electro-Communications,
Tokyo, 182-8585 Japan
[email protected]
Abstract

Bayesian approach for quantum parameter estimation has gained a renewed interest from practical applications of quantum estimation theory. Recently, a lower bound, called the Bayesian Nagaoka-Hayashi bound for the Bayes risk in quantum domain was proposed, which is an extension of a new approach to point estimation of quantum states by Conlon et al. (2021). The objective of this paper is to explore this Bayesian Nagaoka-Hayashi bound further by obtaining its lower bounds. We first obtain one-parameter family of lower bounds, which is an analogue of the Holevo bound in point estimation. Thereby, we derive one-parameter family of Bayesian logarithmic derivative type lower bounds in a closed form for the parameter independent weight matrix setting. This new bound includes previously known Bayesian lower bounds as special cases.

Index Terms:
Quantum estimation, Bayes risk, Bayesian Nagaoka-Hayashi bound

I Introduction

Quantum estimation are pervasive in real-world scenarios, ranging from tracking moving objects to implementing quantum sensors [1, 2]. These problems often involve multiple unknown parameters, such as determining the velocity of an object or quantifying phase information [3, 4]. In tackling these challenges, existing literature has extensively relied on bounds like the multiparameter Cramér-Rao bound and its quantum extension by Helstrom [5]. While powerful, these bounds have limitations, particularly when faced with practical constraints such as limited measurement data and/or moderate prior knowledge.

Recent research has explored alternatives, including the Holevo bound, which offers more information but still requires locally unbiased estimators or infinite numbers of experimental repetitions to make it sense [6]. To overcome these difficulty, Conlon et al. [7] proposed a new approach which is applicable for locally unbiased and finite sample setting. This new bound was named after the historical contribution as the Nagaoka-Hayashi bound. To address these challenges in practical situation, the Bayesian framework emerges as a promising approach due to its ability to incorporate prior information effectively without assumption of unbiasedness. This was initiated by Personick in his seminal work and it established optimal lower bound for one-parameter case [8]. Since Personick’s work, many attempts were made to extend it to multiparameter estimation. However, current solutions still face many challenges, in particular when the Bayes risk is defined by parameter-dependent weight matrix. Tsang [9] proposed a quantum version of the Gill-Levit bound [10] to incorporate parameter-dependent weight matrix. The Tsang bound, however, is expressed as non-trivial optimization. In a recent work, the Bayesian version of the Nagaoka-Hayashi bound was proposed in terms of a semidefinite programming problem, which is efficiently computable [11].

In Suzuki’s work [11], it was also attempted to derive lower bounds which is easy to evaluate than the Bayesian Nagaoka-Hayshi bound, but only a partial solution was given. In this paper, we aim at deriving the more explicit form of the Bayesian lower bound for weighted trace of the mean squared error (MSE) with parameter-independent weight matrices, which is expressed only in terms of averaged quantities over the prior information. This lower bound (Theorem 7) is regarded as a Bayesian version of the Holevo bound in point estimation, since it is given as optimization of a function of the set of Hermitian operators. Based on our result, the close form of one parameter family in parameter-independent weight matrix follows up immediately from our result (Theorem 8). It is shown that the latter bound is a generalization of previously known Bayesian lower bounds, since it is reduced to the Personick bound and its variants as special cases.

II Preliminaries

II-A Optimal Bayesian MSE for classical estimation

In this section, we recall the Bayesian classical case at first. The notations referring to the Bayesian statistics are listed as bellows:

  • pθ(x)=p(x|θ)p_{\theta}(x)=p(x|\theta): A parametric model (likelihood) = A family of probability distribution on 𝒳\mathcal{X}.

  • π(θ)\pi(\theta): A prior distribution for the parameter space Θ\Theta.

  • q(θ|x)q(\theta|x): A posterior distribution for a datum x𝒳x\in{\cal X}.

  • px(x)p_{x}(x): A marginal distribution (evidence).

The expectation value of a random variable XX with respect to the model pθp_{\theta} is denoted by Eθ[X]=xxpθ(x)𝑑xE_{\theta}[X]=\int_{x}x\,p_{\theta}(x)dx, whereas the expectation value with respect to the joint distribution πpθ\pi p_{\theta} is expressed without subscript, and the expectation value with respect to the marginal pXp_{X} is denoted as EX[X]E_{X}[X]. An estimator that returns values on the set Θ\Theta is denoted as θ^(x)=(θ^1(x),θ^2(x),,θ^n(x))\hat{\theta}(x)=(\hat{\theta}_{1}(x),\hat{\theta}_{2}(x),\cdots,\hat{\theta}_{n}(x)). The main objective of this proceedings is to minimize the average of the weighted trace of the MSE matrix. We consider the most general problem with a parameter-dependent weight matrix, which is n×nn\times n positive definite matrix: W(θ)=[Wjk(θ)]>0W(\theta)=[W_{jk}(\theta)]>0. The Bayesian risk is then defined by

RB[θ^|W]=𝑑θπ(θ)Tr[W(θ)Vθ[θ^]],R_{\mathrm{B}}[\hat{\theta}|W]=\int d\theta\pi(\theta)\mathrm{Tr}\left[W(\theta)V_{\theta}[\hat{\theta}]\right], (1)

where Vθ[θ^]V_{\theta}[\hat{\theta}] is the MSE matrix with the j,kj,k elements

Vθ,jk[θ^]=Eθ[(θ^j(X)θj)(θ^k(X)θk)].V_{\theta,jk}[\hat{\theta}]=E_{\theta}\left[(\hat{\theta}_{j}(X)-\theta_{j})(\hat{\theta}_{k}(X)-\theta_{k})\right]. (2)

Throughout the paper, Tr[]\mathrm{Tr}[\cdot] is the trace over nn-dimensional parameter space.

The optimal estimator is known to be given as follows. First define an inner product on the space of nn-dimensional vectors of random variables F(X)=(f1(X),f2(X),,fn(X))F(X)=(f_{1}(X),f_{2}(X),\ldots,f_{n}(X)) and G=(g1(X),g2(X),,gn(X))G=(g_{1}(X),g_{2}(X),\ldots,g_{n}(X)) by F|G:=j=1nEX[fj(X)gj(X)]\langle F|G\rangle:=\sum_{j=1}^{n}E_{X}[f_{j}(X)g_{j}(X)]. The Bayes risk is minimized by completing the square as

RB[θ^|W]=θ^|Wpostθ^2dpost|θ^+w¯dpost|(Wpost)1dpost+w¯,\begin{split}R_{\mathrm{B}}[\hat{\theta}|W]&=\langle\hat{\theta}|{W}^{\mathrm{post}}\hat{\theta}\rangle-2\langle{d}^{\mathrm{post}}|\hat{\theta}\rangle+\overline{w}\\ &\geq\langle{d}^{\mathrm{post}}|({W}^{\mathrm{post}})^{-1}{d}^{\mathrm{post}}\rangle+\overline{w},\end{split}

where the first equality follows by defining the following quantities.

Wijpost(x):=𝑑θq(θ|x)Wij(θ),djpost(x):=i=1n𝑑θq(θ|x)Wij(θ)θi,w¯:=𝑑θπ(θ)i,j=1nθjWij(θ)θj.\begin{split}{W}^{\mathrm{post}}_{ij}(x)&:=\int d\theta q(\theta|x)W_{ij}(\theta),\\ {d}^{\mathrm{post}}_{j}(x)&:=\sum_{i=1}^{n}\int d\theta q(\theta|x)W_{ij}(\theta)\theta_{i},\\ \overline{w}&:=\int\,d\theta\,\pi(\theta)\sum_{i,j=1}^{n}\theta_{j}W_{ij}(\theta)\theta_{j}.\end{split} (3)

Furthermore, the optimal estimator is identified as

θ^opt,i(x)=j=1n((Wpost(x))1)ijdjpost(x).\hat{\theta}_{\mathrm{opt},i}(x)=\sum_{j=1}^{n}\left(({W}^{\mathrm{post}}(x))^{-1}\right)_{ij}{d}^{\mathrm{post}}_{j}(x). (4)

However, this result cannot be extended in the quantum case directly because a quantum version of dpost{d}^{\mathrm{post}} is not uniquely defined due to non-commutativity.

II-B Quantum Bayesian estimation

We aim at deriving lower bounds in another direction. Before that, we need to define the problem in quantum Bayesian setting. Let \mathcal{H} be a finite dimensional Hilbert space and {Sθ|θΘ}\left\{S_{\theta}~{}|~{}\theta\in\Theta\right\} be an nn-parameter quantum model with θ=(θ1,θ2,,θn)\theta=(\theta_{1},\theta_{2},\cdots,\theta_{n}). In the following, we consider SθS_{\theta} is full rank for all θΘ\theta\in\Theta and the map θSθ\theta\rightarrow S_{\theta} is one-to-one. Let a set of positive semidefinite matrices {Πx}\{\Pi_{x}\} be a measurement with the measurement outcome x𝒳x\in{\cal X}. The quantum measurement is normally referred to as a positive operator-valued measure (POVM), which is defined by

Π={Πx},x𝒳,Πx0,𝒳𝑑xΠx=I,\Pi=\{\Pi_{x}\},~{}\forall x\in{\cal X},~{}\Pi_{x}\geq 0,~{}\int_{\cal X}dx\Pi_{x}=I, (5)

where II is the identity operator on \mathcal{H}.

Measurement outcome is described by a random variable XX that obeys the conditional probability distribution:

pθ(x)=tr[SθΠx],p_{\theta}(x)=\mathrm{tr}\left[S_{\theta}\Pi_{x}\right],

where tr[]\mathrm{tr}[\cdot] denotes the trace on the Hilbert space \mathcal{H}. The performance of the estimator in this study is quantified by the quantum Bayes risk:

RB[Π,θ^]=𝑑θπ(θ)Tr[W(θ)Vθ[Π,θ^]].R_{\mathrm{B}}[\Pi,\hat{\theta}]=\int d\theta\,\pi(\theta)\mathrm{Tr}\left[W(\theta)V_{\theta}[\Pi,\hat{\theta}]\right]. (6)

The main objective is to find the best pair (Π,θ^)(\Pi,\hat{\theta}) that minimizes the Bayes risk.

II-C Bayesian Nagaoka-Hayashi bound

In a recent work, we have proposed a new class of Bayesian lower bound, called the Bayesian Nagaoka-Hayashi bound [11]. Before we recall this bound, we obtain the alternative form of the Bayes risk. We first introduce a new set of variables for quantum measurement and estimator by

𝕃jk[Π,θ^]\displaystyle\mathbb{L}_{jk}[\Pi,\hat{\theta}] =dxθ^j(x)Πxθ^k(x)(j,k=1,2,,n),\displaystyle=\int dx\hat{\theta}_{j}(x)\Pi_{x}\hat{\theta}_{k}(x)\quad(j,k=1,2,\dots,n),
Xj[Π,θ^]\displaystyle X_{j}[\Pi,\hat{\theta}] =𝑑xθ^j(x)Πx(j=1,2,,n).\displaystyle=\int dx\hat{\theta}_{j}(x)\Pi_{x}\quad(j=1,2,\ldots,n).

In the following, we will omit the argument [Π,θ^][\Pi,\hat{\theta}] when it is clear. The key idea is to regard the above quantities as the operator-valued matrix and vector [12, 7]. We introduce the extended Hilbert space by =n\mathbb{H}={\mathbb{C}}^{n}\otimes{\cal H}, and then 𝕃\mathbb{L} is identified as a matrix on \mathbb{H}, whose block matrix representation is given by [𝕃jk][\mathbb{L}_{jk}].

Next, define the following quantities defined by the model and prior distribution.

𝕊(θ)\displaystyle\mathbb{S}(\theta) :=[𝕊ij(θ)]with𝕊ij(θ):=Wij(θ)Sθ,\displaystyle:=[\mathbb{S}_{ij}(\theta)]\ \mbox{with}\ \mathbb{S}_{ij}(\theta):=W_{ij}(\theta)S_{\theta},
D(θ)\displaystyle D(\theta) :=[Di(θ)]withDi(θ):=j=1nWij(θ)θjSθ.\displaystyle:=[D_{i}(\theta)]\ \mbox{with}\ D_{i}(\theta):=\sum_{j=1}^{n}W_{ij}(\theta)\theta_{j}S_{\theta}.

Then the next lemma is immediate.

Lemma 1 (Suzuki [11] Lemma 2).

The Bayes risk is expressed as

RB[Π,θ^]\displaystyle R_{\mathrm{B}}[\Pi,\hat{\theta}] =𝕋r[𝕊¯𝕃]𝕋r[D¯XT1]𝕋r[XD¯T1]+w¯,\displaystyle=\mathds{T}\mathrm{r}{\left[\overline{\mathbb{S}}\mathbb{L}\right]}-\mathds{T}\mathrm{r}{\left[\overline{D}X^{T_{1}}\right]}-\mathds{T}\mathrm{r}{\left[X\overline{D}^{T_{1}}\right]}+\overline{w},
where𝕊¯\displaystyle\mathrm{where\ }\overline{\mathbb{S}} :=𝑑θπ(θ)𝕊(θ),\displaystyle:=\int\,d\theta\,\pi(\theta)\mathbb{S}(\theta), (7)
D¯\displaystyle\overline{D} :=𝑑θπ(θ)D(θ).\displaystyle:=\int\,d\theta\,\pi(\theta)D(\theta). (8)

In this lemma, ¯\overline{\cdot} denotes the averaged operators with respect to the prior distribution. XT1X^{T_{1}} means the transpose over the parameter space and 𝕋r[]\mathds{T}\mathrm{r}[\cdot] is the trace over both the Hilbert space and the parameter space.

The fundamental theorem of Ref. [11] is the lower bound for the Bayes risk RB[Π,θ^]R_{\mathrm{B}}[\Pi,\hat{\theta}].

Theorem 2 (Bayesian Nagaoka-Hayashi bound, Suzuki [11] Theorem 1).

For any POVM Π\Pi and estimator θ^\hat{\theta}, the following lower bound holds for the Bayes risk.

RB[Π,θ^]𝒞BNH\displaystyle R_{\mathrm{B}}[\Pi,\hat{\theta}]\geq{\cal C}_{\mathrm{BNH}}
𝒞BNH:=min𝕃,X{𝕋r[𝕊¯𝕃]𝕋r[D¯XT1]𝕋r[XD¯T1]}+w¯.\displaystyle{\cal C}_{\mathrm{BNH}}:=\min_{\mathbb{L},X}\left\{\mathds{T}\mathrm{r}{\left[\overline{\mathbb{S}}\mathbb{L}\right]}-\mathds{T}\mathrm{r}{\left[\overline{D}X^{T_{1}}\right]}-\mathds{T}\mathrm{r}{\left[X\overline{D}^{T_{1}}\right]}\right\}+\overline{w}.

Here optimization is subject to the constraints: j,k,𝕃jk=𝕃kj\forall j,k,\mathbb{L}_{jk}=\mathbb{L}_{kj}, 𝕃jk\mathbb{L}_{jk} is Hermitian, XjX_{j} is Hermitian, and 𝕃XXT1\mathbb{L}\geq{X}X^{T_{1}}.

II-D Bayesian Holevo-type bound

In Ref. [11], a Bayesian version of the Holevo bound was also proposed. We recall that the Bayesian Holevo-type bound is lower than the Bayesian Nagaoka-Hayashi bound in general.

Theorem 3 (Bayesian Holevo-type bound, Suzuki [11] Theorem 3).
𝒞BNH\displaystyle{\cal C}_{\mathrm{BNH}} 𝒞BH,\displaystyle\geq\mathcal{C}_{\mathrm{BH}},
𝒞BH\displaystyle\mathcal{C}_{\mathrm{BH}} =minXj:Hermitian{dθπ(θ)Tr[ReZθ(X)]\displaystyle=\min_{X_{j}:\mathrm{\,Hermitian}}\{\int d\theta\pi(\theta)\mathrm{Tr}\left[\mathrm{Re}\,Z_{\theta}(X)\right]
+𝑑θ\displaystyle+\int d\theta π(θ)Tr|ImZθ(X)|𝕋r[D¯XT1]𝕋r[XD¯T1]}+w¯,\displaystyle\pi(\theta)\mathrm{Tr}|\mathrm{Im}\,Z_{\theta}(X)|-\mathds{T}\mathrm{r}\left[\overline{D}X^{T_{1}}\right]-\mathds{T}\mathrm{r}{\left[X\overline{D}^{T_{1}}\right]}\}+\overline{w},

where

Zθ(X)=tr(W(θ)SθXXT1W(θ)Sθ).Z_{\theta}(X)=\mathrm{tr}\left(\sqrt{W(\theta)}\otimes\sqrt{S_{\theta}}XX^{T_{1}}\sqrt{W(\theta)}\otimes\sqrt{S_{\theta}}\right).

Here Re(Im)\mathrm{Re}\,(\mathrm{Im}\,) denote the element-wise real (imaginary) part of a matrix and |A|=AA|A|=\sqrt{A^{\dagger}A}. In other words, Tr|A|\mathrm{Tr}|A| is the absolute sum of all eigenvalues of AA.

This lower bound is only expressed in terms of minimization over the set of Hermitian matrices X=(X1,X2,,Xn)X=(X_{1},X_{2},\ldots,X_{n}). However, it is still hard to analyze it further, since the second term is expressed as an integration of functions containing variable XX.

III Results

In this section, we shall propose a method of deriving a family of Holevo-type lower bounds based on a new technique. This is to apply an appropriate map that satisfies a certain conditions. Before we present our result, we first examine the direct analogue of the optimal classical bound explained in Sec. II-A.

III-A Direct generalization of the Bayesian classically optimum bound

Firstly, we consider to generalize the Bayesian classically optimum bound in quantum state with θ\theta-dependent weight matrix W(θ)W(\theta). This is done by apply the constraint 𝕃XXT1\mathbb{L}\geq{X}X^{T_{1}} in the definition of the Bayesian Nagaoka-Hayashi bound.

Theorem 4 (Quantum Bayesian lower bound 1).
R𝖡𝒞direct=D¯|𝕊¯1D¯𝕊¯+w¯,R_{\mathsf{B}}\geq\mathcal{C}_{\mathrm{direct}}=-\langle\overline{D}|\overline{\mathbb{S}}^{-1}\overline{D}\rangle_{\overline{\mathbb{S}}}+\overline{w}, (9)

where the inner product is defined for operator-valued vector X=[Xj]X=[X_{j}] and Y=[Yj]Y=[Y_{j}] as

X,Y𝕊¯:=tr[XT1𝕊¯Y].\langle X,Y\rangle_{\overline{\mathbb{S}}}:=\mathrm{tr}\left[X^{T_{1}}\overline{\mathbb{S}}Y\right]. (10)
Proof.

This can be proved by completing the square as in the classical case after the use of 𝕃XXT1\mathbb{L}\geq{X}X^{T_{1}}.

R𝖡𝕋r[𝕊¯XXT1]2𝕋r(D¯XT1)+w¯=tr(XT1𝕊¯X)2tr[XT1D¯]+w¯=X|X𝕊¯2𝕊¯1D¯|X𝕊¯+w¯D¯|𝕊¯1D¯𝕊¯+w¯,\begin{split}R_{\mathsf{B}}&\geq\mathds{T}\mathrm{r}[\overline{\mathbb{S}}XX^{T_{1}}]-2\mathds{T}\mathrm{r}\left(\overline{D}X^{T_{1}}\right)+\overline{w}\\ &=\mathrm{tr}\left(X^{T_{1}}\overline{\mathbb{S}}X\right)-2\mathrm{tr}\left[X^{T_{1}}\overline{D}\right]+\overline{w}\\ &=\langle X|X\rangle_{\overline{\mathbb{S}}}-2\langle\overline{\mathbb{S}}^{-1}\overline{D}|X\rangle_{\overline{\mathbb{S}}}+\overline{w}\\ &\geq-\langle\overline{D}|\overline{\mathbb{S}}^{-1}\overline{D}\rangle_{\overline{\mathbb{S}}}+\overline{w},\end{split} (11)

where the optimizer for the inequality is given by

Xopt,i=j=1n(𝕊¯1)ijDj¯.X_{\mathrm{opt},i}=\sum_{j=1}^{n}({\overline{\mathbb{S}}}^{-1})_{ij}\overline{D_{j}}.

The proof of this bound is straightforward. However, this cannot be tight because Xj,(𝕊¯1)ijX_{j},({\overline{\mathbb{S}}}^{-1})_{ij} (i,ji,jth element of (𝕊¯)1(\overline{\mathbb{S}})^{-1}), and Di¯\overline{D_{i}} are Hermitian, but 𝕊¯ij1\overline{\mathbb{S}}^{-1}_{ij} and Di¯\overline{D_{i}} are not commute. This means that optimal XoptX_{\mathrm{opt}} cannot be Hermitian. This bound is the direct generalization of the Bayesian classical one. If we set the state as classical we can immediately degenerate in the previous bound.

III-B New quantum Bayesian lower bounds

Next, we derive a new one-parameter family of lower bounds which is tighter than the direct generalization. The key idea in this proceedings is to define a map \mathcal{E} from (n)\mathcal{L}(\mathds{C}^{n}\otimes\mathcal{H}) to (n)\mathcal{L}(\mathds{C}^{n}) such that satisfies the condition.

Definition 5.

We say that \mathcal{E} is real symmetric, if (𝕃)n×n\mathcal{E}(\mathbb{L})\in{\mathbb{C}}^{n\times n} is real symmetric for all 𝕃sym(n)\mathbb{L}\in\mathcal{L}_{\mathrm{sym}}(\mathds{C}^{n}\otimes\mathcal{H}).

Here sym(n)\mathcal{L}_{\mathrm{sym}}(\mathds{C}^{n}\otimes\mathcal{H}) denotes the set of all Hermitian 𝕃\mathbb{L} which is invariant under the partial transposition T1\cdot^{T_{1}}. Using the block-matrix representation, this is equivalent to i,j,𝕃ij=𝕃ji\forall i,j,\,\mathbb{L}_{ij}=\mathbb{L}_{ji}.

Given a positive operator 𝕊\mathbb{S} on =n\mathbb{H}=\mathds{C}^{n}\otimes\mathcal{H}, we define one-parameter family of maps for 𝕏(n)\mathds{X}\in\mathcal{L}(\mathds{C}^{n}\otimes\mathcal{H}).

(λ)(𝕏):=1+λ2(+)(𝕏)+1λ2()(𝕏),\mathcal{E}^{(\lambda)}(\mathds{X}):=\frac{1+\lambda}{2}\mathcal{E}^{(+)}(\mathds{X})+\frac{1-\lambda}{2}\mathcal{E}^{(-)}(\mathds{X}), (12)

where

(+)(𝕏)\displaystyle\mathcal{E}^{(+)}(\mathds{X}) :=tr[(IS)𝕏(IS)]\displaystyle:=\mathrm{tr}\left[(I\otimes\sqrt{S})\mathds{X}(I\otimes\sqrt{S})\right] (13)
()(𝕏)\displaystyle\mathcal{E}^{(-)}(\mathds{X}) :=tr[(IS)𝕏T1(IS)]\displaystyle:=\mathrm{tr}\left[(I\otimes\sqrt{S})\mathds{X}^{T_{1}}(I\otimes\sqrt{S})\right] (14)

We can show the next property.

Lemma 6.

A map (λ)\mathcal{E}^{(\lambda)} is real symmetric and positive for λ[1,1]\lambda\in[-1,1].

Proof.

To show (+)\mathcal{E}^{(+)} is real symmetry we decompose 𝕏=iαiAiBi\mathds{X}=\sum_{i}\alpha_{i}A_{i}\otimes B_{i}. Then we obtain the form of (+)(𝕏)\mathcal{E}^{(+)}(\mathds{X}) in

(+)(𝕏)=iαiAitrSBiand(+)(𝕏)T=iαiAiTtrSBi.\displaystyle\mathcal{E}^{(+)}(\mathds{X})=\sum_{i}\alpha_{i}A_{i}\mathrm{tr}SB_{i}\ \mathrm{and}\ \mathcal{E}^{(+)}(\mathds{X})^{T}=\sum_{i}\alpha_{i}A_{i}^{T}\mathrm{tr}SB_{i}. (15)

We notice that 𝕏sym,h()\mathds{X}\in\mathcal{L}_{sym,h}(\mathcal{H}), this means AiT=AiA_{i}^{T}=A_{i}. Thus (+)(𝕏)=(+)(𝕏)T\mathcal{E}^{(+)}(\mathds{X})=\mathcal{E}^{(+)}(\mathds{X})^{T}.

On the other hand, (+)(𝕏)\mathcal{E}^{(+)}(\mathds{X}) is a real matrix. Since AA is positive and AT=AA^{T}=A, we obtain An×nA\in\mathds{R}^{n\times n}. By BB is hermitian, we get trSB\mathrm{tr}SB\in\mathds{R}. Thus (+)(𝕏)n×n\mathcal{E}^{(+)}(\mathds{X})\in\mathds{R}^{n\times n}. ∎

After identifying a real symmetric map, we can derive a family of lower bounds, which has a similar structure as the Holevo bound in point estimation.

Theorem 7 (Quantum Bayesian lower bound 2).
𝒞BNH𝒞BH(λ):=minX{Tr[WReZB(λ)(X)]+Tr|WImZB(λ)(X)W|𝕋r[WD¯XT1]𝕋r[WXD¯T1]+Tr[WM]},\begin{split}\mathcal{C}_{\mathrm{BNH}}\geq\mathcal{C}_{\mathrm{BH}}^{(\lambda)}:=\min_{X}\{\mathrm{Tr}\left[W\mathrm{Re}\,Z_{B}^{(\lambda)}(X)\right]\\ +\mathrm{Tr}|\sqrt{W}\mathrm{Im}\,Z_{B}^{(\lambda)}(X)\sqrt{W}|\\ -\mathds{T}\mathrm{r}\left[W\overline{D}X^{T_{1}}\right]-\mathds{T}\mathrm{r}{\left[WX\overline{D}^{T_{1}}\right]}+\mathrm{Tr}[WM]\},\end{split} (16)

where

ZB(λ)(X)\displaystyle Z_{B}^{(\lambda)}(X) :=[1+λ2tr(XiSBXj)+1λ2tr(XjSBXi)],\displaystyle:=\left[\frac{1+\lambda}{2}\mathrm{tr}(X_{i}S_{B}X_{j})+\frac{1-\lambda}{2}\mathrm{tr}(X_{j}S_{B}X_{i})\right], (17)
=[Xi,XjSB(λ)]\displaystyle=\left[\langle X_{i},X_{j}\rangle^{(\lambda)}_{S_{B}}\right] (18)

is an n×nn\times n complex positive semidefinite matrix, and minimization is subject to XjX_{j}: Hermitian. Furthermore, the right hand side the bound (16) takes the maximum value when λ=±1\lambda=\pm 1.

maxλ[1,1]=𝒞BH(λ=±1).\max_{\lambda\in[-1,1]}=\mathcal{C}_{\mathrm{BH}}^{(\lambda=\pm 1)}.
Proof.

First, let us prove the inequality (16). From the original form of the Bayesian Nagaoka-Hayashi bound of Theorem 2, we see that 𝕃,X\mathbb{L},X satisfying the constraints also satisfy (λ)(𝕃XXT1)0\mathcal{E}^{(\lambda)}(\mathbb{L}-XX^{T_{1}})\geq 0. Define V:=(λ)(𝕃)=𝕋r[𝕊¯𝕃]V:=\mathcal{E}^{(\lambda)}(\mathbb{L})=\mathds{T}\mathrm{r}{\left[\overline{\mathbb{S}}\mathbb{L}\right]}, then VV is real symmetric and hence we can relax the minimization problem in terms of VV and XX:

𝒞BNH\displaystyle\mathcal{C}_{\mathrm{BNH}} minV,X{Tr[WV]𝕋r[WD¯XT1]𝕋r[WXD¯T1]\displaystyle\geq\min_{V,X}\{\mathrm{Tr}[WV]-\mathds{T}\mathrm{r}\left[W\overline{D}X^{T_{1}}\right]-\mathds{T}\mathrm{r}\left[WX\overline{D}^{T_{1}}\right] (19)
+Tr[WM]},\displaystyle+\mathrm{Tr}[WM]\}, (20)

where the constraints are VV: real symmetric, VZ(λ)(X)V\geq Z^{(\lambda)}(X), and XjX_{j}: Hermitian. From the well-known lemma (Lemma 6.6.1 in Ref. [6]), we have minV{Tr[WV]|VZ(λ)(X)}=Tr[WReZ(λ)(X)]+Tr|WImZ(λ)(X)W|\min_{V}\{\mathrm{Tr}[WV]\,|\,V\geq Z^{(\lambda)}(X)\}=\mathrm{Tr}\left[W\mathrm{Re}\,Z^{(\lambda)}(X)\right]+\mathrm{Tr}|\sqrt{W}\mathrm{Im}\,Z^{(\lambda)}(X)\sqrt{W}|. This proves the first part of the theorem.

Next, we show that this family of bounds takes the maximum value when λ=±1\lambda=\pm 1, i.e.,

𝒞BH(λ=±1)𝒞BH(λ)forallλ[1,1].\mathcal{C}_{\mathrm{BH}}^{{(\lambda=\pm 1)}}\geq\mathcal{C}_{\mathrm{BH}}^{(\lambda)}~{}\rm{for~{}all~{}}\lambda\in[-1,1]. (21)

By separating the items, we see the real part of Z(λ)Z^{(\lambda)} is independent of λ\lambda as

ReZ(λ)(X)=12Z(λ)(X)+12(Z(λ)(X)T)=12Z(λ)(X)+12Z(λ)(X)=12tr[{𝕊¯,12[XXT1+(XXT1)T1]}],\begin{split}\mathrm{Re}\,Z^{(\lambda)}(X)&=\frac{1}{2}Z^{(\lambda)}(X)+\frac{1}{2}(Z^{(\lambda)}(X)^{T})\\ &=\frac{1}{2}Z^{(\lambda)}(X)+\frac{1}{2}Z^{(-\lambda)}(X)\\ &=\frac{1}{2}\mathrm{tr}\left[\left\{\overline{\mathbb{S}}\,,\,\frac{1}{2}\left[XX^{T_{1}}+(XX^{T_{1}})^{T_{1}}\right]\right\}\right],\end{split} (22)

where the second line is due to (tr[𝕊¯𝕏])T=tr[𝕏T1𝕊¯](\mathrm{tr}[\overline{\mathbb{S}}\mathbb{X}])^{T}=\mathrm{tr}[\mathbb{X}^{T_{1}}\overline{\mathbb{S}}] and symmetrized multiplication of 𝕊¯,𝕏\bar{\mathbb{S}},\mathbb{X}. The imaginary part is, on the other hand, proportional to λ\lambda, since

ImZ(λ)(X)=12i(Z(λ)(X)Z(λ)(X)T)=12i(Z(λ)(X)Z(λ)(X))=λ2tr[{𝕊¯,i2[XXT1(XXT1)T1]}],\begin{split}\mathrm{Im}\,Z^{(\lambda)}(X)&=\frac{1}{2i}\left(Z^{(\lambda)}(X)-Z^{(\lambda)}(X)^{T}\right)\\ &=\frac{1}{2i}(Z^{(\lambda)}(X)-Z^{(-\lambda)}(X))\\ &=\frac{\lambda}{2}\mathrm{tr}\left[\left\{\overline{\mathbb{S}}\,,\,\frac{i}{2}\big{[}XX^{T_{1}}-(XX^{T_{1}})^{T_{1}}\big{]}\right\}\right],\end{split} (23)

where {X,Y}:=XY+YX\{X,Y\}:=XY+YX is the anti-commutator. Due to it, we can see the trace norm of imaginary part is proportional to absolute value of λ\lambda. In other words, Tr|ImZ(λ)(X)||λ|\mathrm{Tr}|\mathrm{Im}\,Z^{(\lambda)}(X)|\propto|\lambda|. This means that the one-parameter family of the quantum Bayesian lower bounds takes the maximum value at λ=±1\lambda=\pm 1. ∎

We now discuss the advantages of the proposed bound:

  • The two Bayesian Holevo-type bounds are similar; both are expressed as minimization over XX operators. However as an optimization problem, the new one is easier to calculate since the previous bound (Theorem 3) requires to compute the integral over a complicated function of XX, namely 𝑑θπ(θ)Tr|ImZθ(X)|\int d\theta\pi(\theta)\mathrm{Tr}|\mathrm{Im}\,Z_{\theta}(X)|. In contrast, the new bound refers to the averaged state 𝕊¯=𝑑θW(θ)Sθ\overline{\mathbb{S}}=\int d\theta W(\theta)\otimes S_{\theta} only. Thus, what we need is to calculate the averaged state in advance.

  • This new bound is tighter than the direct quantum version of the optimal Bayesian bound (Theorem 4).

    𝒞BH(λ)𝒞direct.\mathcal{C}_{\mathrm{BH}}^{(\lambda)}\geq\mathcal{C}_{\mathrm{direct}}. (24)

    The reason is that we can show the real part of the bound (22) is equal to the first term of function in the proof (11) and the imaginary part (23) is non-negative.

    Tr[ReZ(λ)(X)]+Tr|ImZ(λ)(X)|Tr[ReZ(λ)(X)]=tr[XT1𝕊¯X].\mathrm{Tr}[\mathrm{Re}\,Z^{(\lambda)}(X)]+\mathrm{Tr}|\mathrm{Im}\,Z^{(\lambda)}(X)|\\ \geq\mathrm{Tr}[\mathrm{Re}\,Z^{(\lambda)}(X)]=\mathrm{tr}[X^{T_{1}}\overline{\mathbb{S}}X].
  • Lastly, the family of our new bounds immediately gives one-parameter family generalization of the previously established Bayesian lower bounds when the weight matrix is parameter independent. This will be explored in the next subsection.

III-C Bayesian λ\lambdaLD bound in parameter-independent weight

In this subsection, we take the θ\theta-independent weight matrix. The Bayesian MSE is defined as

VB[Π,θ^]:=𝑑θπ(θ)Vθ[Π,θ^].V_{B}[\Pi,\hat{\theta}]:=\int d\theta\pi(\theta)V_{\theta}[\Pi,\hat{\theta}]. (25)

The goal here is to derive a family of quantum Bayesian lower bounds in a closed form by approximating the optimization over XX operators. The key ingredient in our approach is to utilize the logarithmic derivative (LD) type equation which was shown to be useful for point estimation [13, 14]. For convenience, we call this family as the Bayesian λ\lambdaLD bound, even though this is different from the quantum Cramér-Rao bound based on quantum LDs.

We substitute parameter-independent weight matrix in the quantum Bayesian bound of Theorem 7 to get the following theorem.

Theorem 8 (Bayesian λ\lambdaLD bound).

For parameter independent weight matrix, the Bayes risk is bounded by 𝒞BLD(λ)(W)\mathcal{C}^{(\lambda)}_{\mathrm{BLD}}(W).

𝒞BLD(λ)(W):=Tr[WReK(λ)]+Tr|WImK(λ)W|+w¯,\mathcal{C}^{(\lambda)}_{\mathrm{BLD}}(W):=-\mathrm{Tr}\left[W\mathrm{Re}\,K^{(\lambda)}\right]+\mathrm{Tr}|\sqrt{W}\mathrm{Im}\,K^{(\lambda)}\sqrt{W}|+\overline{w},

for λ[1,1]\lambda\in[-1,1], where K(λ)K^{(\lambda)} is defined by

Kjk(λ)\displaystyle K^{(\lambda)}_{jk} :=tr[DB,kLj(λ)],\displaystyle:=\mathrm{tr}[{D}_{\mathrm{B},k}L^{(\lambda)}_{j}],
DB,j\displaystyle{D}_{\mathrm{B},j} =1+λ2SBLj(λ)+1λ2Lj(λ)SB,\displaystyle=\frac{1+\lambda}{2}{S}_{\mathrm{B}}L^{(\lambda)}_{j}+\frac{1-\lambda}{2}L^{(\lambda)}_{j}{S}_{\mathrm{B}},
SB\displaystyle{S}_{\mathrm{B}} =𝑑θπ(θ)Sθ,\displaystyle=\int\,d\theta\,\pi(\theta)S_{\theta},
DB,j\displaystyle{D}_{\mathrm{B},j} =𝑑θπ(θ)θjSθ.\displaystyle=\int\,d\theta\,\pi(\theta)\theta_{j}S_{\theta}.

In addition, the Bayes risk is bounded by

RB[Π,θ^]maxλ𝒞BLD(λ)(W)=:CBLDmax.R_{\mathrm{B}}[\Pi,\hat{\theta}]\geq\max_{\lambda}\mathcal{C}^{(\lambda)}_{\mathrm{BLD}}(W)=:C^{\max}_{\mathrm{BLD}}. (26)

We have two remarks before proving this theorem. First, note that the last form of the bound is a Bayesian version of the maximum logarithmic derivative Cramé-Rao bound proposed by Yamagata in point estimation [13]. Second, due to the process of deriving the Bayesian λ\lambdaLD bound, it is tighter than the Personick bound [8] and its generalization [15, 16], and the Bayesian bound proposed by Holevo [17, 18]. In other words, the above one-parameter family of bounds includes well-known bounds as special cases. This will be shown later.

Proof.

When WW is θ\theta-independent, we have

𝕊¯\displaystyle\overline{\mathbb{S}} =WSB,\displaystyle=W\otimes{S}_{\mathrm{B}},
D¯j\displaystyle\overline{D}_{j} =kWjkDB,k,\displaystyle=\sum_{k}W_{jk}{D}_{\mathrm{B},k},
w¯\displaystyle\overline{w} =Tr{WM},\displaystyle=\mathrm{Tr}\left\{WM\right\},

where Mjk:=𝑑θπ(θ)θjθkM_{jk}:=\int d\theta\,\pi(\theta)\theta_{j}\theta_{k}. Define one-parameter family of inner products by

X,YSB(λ):=tr[SB(1+λ2YX+1λ2XY)],\langle X,Y\rangle_{{S}_{\mathrm{B}}}^{(\lambda)}:=\mathrm{tr}\left[{S}_{\mathrm{B}}(\frac{1+\lambda}{2}YX^{\dagger}+\frac{1-\lambda}{2}X^{\dagger}Y)\right], (27)

for λ[1,1]\lambda\in[-1,1]. Then the Bayesian bound of Theorem 7 is expressed by using Eq. (19) as

𝒞BH(λ)=minV,X{Tr[WV]+w¯|VZB[X]HB[X]HB[X]T},\mathcal{C}_{\mathrm{BH}}^{(\lambda)}=\min_{V,X}\{\mathrm{Tr}[WV]+\overline{w}\,|\,V\geq{Z}_{\mathrm{B}}[X]-{H}_{\mathrm{B}}[X]-{H}_{\mathrm{B}}[X]^{T}\},

where ZB,jk(λ)[X]:=Xj,XkSB(λ)Z^{(\lambda)}_{\mathrm{B},jk}[X]:=\langle X_{j},X_{k}\rangle_{{S}_{\mathrm{B}}}^{(\lambda)}, HB,jk[X]:=tr[DB,jXk]H_{\mathrm{B},jk}[X]:=\mathrm{tr}{[{D}_{\mathrm{B},j}X_{k}]}, and VV is real symmetric. We can carry out the optimization over XX first to get a lower bound. To do this, we note the constraint is also written as

ZB[X]HB[X]HB[X]T=ZB[XL(λ)]ZB[L(λ)],{Z}_{\mathrm{B}}[X]-{H}_{\mathrm{B}}[X]-{H}_{\mathrm{B}}[X]^{T}={Z}_{\mathrm{B}}[X-L^{(\lambda)}]-{Z}_{\mathrm{B}}[L^{(\lambda)}],

which follows from the definition of Lj(λ)L^{(\lambda)}_{j}. Note that Lj(λ)L^{(\lambda)}_{j} are not Hermitian in general, yet we can obtain a lower bound by removing this condition on XX. Hence we have

𝒞BH(λ)minV{Tr[WV]+w¯|VZB[L(λ)]},\mathcal{C}_{\mathrm{BH}}^{(\lambda)}\geq\min_{V}\{\mathrm{Tr}[WV]+\overline{w}\,|\,V\geq-{Z}_{\mathrm{B}}[L^{(\lambda)}]\}, (28)

where VV real symmetric. The last step follows from the Lemma (Lemma 6.6.1 in Ref. [6]) to get

𝒞BH(λ)\displaystyle\mathcal{C}_{\mathrm{BH}}^{(\lambda)} Tr[WReZB[L(λ)]]+Tr|WImZB[L(λ)]W|+w¯\displaystyle\geq-\mathrm{Tr}[W\mathrm{Re}\,{Z}_{\mathrm{B}}[L^{(\lambda)}]]+\mathrm{Tr}|\sqrt{W}\mathrm{Im}\,{Z}_{\mathrm{B}}[L^{(\lambda)}]\sqrt{W}|+\overline{w}
=𝒞BLD(λ)(W),\displaystyle=\mathcal{C}^{(\lambda)}_{\mathrm{BLD}}(W),

since ZB[L(λ)]=K(λ){Z}_{\mathrm{B}}[L^{(\lambda)}]=K^{(\lambda)}. ∎

By the above proof, we can immediately show that the Personick bound is recovered when λ=0\lambda=0, whereas the Holevo’s result corresponds to λ=1\lambda=1. To do this, we can set θ\theta-independent WW as the rank-1 projector W=ccW=cc^{\dagger} with cnc\in\mathds{C}^{n}. Next, we obtain a lower bound for 𝒞BLD(λ)\mathcal{C}_{\mathrm{BLD}}^{(\lambda)}. This gives the desired result.

𝒞BLD(λ)cMc\displaystyle\mathcal{C}_{\mathrm{BLD}}^{(\lambda)}-c^{\dagger}Mc Lc(λ),Lc(λ)SB(λ)=cK(λ)c,\displaystyle\geq-\langle L_{c}^{(\lambda)},L_{c}^{(\lambda)}\rangle_{{S}_{\mathrm{B}}}^{(\lambda)}=-c^{\dagger}K^{(\lambda)}c,

where we set Lc(λ)=jcjLj(λ)L_{c}^{(\lambda)}=\sum_{j}c_{j}L_{j}^{(\lambda)}. Since this is true for any choice of cc, we prove the matrix inequality VBMK(λ)V_{\mathrm{B}}\geq M-K^{(\lambda)}. If we choose λ=0\lambda=0, thereby, we recover the Personick bound [8].

IV Conclusion

In summary, we introduce one-parameter family of the Bayesian Holevo type bounds, a new lower bound for the Bayes risk, which addresses the parameter-independent weight matrices in this domain. Our bound exhibits superior tightness compared to the direct extension of the Bayesian classical optimum bound, while maintaining computational simplicity relative to the Bayesian Holevo-type bound. Moreover, we derive the closed form of the Bayesian λ\lambdaLD type lower bounds. Our analysis demonstrates its superiority over existing lower bounds, including the symmetric logarithmic derivative type bound [8] and the right logarithmic derivative type bound [17, 18], which are obtained by setting λ\lambda to 0 and 1, respectively. In other words, the derived closed form offers a generalized framework that encompasses other bounds, enhancing its practical utility and theoretical significance.

References

  • [1] T. J. Proctor, P. A. Knott, and J. A. Dunningham, “Multiparameter estimation in networked quantum sensors,” Physical review letters, vol. 120, no. 8, p. 080501, 2018.
  • [2] H. T. Dinani, D. W. Berry, R. Gonzalez, J. R. Maze, and C. Bonato, “Bayesian estimation for quantum sensing in the absence of single-shot detection,” Physical Review B, vol. 99, no. 12, p. 125413, 2019.
  • [3] Q. Zhuang, Z. Zhang, and J. H. Shapiro, “Entanglement-enhanced lidars for simultaneous range and velocity measurements,” Physical Review A, vol. 96, no. 4, p. 040304, 2017.
  • [4] M. Szczykulska, T. Baumgratz, and A. Datta, “Reaching for the quantum limits in the simultaneous estimation of phase and phase diffusion,” Quantum Science and Technology, vol. 2, no. 4, p. 044004, 2017.
  • [5] C. W. Helstrom, Quantum detection and estimation theory.   Academic press, 1976.
  • [6] A. S. Holevo, Probabilistic and statistical aspects of quantum theory.   Edizioni della Normale, 2011.
  • [7] L. O. Conlon, J. Suzuki, P. K. Lam, and S. M. Assad, “Efficient computation of the nagaoka–hayashi bound for multiparameter estimation with separable measurements,” npj Quantum Information, vol. 7, no. 1, p. 110, 2021.
  • [8] S. Personick, “Application of quantum estimation theory to analog communication over quantum channels,” IEEE Transactions on Information Theory, vol. 17, no. 3, pp. 240–246, 1971.
  • [9] M. Tsang, “Physics-inspired forms of the bayesian cramér-rao bound,” Physical Review A, vol. 102, no. 6, p. 062217, 2020.
  • [10] R. D. Gill and B. Y. Levit, “Applications of the van trees inequality: a bayesian cramér-rao bound,” Bernoulli, pp. 59–79, 1995.
  • [11] J. Suzuki, “Bayesian nagaoka-hayashi bound for multiparameter quantum-state estimation problem,” IEICE Transactions on Fundamentals of Electronics, Communications and Computer Sciences, vol. 107, no. 3, pp. 510–518, 2024.
  • [12] M. Hayashi, “On simultaneous measurement of noncommutative observables,” Surikaisekikenkyusho (RIMS), Kyoto Univ., Kokyuroku (in japanese), no. 96, p. 1099, 1999.
  • [13] K. Yamagata, “Maximum logarithmic derivative bound on quantum state estimation as a dual of the holevo bound,” Journal of Mathematical Physics, vol. 62, no. 6, 2021.
  • [14] J. Suzuki, “Non-monotone metric on the quantum parametric model,” The European Physical Journal Plus, vol. 136, no. 1, pp. 1–26, 2021.
  • [15] J. Rubio and J. Dunningham, “Quantum metrology in the presence of limited data,” New Journal of Physics, vol. 21, no. 4, p. 043037, 2019.
  • [16] R. Demkowicz-Dobrzański, W. Górecki, and M. Guţă, “Multi-parameter estimation beyond quantum fisher information,” Journal of Physics A: Mathematical and Theoretical, vol. 53, no. 36, p. 363001, 2020.
  • [17] A. S. Holevo, “Noncommutative analogues of the cramér-rao inequality in the quantum measurement theory,” Proceedings of the Third Japan — USSR Symposium on Probability Theory, Lecture Notes in Mathematics, vol. 550, pp. 194–222, 1976.
  • [18] ——, “Commutation superoperator of a state and its applications to the noncommutative statistics,” Reports on mathematical physics, vol. 12, no. 2, pp. 251–271, 1977.