Logical Entropy
Open Access
Issue
4open
Volume 5, 2022
Logical Entropy
Article Number 2
Number of page(s) 14
Section Physics - Applied Physics
DOI https://doi.org/10.1051/fopen/2021005
Published online 25 January 2022

© B. Tamir et al., Published by EDP Sciences, 2022

Licence Creative CommonsThis is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

1 Introduction

Entropy is one of the paramount concepts in probability theory and physics. Even though information does not seem to have a precise definition, the Shannon entropy is seen as an important measure of information about a system of interest, while the Gibbs entropy plays a similar role in statistical mechanics. A possible and, in a sense, natural extension of these classical measures to the quantum realm is the von Neumann entropy. Despite playing a fundamental role in many applications in quantum information, the von Neumann entropy was criticized on several different grounds [13]. In a nutshell, while classical entropy indicates one’s ignorance about the system [4], quantum entropy seems to have a fundamentally different flavor, corresponding to an a priori inaccessibility of information or the existence of non-local correlations. In this perspective, classical entropy concerns subjective/epistemic indefiniteness, while quantum entropy is associated with some form of objective/ontological indefiniteness [5], although this reasoning can be contested. To address conceptual problems like this one, the non-additive Tsallis entropy and other measures were proposed [1, 6, 7].

Classical logical entropy was recently introduced by Ellerman [8, 9] as an informational measure arising from the logic of partitions. As such, this entropy gives the distinction of partitions of a set U. A partition π is defined as a set of disjoint parts of a set, as exemplified in Figure 1a. The set can be thought of as being originally fully distinct, while each partition collects together blocks whose distinctions are factored out. Each block represents elements that are associated with an equivalence relation on the set. Then, the elements of a block are indistinct among themselves while different blocks are distinct from each other, given an equivalence relation.

thumbnail Figure 1

Pictorial representation of a set with a partition and its distinction. (a) A set U with six elements ui is divided in a partition π with three blocks (green areas). (b) The set U × U has each of its points characterized by a square in a 2D mesh. Points that belong to dit(π), i.e., pairs whose components belong to different blocks are colored in green.

With these concepts in mind, it seems that the extension of this framework of partitions and distinctions to the study of quantum systems may bring new insights into problems of quantum state discrimination, quantum cryptography, and quantum channel capacity. In fact, in these problems, we are, in one way or another, interested in a distance measure between distinguishable states, which is exactly the kind of knowledge the logical entropy is associated with.

This work is an updated and much extended version of a previous preprint that proposed the study of quantum logical entropy [10]. In this new version, like in the original one, we focus mostly on basic definitions and properties of this quantity. Other advanced topics were either treated in previous research, like in [11], or are left for future investigation. However, as it will be further elaborated upon across the text, the results presented here lay the groundwork for various theoretical applications – even for scenarios involving post-selected systems.

To set the framework, we start with a brief overview of classical logical entropy. For that, consider a finite set U and a partition π = {Bi} of U, where each block Bi is a disjoint part of U. Moreover, denote by dit(π) the distinction of the partition π, i.e., the set of all pairs (u, u′) ∈ U × U such that u and u′ are not in the same block Bi of the partition π, as illustrated in Figure 1b.

The logical entropy Lπ (U) is defined as L π ( U ) = | dit ( π ) | | U × U | , $$ {L}_{\pi }(U)=\frac{|\mathrm{dit}(\pi )|}{|U\times U|}, $$(1)where | | $ |\cdot |$ denotes the cardinality of the set. In other words, Lπ (U) is the probability of getting elements from two different blocks Bi after uniformly drawing two elements from U. Therefore, it is a measure of average distinction.

If each element of U is assumed to be equally probable, we can write p B i = | B i | / | U | $ {p}_{{B}_i}=|{B}_i|/|U|$ and, thus, L π ( U ) = 1 - B i π p B i 2 . $$ {L}_{\pi }(U)=1-\sum_{{B}_i\in \pi } {p}_{{B}_i}^2. $$(2)

Furthermore, if a probability pk is assigned to each element uk ∈ U, the above expression can be applied to the partition π = 1U with element-blocks given by {ui}. This leads to L ( { p i } ) L 1 U ( U ) = 1 - i p i 2 = i p i ( 1 - p i ) . $$ L(\{{p}_i\})\equiv {L}_{{1}_U}(U)=1-\sum_i {p}_i^2=\sum_i {p}_i(1-{p}_i). $$(3)

Therefore, L({pi}) is the probability to draw two different elements ui consecutively. To generalize it even further, one may consider an arbitrary set U with a countable number of blocks. In this case, the idea of logical entropy is extended to countable probability distributions.

Starting from this definition, other concepts, like logical divergence, conditional logical entropy, and mutual logical information [8], can also be introduced along the same lines.

It should be noted that the formula for logical entropy is rooted in the history of information theory, preceding by at least a century its relation to the logic of partitions introduced by Ellerman. In fact, it can be traced back to Gini’s index of mutability [12]. Also, Polish Enigma crypto-analysts and, later, Turing used the term “repeat rate” [1315], which is just the complement of the logical entropy. Moreover, it coincides with the Tsallis entropy of index 2 and, as a result, resembles the information measure suggested by Brukner and Zeilinger [1, 2] and Manfredi [7] for applications in quantum mechanics.

Here, we follow standard methods from quantum information (e.g., [16]) to extend the notion of logical entropy to quantum states. In this approach, the set U becomes the state of a quantum system, i.e., an element of a Hilbert space. With that, besides adding new results, we generalize the ones originally presented in [17], supporting them with formal proofs regarding quantum density matrices. We hope that the discussion we present will shed new light on this intriguing informational measure.

Since the writing of the original manuscript [10], additional works in this area have elaborated on the topic [1823]. Still, we expect that the updates added here will provide new insights into the subject. In fact, within the next section, we present a new road to the introduction of quantum logical entropy that makes its connection with the idea of partitions present in the classical definition more evident. After that, in the section with the properties of quantum logical entropies, we prove some results that were not present in the original manuscript and simplify the proofs to others that were already on it whenever possible. Moreover, we added a section where new definitions of logical entropy for post-selected systems are introduced. To conclude this work, we further discussions some aspects of our study in the final section.

2 Quantum logical entropy

Every projection-valued measure (PVM) defines a partition π in a Hilbert space in the sense that it provides a direct-sum decomposition of the space [24, 25]. In fact, recall that a PVM is characterized by parts Bi = |bi〉〈bi| such that i B i = I $ {\sum }_i {B}_i=I$. Then, the probability that two consecutive measurements of copies of a state ρ in such a Hilbert space will belong to distinct parts is L π ( ρ ) i tr ( B i ρ ) [ 1 - tr ( B i ρ ) ] . $$ {L}_{\pi }(\rho )\equiv \sum_i \mathrm{tr}({B}_i\rho )[1-\mathrm{tr}({B}_i\rho )]. $$(4)

Observing that tr(Bi ρBi) = tr(Bi ρ) and Bi ρBi = tr(Bi ρ)Bi, which implies that [tr(Bi ρBi)]2 = tr(Bi ρBi)2, the quantity Lπ(ρ) can be rewritten as L π ( ρ ) = i tr ( B i ρ B i ) - i [ tr ( B i ρ B i ) ] 2 = tr ( i B i ρ B i ) - tr [ i ( B i ρ B i ) 2 ] = tr ρ - tr ( ρ ) 2 = tr [ ρ ( I - ρ ) ] , $$ \begin{array}{ll}{L}_{\pi }(\rho )& =\sum_i \mathrm{tr}({B}_i\rho {B}_i)-\sum_i [\mathrm{tr}({B}_i\rho {B}_i){]}^2\\ & =\mathrm{tr}\left(\sum_i {B}_i\rho {B}_i\right)-\mathrm{tr}\left[\sum_i ({B}_i\rho {B}_i{)}^2\right]\\ & =\mathrm{tr}\rho \mathrm{\prime}-\mathrm{tr}(\rho \mathrm{\prime}{)}^2\\ & =\mathrm{tr}[\rho \mathrm{\prime}(I-\rho \mathrm{\prime})],\end{array} $$(5)where ρ′ = ∑i(Bi ρBi) is the measured density matrix ρ with the PVM associated with π. The operator ρ′ is sometimes referred to as Lüders mixture (see, e.g., [26]) since it was introduced by Lüders [27] as a generalization of the state update rule proposed by von Neumann in his seminal book [28].

Expression (4) already defines a logical entropy for the state ρ for a given partition π (characterized by a PVM), as represented in Figure 2. Observe that this definition can be applied even in cases of coarse PVMs, i.e., PVMs with degeneracy. In this sense, and perhaps somewhat differently from the von Neumann entropy, the logical entropy has a clear operational meaning directly associated with its definition.

thumbnail Figure 2

Representation of the probability distribution used in the definition of quantum logical entropy. Any state ρ, together with a partition characterized by a PVM, which is associated with an orthonormal basis in the corresponding Hilbert space, defines a probability distribution. This is the case even in case of coarse (i.e., degenerate) PVMs, as illustrated by the yellow partition. These distributions allow the introduction of (PVM-dependent) logical entropies for quantum systems.

The dependency on PVMs should not come as a surprise since, even classically, as already seen, the logical entropy, in general, depends on the partition. However, if we wish to have a PVM-independent quantum logical entropy, we can define L ( ρ ) min π { L π ( ρ ) | π   is   non - degenerate } . $$ L\left(\rho \right)\equiv \underset{\pi }{\mathrm{min}}\left\{{L}_{\pi }\left(\rho \right)\right|{\pi }\enspace \mathrm{is}\enspace \mathrm{non}-\mathrm{degenerate}\}. $$(6)

While this is not the only possible definition, it coincides with the one that is typically considered in the literature, as shown next. It is, also, in a certain sense, a “natural” choice. In fact, it turns out that the PVM that minimizes Lπ (ρ) is the one associated with a basis for which ρ is diagonal, in which case ρ = ρ′. To see that, let ρ = ∑ij ρij|i〉〈j| be a representation of ρ in the basis associated with the projectors Bi. In this same basis, ρ′ = ∑i ρii|i〉〈i|. Moreover, tr ρ 2 = ij | ρ ij | 2 = tr ρ 2 + i j i | ρ ij | 2 , $$ {\mathrm{tr}{\rho }}^2=\sum_{{ij}} |{\rho }_{{ij}}{|}^2=\mathrm{tr}\rho {\prime}^2+\sum_i \sum_{j\ne i} |{\rho }_{{ij}}{|}^2, $$(7)which implies that Lπ(ρ) = 1 − trρ 2 + ∑iji|ρij|2, which is minimized whenever ρ is diagonal in the basis associated with the projectors Bi. Since this is the case for every non-degenerate PVM, we conclude that L ( ρ ) = tr [ ρ ( I - ρ ) ] . $$ L(\rho )=\mathrm{tr}[\rho (I-\rho )]. $$(8)

This is the expression that was defined as the quantum logical entropy in the original version of this work [10] and has, ever since, been used in the literature (see, e.g., [23]). In this update of the work, however, we introduce the quantum logical entropy through definition (4). In this way, we keep the conceptual connection to the idea of partitions present in the original (classical) formulation of the logical entropy.

Observe that equation (7) can be rewritten as L ( ρ ) - L ( ρ ) = i j i | ρ ij | 2 . $$ L(\rho \mathrm{\prime})-L(\rho )=\sum_i \sum_{j\ne i} |{\rho }_{{ij}}{|}^2. $$(9)

This means that the change in logical entropy due to a PVM is associated with the coherence of the initial state in the basis characterized by the PVM. In particular, PVMs do not decrease the logical entropy, i.e., L ( ρ ) L ( ρ ) $ L(\rho )\le L(\rho \mathrm{\prime})$.

It should be also noted that quantum logical entropy can be reduced to the complement of the purity of the state ρ, which is given by γ(ρ) = trρ 2. In fact, the expression for this entropy coincides with what is sometimes referred to as impurity or mixedness [29]. Furthermore, similarly to the classical case, the logical entropy is also equivalent to the quantum Tsallis entropy of index two, which, in turn, is equivalent to the so-called linear entropy – although its most appropriate name should be quadratic entropy, as it was referred to in [30]. While this seems to weaken the novelty in introducing the quantum logical entropy, observe that the connection with partitions, which is the starting point for that, may bring new insights and understanding to those already known quantities. Moreover, other definitions, like logical divergence and conditional logical entropy presented next, are added on top of it. They have the potential to help in the study of various scenarios in quantum information.

With that, we introduce the quantum logical divergence as d ( ρ | | σ ) 2   tr ρ ( I - σ ) - L ( ρ ) - L ( σ ) . $$ d(\rho ||\sigma )\equiv 2\enspace \mathrm{tr}\rho (I-\sigma )-L(\rho )-L(\sigma ). $$(10)

Note that there exists a relation between the logical divergence and quantum fidelity, defined as F ( ρ , σ ) = ( tr σ ρ σ ) 2 . $$ F(\rho,\sigma )={\left(\mathrm{tr}\sqrt{\sqrt{\sigma }\rho \sqrt{\sigma }}\right)}^2. $$(11)

In fact, definition (10) can be rewritten as d ( ρ | | σ ) = γ ( ρ ) + γ ( σ ) - 2 tr ( ρ σ ) . $$ d(\rho ||\sigma )=\gamma (\rho )+\gamma (\sigma )-2\mathrm{tr}({\rho \sigma }). $$(12)

While the term tr(ρσ) is not equivalent to F(ρ,σ) in general, there exists an apparent similarity between them. This becomes more evident in cases where ρ and σ commute, for which F ( ρ , σ ) = ( tr ρ σ ) 2 $ F(\rho,\sigma )={\left(\mathrm{tr}\sqrt{{\rho \sigma }}\right)}^2$. Moreover, the two are equivalent if ρ and σ are pure states.

It should be noted that the logical divergence between a density matrix before and after a PVM measurement is simply d ( ρ | | ρ ) = L ( ρ ) - L ( ρ ) . $$ d(\rho ||\rho \mathrm{\prime})=L(\rho \mathrm{\prime})-L(\rho ). $$(13)

Finally, we can also introduce the quantum conditional logical entropy as L ( A / B ) L ( ρ AB ) - L ( I / d ρ B ) , $$ L(A/B)\equiv L({\rho }_{{AB}})-L(I/d\otimes {\rho }_B), $$(14)where d is the dimension of A. Moreover, here and throughout the text, we denote ρA = trB ρAB and ΡB = trA ρAB.

Observe that the conditional logical entropy can be reduced to a special case of the logical divergence. In fact, since tr [ ρ AB ( I - I / d ρ B ) ] = 1 - 1 d tr ρ B 2 , $$ \mathrm{tr}[{\rho }_{{AB}}(I-I/d\otimes {\rho }_B)]=1-\frac{1}{d}\mathrm{tr}{\rho }_B^2, $$(15)and L ( I / d ρ B ) = 1 - 1 d tr ρ B 2 , $$ L(I/d\otimes {\rho }_B)=1-\frac{1}{d}\mathrm{tr}{\rho }_B^2, $$(16)it follows from the definition of both entropies that L ( A / B ) = - d ( ρ AB | | I / d ρ B ) , $$ L(A/B)=-d({\rho }_{{AB}}||I/d\otimes {\rho }_B), $$(17)proving the relation between conditional logical entropy and logical divergence. As a result, they share multiple properties.

Observe that, different from other logical entropies, the conditional logical entropy is negative. However, this does not affect its close relation to probabilities since the function is always non-positive. Hence, the overall negative sign can be mitigated. In fact, one could, for instance, replace the original definition by its additive inverse. This is not done here to maintain a similar algebraic structure between the definition of conditional von Neumman entropy and conditional logical entropy.

In the following section, we state and prove various properties of quantum logical entropies.

3 Properties of quantum logical entropies

Proposition 3.1 (Basic properties).

  1. Logical entropy is non-negative and L(ρ) = 0 for a pure state.

  2. The maximal value of the logical entropy is 1 −1/d, where d is the dimension of the Hilbert space. This value is the logical entropy of the maximally mixed state I/d.

  3. Given a composite pure state ρAB on the space (A, B), it follows that L ( ρ A ) = L ( ρ B ) $ L({\rho }_A)=L({\rho }_B)$.

  4. If ρ AB = ρ A ρ B $ {\rho }_{{AB}}={\rho }_A\otimes {\rho }_B$ , then

L ( ρ A ρ B ) = L ( ρ A ) + L ( ρ B ) - L ( ρ A ) L ( ρ B ) . $$ L({\rho }_A\otimes {\rho }_B)=L({\rho }_A)+L({\rho }_B)-L({\rho }_A)\cdot L({\rho }_B). $$(18)

Proof.

  1. This follows from the definition since the logical entropy is a probability. However, another way to verify it is by observing that for every density matrix, trρ = 1 and trρ 2 ≤ 1, with equality if and only if ρ is pure.

  2. Let {λi} be the set of eigenvalues of a density matrix ρ. Then, using the Cauchy–Schwarz inequality for two vectors u and v whose components are, respectively, ui = λi and vi = 1/d, it holds that trρ 2 ≥ 1/d. Therefore, L(ρ) ≤ 1 − 1/d. Note also that L(I/d) = 1 − 1/d.

  3. The result follows immediately from the Schmidt decomposition since A and B have the same orthonormal set of eigenvectors.

  4. The result follows from writing ρA and ρB in their diagonal form and, then, using the identity

( 1 - x ) + ( 1 - y ) - ( 1 - x ) ( 1 - y ) = 1 - xy . $$ (1-x)+(1-y)-(1-x)(1-y)=1-{xy}. $$(19)

Observe that part (d) of the previous proposition states, in particular, that the logical entropy of separable bipartite systems is subadditive, i.e., L ( ρ AB ) L ( ρ A ) + L ( ρ B ) . $$ L({\rho }_{{AB}})\le L({\rho }_A)+L({\rho }_B). $$(20)

The next proposition generalizes this result for generic bipartite systems.

Proposition 3.2 (Subadditivity).

The logical entropy of a density matrix ρAB is subadditive.

Proof. This result was proved in 2007 by Audenaert for any quantum Tsallis entropy of index greater than one [31], which, in particular, includes the logical entropy.□

Although the quantum logical entropy satisfies the subadditivity property, it should be noted that, differently from the von Neumann entropy, it does not satisfy the strong subadditivity, i.e., in general, it does not hold that L ( ρ ABC ) + L ( ρ B ) L ( ρ AB ) + L ( ρ BC ) . $$ L({\rho }_{{ABC}})+L({\rho }_B)\le L({\rho }_{{AB}})+L({\rho }_{{BC}}). $$(21)

However, the logical entropy satisfies a condition called firm subadditivity, which is characterized as follows: Given a PVM defined by blocks Ak in a subsystem A of a bipartite state ρAB, the logical entropy is said to be firm subadditive if L ( ρ AB ) L ( ρ A ) + k p k L ( ρ B ( k ) ) , $$ L({\rho }_{{AB}})\le L({\rho }_A)+\sum_k {p}_kL({\rho }_B^{(k)}), $$(22)where p k = tr ( A k ρ A ) $ {p}_k=\mathrm{tr}({A}_k{\rho }_A)$ and p k ρ B ( k ) = t r A ( A k ρ AB ) $ {p}_k{\rho }_B^{(k)}=\mathrm{t}{\mathrm{r}}_A({A}_k{\rho }_{{AB}})$.

Proposition 3.3 (Firm subadditivity).

The logical entropy of a density matrix ρ AB $ {\rho }_{{AB}}$ is firm subadditive.

Proof. This result was proved in 2011 by Coles in Theorem 5 of [30]. □

In our next result, we show that the logical entropy satisfies a triangle inequality.

Proposition 3.4 (Triangle inequality).

For any density matrix ρAB , it holds that | L ( ρ A ) - L ( ρ B ) | L ( ρ AB ) . $$ |L({\rho }_A)-L({\rho }_B)|\le L({\rho }_{{AB}}). $$(23)

Proof. Let R be a system such that ρABR is pure. From Proposition 3.2, we deduce that L ( ρ BR ) L ( ρ B ) + L ( ρ R ) . $$ L({\rho }_{{BR}})\le L({\rho }_B)+L({\rho }_R). $$(24)

Moreover, Proposition 3.1(c) implies that L(ρR) = L(ρAB) and L(ρA) = L(ρBR). As a result, L ( ρ A ) - L ( ρ B ) L ( ρ AB ) . $$ L({\rho }_A)-L({\rho }_B)\le L({\rho }_{{AB}}). $$(25)

A similar reasoning leads to L ( ρ B ) - L ( ρ A ) L ( ρ AB ) . $$ L({\rho }_B)-L({\rho }_A)\le L({\rho }_{{AB}}). $$(26)

Therefore, inequality (23) holds. □

As discussed earlier, by definition, a PVM cannot decrease the logical entropy of a system. Our next result shows that this is the case not only for PVMs but for any unital map ΛU, which is a map that preserves the maximally mixed state, i.e., ΛU (I/d) = I/d, where d is the dimension of the Hilbert space. Note that for every positive operator-valued measure (POVM), which is given by a set of semi-definite matrices { E i } i = 1 n $ {\left\{{E}_i\right\}}_{i=1}^n$ with i = 1 n E i = I $ {\sum }_{i=1}^n {E}_i=I$, there are many corresponding implementations Mi such that E i = M i M i $ {E}_i={M}_i^{\dagger }{M}_i$. Ei are positive semidefinite and so Mi = Pi Ui where Pi is a unique Hermitian positive semidefinite (also denoted P i = E i $ {P}_i=\sqrt{{E}_i}$) and Ui can be any unitary. Choosing Ui = I, we obtain a unital implementation Λ U ( ρ ) = i = 1 n M i ρ M i $ {\mathrm{\Lambda }}_U\left(\rho \right)={\sum }_{i=1}^n {M}_i\rho {M}_i^{\dagger }$ of the POVM.

Proposition 3.5 (Entropy after a unital map).

For any unital map Λ U $ {\Lambda }_U$ , it holds that L ( ρ ) L ( ρ ) , $$ L\left(\rho \right)\le L\left(\rho \mathrm{\prime}\right), $$(27) where ρ′ = ΛU(ρ).

Proof. Recall that given two vectors x and y of dimension n, y is said to majorize x, which is denoted by x ≺ y, if i = 1 k x i i = 1 k y i $ {\sum }_{i=1}^k {x}_i^{\downarrow }\le {\sum }_{i=1}^k {y}_i^{\downarrow }$ for every k ≤ n and i = 1 n x i = i = 1 n y i $ {\sum }_{i=1}^n {x}_i={\sum }_{i=1}^n {y}_i$, where x $ {x}^{\downarrow }$ is the vector x reordered so that its components are in non-increasing order. Then, as stated in Lemma 3 of [32] and proven in [33, 34], given two density matrices ρ and σ of equal dimension, ρ can be converted into σ via some unital map Λ if and only if the vector of eigenvalues of ρ majorizes the vector of eigenvalues of σ. In our case of interest, if x is the vector of eigenvalues of ρ′ and y is the vector of eigenvalues of ρ, x ≺ y.

Moreover, as shown in Proposition 12.11 of [16], x ≺ y if and only if x = j = 1 m p j P j y $ x={\sum }_{j=1}^m {p}_j{P}_jy$ for some probability distribution pj and permutation matrices Pj. Now, letting Pj be the permutation matrix associated with the permutation σj, we can write x i = j = 1 m p j y σ j ( i ) $ {x}_i={\sum }_{j=1}^m {p}_j{y}_{{\sigma }_j(i)}$. As a result, x 2 = i = 1 d x i 2 $ {x}^2={\sum }_{i=1}^d {x}_i^2$ gives x 2 = i = 1 d ( j = 1 m p j y σ j ( i ) ) 2 = i = 1 d j = 1 m p j 2 y i 2 + i = 1 d j = 1 m k j p j p k y σ j ( i ) y σ k ( i ) . $$ \begin{array}{ll}{x}^2& =\sum_{i=1}^d {\left(\sum_{j=1}^m {p}_j{y}_{{\sigma }_j(i)}\right)}^2\\ & =\sum_{i=1}^d \sum_{j=1}^m {p}_j^2{y}_i^2+\sum_{i=1}^d \sum_{j=1}^m \sum_{k\ne j} {p}_j{p}_k{y}_{{\sigma }_j(i)}{y}_{{\sigma }_k(i)}.\end{array} $$(28)

Furthermore, since j = 1 m p j = 1 $ {\sum }_{j=1}^m {p}_j=1$, y 2 = i = 1 d ( j = 1 m p j ) 2 y i 2 = i = 1 d j = 1 m p j 2 y i 2 + i = 1 d j = 1 m k j p j p k y i 2 . $$ \begin{array}{ll}{y}^2& =\sum_{i=1}^d {\left(\sum_{j=1}^m {p}_j\right)}^2{y}_i^2\\ & =\sum_{i=1}^d \sum_{j=1}^m {p}_j^2{y}_i^2+\sum_{i=1}^d \sum_{j=1}^m \sum_{k\ne j} {p}_j{p}_k{y}_i^2.\end{array} $$(29)

Therefore, y 2 - x 2 = i = 1 d j = 1 m k j p j p k ( y i 2 - y σ j ( i ) y σ k ( i ) ) = i = 1 d j = 1 m k j p j p k ( 1 2 y σ j ( i ) 2 + 1 2 y σ k ( i ) 2 - y σ j ( i ) y σ k ( i ) ) = 1 2 i = 1 d j = 1 m k j p j p k ( y σ j ( i ) - y σ k ( i ) ) 2 0 . $$ \begin{array}{ll}{y}^2-{x}^2& =\sum_{i=1}^d \sum_{j=1}^m \sum_{k\ne j} {p}_j{p}_k\left({y}_i^2-{y}_{{\sigma }_j(i)}{y}_{{\sigma }_k(i)}\right)\\ & =\sum_{i=1}^d \sum_{j=1}^m \sum_{k\ne j} {p}_j{p}_k\left(\frac{1}{2}{y}_{{\sigma }_j(i)}^2+\frac{1}{2}{y}_{{\sigma }_k(i)}^2-{y}_{{\sigma }_j(i)}{y}_{{\sigma }_k(i)}\right)\\ & =\frac{1}{2}\sum_{i=1}^d \sum_{j=1}^m \sum_{k\ne j} {p}_j{p}_k{\left({y}_{{\sigma }_j(i)}-{y}_{{\sigma }_k(i)}\right)}^2\\ & \ge 0.\end{array} $$(30)

Since x and y are, respectively, the vectors of the eigenvalues of ρ′ and ρ, tr ( ρ ) 2 tr ρ 2 , $$ {\mathrm{tr}\left(\rho \mathrm{\prime}\right)}^2\le \mathrm{tr}{\rho }^2, $$(31)which implies that inequality (27) holds. □

Suppose a system of interest interacts unitarily with other systems whose individual states are not under experimental control, e.g., the environment. Such interactions can be effectively represented by unital maps acting on the system of interest, i.e., a system of interest that starts in the state ρS can be repersented by ρ S = i E i ρ S E i $ {\rho \mathrm{\prime}}_S={\sum }_i {E}_i{\rho }_S{E}_i^{\dagger }$ after the interaction. Then, from the previous proposition, the system of interest’s logical entropy increases. In particular, the lower bound for the logical entropy should no longer be null. This lower bound as well as an upper bound in case the joint system is pure is the concern of the next result.

Proposition 3.6 (Entropy after unitary interaction).

Assume a system S interacts with a system R through a unitary U. Also, let the joint system after the interaction be ρ ' SR = U ρ SR U $ {{\rho \prime}}_{{SR}}=U{\rho }_{{SR}}{U}^{\dagger }$ and define B ij = i | ρ ' SR | j $ {{B}_{{ij}}}=\left\langle i|{\rho \prime}_{{SR}}|j\right\rangle $ , where ρSR is the state of the joint system before the interaction and {|i〉} is an orthonormal basis of system R. Then, L ( ρ S ) 2 i j < i { tr ( B ij B ij ) - Re [ tr ( B ii B jj ) ] } , $$ {L(\rho \mathrm{\prime}}_S)\ge 2\sum_i \sum_{j < i} \left\{{tr}({B}_{{ij}}{B}_{{ij}}^{\dagger })-{Re}[{tr}({B}_{{ii}}{B}_{{jj}})]\right\}, $$(32) where ρ ' S = t r R ρ ' SR $ {{{\rho \prime}}}_S=t{r}_R{\rho \prime}_{{SR}}$ . Furthermore, if ρSR is a pure state, L ( ρ S ) 2 i j < i tr ( B ij B ij ) . $$ {L(\rho \mathrm{\prime}}_S)\le 2\sum_i \sum_{j < i} {tr}({B}_{{ij}}{B}_{{ij}}^{\dagger }). $$(33)

Proof. Observe that ρ ' S = i B ii $ {{\rho \prime}}_S={\sum }_i {B}_{{ii}}$ and hence ρ S 2 = ij B ii B jj $ {{\rho }_S^\mathrm{\prime}}^2={\sum }_{{ij}} {B}_{{ii}}{B}_{{jj}}$. As a result, L ( ρ ' S ) = 1 - i tr ( B ii 2 ) - i j i tr ( B ii B jj ) = 1 - i tr ( B ii 2 ) - 2 i j < i Re [ tr ( B ii B jj ) ] . $$ \begin{array}{ll}L({{\rho \prime}}_S)& =1-\sum_i \mathrm{tr}({B}_{{ii}}^2)-\sum_i \sum_{j\ne i} \mathrm{tr}({B}_{{ii}}{B}_{{jj}})\\ & =1-\sum_i \mathrm{tr}({B}_{{ii}}^2)-2\sum_i \sum_{j < i} \mathrm{Re}[\mathrm{tr}({B}_{{ii}}{B}_{{jj}})].\end{array} $$(34)

Moreover, ρ SR = ij B ij | i j | $ \rho \mathrm{\prime}_{{SR}}={\sum }_{{ij}} {B}_{{ij}}\otimes \left.|i\right\rangle\left\langle\right. j| $ and ρ SR 2 = ijk B ik B kj | i j | $ {{\rho \mathrm{\prime}}_{{SR}}}^2={\sum }_{{ijk}} {B}_{{ik}}{B}_{{kj}}\otimes |\left.i\right\rangle\left\langle j\right.| $. Then, since tr ρ SR 2 1 $ {{\mathrm{tr}\rho \mathrm{\prime}}}_{{SR}}^2\le 1 $, 1 - i tr ( B ii 2 ) i j i tr ( B ij B ji ) . $$ 1-\sum_i \mathrm{tr}({B}_{{ii}}^2)\ge \sum_i \sum_{j\ne i} \mathrm{tr}({B}_{{ij}}{B}_{{ji}}). $$(35)

Finally, combining expressions (34) and (35), we obtain inequality (32), which completes the first part of the proof.

Now, if ρSR is a pure state, which implies that ρSR is also pure, tr ρ SR 2 = 1 $ {{\mathrm{tr}\rho \mathrm{\prime}}}_{{SR}}^2=1$ and, as a result, expression (35) becomes an equality. The left-hand side of the latter is greater or equals to L(ρS), as can be directly seen from (34). Therefore, we are lead to inequality (33) holds, completing the proof. □

Proposition 3.7 (Entropy of classical mixtures of quantum systems).

Let ρ = i p i ρ i $ \rho ={\sum }_i {p}_i{\rho }_i$, then L ( ρ ) L ( { p i } ) + k p k 2 L ( ρ k ) , $$ L(\rho )\le L(\{{p}_i\})+\sum_k {p}_k^2L({\rho }_k), $$(36)where L({pi}) is defined in (3). Moreover, the equality holds whenever the matrices ρi have orthogonal support.

Proof. To start, assume the states ρi are pure, i.e., ρi = |ψi〉〈ψi|. Then, ρ = ∑i pi|ψi〉〈ψi|. Introducing an auxiliary system R that purifies ρ, the state of the joint system can be written as | Ψ SR   = i p i   | ψ i   | i , $ |{\mathrm{\Psi }}_{{SR}}\left.\right\rangle\enspace ={\sum }_i \sqrt{{p}_i}\enspace\left. |{\psi }_i\right\rangle\enspace |i\rangle, $ where {|i〉} is an orthonormal basis of system R. Observe that ρ = ρS and ρ R = ij l (     l |   ψ i     ψ j |   l )   | i     j |   p i p j , $$ {\rho }_R=\sum_{{ij}} \sum_{\mathcal{l}} \left(\enspace \left\langle \enspace \mathcal{l}|\enspace {\psi }_i\right\rangle\enspace \left\langle \enspace {\psi }_j|\enspace \mathcal{l}\right\rangle\right)\enspace\left. |i\right\rangle\enspace \left\langle \right.\enspace j|\enspace \sqrt{{p}_i}\sqrt{{p}_j}, $$(37)where {   | l   } $ \{\enspace\left. |\mathcal{l}\right\rangle\enspace \} $ is an orthonormal basis of S. Also, by Proposition 3.1(c), L ( ρ S ) = L ( ρ R ) $ L({\rho }_S)=L({\rho }_R)$.

If ρR is measured with a PVM characterized by Pi = |i〉〈i| , $,$ it holds that ρ R = i P i ρ R P i = p i   | i     i |   . $$ {\rho \mathrm{\prime}}_R=\sum_i {P}_i{\rho }_R{P}_i=\sum {p}_i\enspace |\left.i\right\rangle\enspace \left\langle \enspace i|\right.\enspace. $$(38)

This implies that L(ρR) = L({pi}). Since PVMs increase the logical entropy, L ( { p i } ) = L ( ρ R ) L ( ρ R ) = L ( ρ S ) . $$ {L(\{{p}_i\})}=L(\rho \mathrm{\prime}_R)\ge L({\rho }_R)=L({\rho }_S). $$(39)

Since the logical entropy of pure states vanish, the above expression proves the proposition for cases where the matrices ρi are pure states.

For the general case, let ρi = ∑ij λij|λij〉〈λij| , $,$ where the states |λij are orthonormal vectors in the subspace associated with ρi. Hence, ρ can be written as the sum of pure states ρ = ∑ij pi λij|λij〉〈λij| (not necessarily all orthogonal). Now, using the result for pure states we just obtained, it holds that L ( ρ ) L ( { p i λ ij } ) = 1 - ij p i 2 λ ij 2 = 1 - i p i 2 + k p k 2 ( 1 - j λ kj 2 ) = L ( { p i } ) + k p k 2 L ( ρ k ) , $$ \begin{array}{ll}L(\rho )& \le L(\{{p}_i{\lambda }_{{ij}}\})\\ & =1-\sum_{{ij}} {p}_i^2{\lambda }_{{ij}}^2\\ & =1-\sum_i {p}_i^2+\sum_k {p}_k^2(1-\sum_j {\lambda }_{{kj}}^2)\\ & =L(\{{p}_i\})+\sum_k {p}_k^2L({\rho }_k),\end{array} $$(40)as we wanted to show.

Finally, it follows from direct computation that the equality holds whenever the matrices ρi have orthogonal support. In fact, in this case, ρ 2 = i p i 2 ρ i 2 $ {\rho }^2={\sum }_i {p}_i^2{\rho }_i^2$ and, as a result, L ( ρ ) = 1 - i p i 2 tr ρ i 2 = 1 - i p i 2 + k p k 2 ( 1 - tr ρ k 2 ) , $$ \begin{array}{ll}L(\rho )& =1-\sum_i {p}_i^2\mathrm{tr}{\rho }_i^2\\ & =1-\sum_i {p}_i^2+\sum_k {p}_k^2(1-\mathrm{tr}{\rho }_k^2),\end{array} $$(41)which concludes the proof. □

In the next result, we prove the non-negativity of the logical divergence.

Proposition 3.8 (Klein’s inequality).

The logical divergence is always non-negative, i.e., d ( ρ | | σ ) 0 $$ d(\rho ||\sigma )\ge 0 $$(42) for every pair of density matrices ρ and σ, with equality holding if and only if ρ = σ.

Proof. It follows by direct computation that d ( ρ | | σ ) = tr ( ρ - σ ) 2 . $$ d(\rho ||\sigma )=\mathrm{tr}(\rho -\sigma {)}^2. $$(43)

Then, recall that for any Hermitian matrix A, tr(A2) ≥ 0, with tr(A 2) = 0 if and only if A = 0. □

It should be noted that (43) defines the logical divergence as the square of the Hilbert–Schmidt norm of the difference of the two density matrices under consideration. Moreover, it can be shown that this norm coincides with a natural definition of the Hamming distance between density matrices [23].

In the next three propositions, we study the concavity of logical entropies.

Proposition 3.9 (Concavity of logical entropy).

Let ρ = ∑pi pi , where ρi is a density matrix for each i and 0 ≤ pi ≤ 1, such thatpi = 1. Moreover, set L ( ρ ) ̅ = p i L ( ρ i ) $ \overline{L(\rho )}=\sum {p}_iL({\rho }_i)$.

  1. If ρi have orthogonal support, then

    L ( ρ ) ̅ < L ( ρ ) . $$ \overline{L(\rho )} < L(\rho ). $$(44)

  2. In general,

    L ( ρ ) ̅ - L ( { p i } ) < L ( ρ ) < L ( ρ ) ̅ + L ( { p i } ) , $$ \overline{L(\rho )}-L(\{{p}_i\})<L(\rho ) < \overline{L(\rho )}+L(\{{p}_i\}), $$(45)

where L({pi}) is the classical logical entropy of the distribution {pi}i. In other words, L(ρ) is in the L({pi}) neighborhood of L ( ρ ) ̅ $ \overline{L(\rho )}$.

Proof.

  1. We will demonstrate the argument on two density matrices ρ1 and ρ2 having an orthogonal support. Set ρ1 = ∑i pi|i〉〈i| , $,$ ρ 2 = j q j   | j     j | , $ {\rho }_2={\sum }_j{q}_j\enspace \left.|j\right\rangle\enspace \left\langle \enspace {j}|\right., $ where |i   $ \enspace $ and |j〉 are two bases with orthogonal support, also set ρ = λρ1 + (1 − λ)ρ2 , where 0 < λ < 1. Now, L ( ρ ) ̅ = λ L ( ρ 1 ) + ( 1 - λ ) L ( ρ 2 ) = λ ( 1 - i p i 2 ) + ( 1 - λ ) ( 1 - j q j 2 ) = 1 - λ i p i 2 - ( 1 - λ ) j q j 2 . $$ \begin{array}{ll}\overline{L(\rho )}& =\lambda L({\rho }_1)+(1-\lambda )L({\rho }_2)\\ & =\lambda (1-\sum_i {p}_i^2)+(1-\lambda )(1-\sum_j {q}_j^2)\\ & =1-\lambda \sum_i {p}_i^2-(1-\lambda )\sum_j {q}_j^2.\end{array} $$(46)

    However, L ( ρ ) = 1 - λ 2 i p i 2 - ( 1 - λ ) 2 j q j 2 . $$ L(\rho )=1-{\lambda }^2\sum_i {p}_i^2-(1-\lambda {)}^2\sum_j {q}_j^2. $$(47)

    Therefore, inequality (44) holds.

  2. Consider ρ AB = i p i ρ i   | i     i | , $ {\rho }_{{AB}}={\sum }_i {p}_i{\rho }_i\otimes \enspace\left. |i\right\rangle\enspace \left\langle \enspace i|\right., $ so ρAB is a sum of densities with an orthogonal support. From the result in part (a) and Proposition 3.2, it follows that

L ( ρ ) ̅ L ( ρ AB ) L ( ρ A ) + L ( ρ B ) . $$ \overline{L(\rho )}\le L({\rho }_{{AB}})\le L({\rho }_A)+L({\rho }_B). $$(48)

However, ρA = ρ and L(ρB) = L({pi}), which leads to L ( ρ ) ̅ - L ( { p i } ) L ( ρ ) , $$ \overline{L(\rho )}-L(\{{p}_i\})\le L(\rho ), $$(49)proving the first part of inequality (45). To conclude the proof, we first show that L ( ρ ) L ( ρ ) ̅ + L ( { p i } ) $ L(\rho )\le \overline{L(\rho )}+L(\{{p}_i\})$ for the case where the matrices ρi are pure states, i.e., ρi = |ψi〉〈ψi|   $ \enspace $. For that, consider the following purification of the system: | η   = i p i   | ψ i     | i , $$ \left.|\eta \right\rangle\enspace =\sum_i \sqrt{{p}_i}\enspace\left. |{\psi }_i\right\rangle\enspace \otimes \enspace |i\rangle, $$(50)where the vectors |i   $ \enspace $ are orthonormal in some system B. Moreover, denote ρ ̃ =   | η     η |   $ \tilde{\rho}=\enspace\left. |\eta \right\rangle\enspace \left\langle \enspace \eta |\right.\enspace $. It is clear, then, that ρ = ρ ̃ A = i p i   | ψ i     ψ i |   $ \rho ={\tilde{\rho}}_A={\sum }_i {p}_i\enspace \left.|{\psi }_i\right\rangle\enspace \left\langle \enspace {\psi }_i|\right.\enspace $ and ρ ̃ B = ij p i p j     ψ j |   ψ i   | i     j |   . $$ {\tilde{\rho}}_B=\sum_{{ij}} \sqrt{{p}_i{p}_j}\enspace\left. \left\langle \enspace {\psi }_j| {\psi }_i\right\rangle\enspace |i\right\rangle\enspace \left\langle \enspace j|\right.\enspace. $$(51)

Note that the vectors |ψi   $ \enspace $ are not necessarily orthogonal. Measuring system B with the operators Pi = |i〉〈i| , $,$ we obtain ρ ̃ B = i p i   | i     i |   $ {\tilde{\rho}}_{B\mathrm{\prime}}={\sum }_i {p}_i\enspace\left. |i\right\rangle\enspace \left\langle \enspace i|\right.\enspace $. By Proposition 3.5, L ( ρ ̃ B ) = L ( { p i } ) L ( ρ ̃ B ) = L ( ρ ) $ L({\tilde{\rho}}_{B\mathrm{\prime}})=L(\{{p}_i\})\ge L({\tilde{\rho}}_B)=L(\rho ) $, where we used Proposition 3.1(c) in the last step. Therefore, for ρ which is a sum of pure states, we have L(ρ) ≤ L({pi}), which is consistent with (45) since, in this case, L ( ρ ) ̅ = 0 $ \overline{L(\rho )}=0$.

Consider now the general case where ρ = ∑i pi ρi with ρ i = j p i j   | e i j     e i j |   . $$ {\rho }_i=\sum_j {p}_i^j\enspace\left. |{e}_i^j\right\rangle\enspace \left\langle \enspace {e}_i^j|\right.\enspace. $$(52)

Here, {   | e i j   } j $ \left.\{\enspace |{e}_i^j\right\rangle\enspace {\}}_j $ are orthonormal vectors for each i $ i$. Hence, ρ = ij p i p i j   | e i j     e i j | , $$ \rho =\sum_{{ij}} {p}_i{p}_i^j\enspace\left. |{e}_i^j\right\rangle\enspace \left\langle \enspace {e}_i^j|\right., $$(53)where | e i j     e i j |   $ \left.|{e}_i^j\right\rangle\enspace \left\langle \enspace {e}_i^j|\right.\enspace $ are pure states for every i and j. We can use the previous result for pure states to conclude that L ( ρ ) L ( { p i p i j } ) = ij p i p i j ( 1 - p i p i j ) . $$ L(\rho )\le L(\{{p}_i{p}_i^j\})=\sum_{{ij}} {p}_i{p}_i^j(1-{p}_i{p}_i^j). $$(54)

Using the fact that, for x i , x j [ 0,1 ] $ {x}_i,{x}_j\in [\mathrm{0,1}]$, 1 - x i x j ( 1 - x i ) + ( 1 - x j ) , $$ 1-{x}_i{x}_j\le (1-{x}_i)+(1-{x}_j), $$(55)we are led to L ( ρ ) ij p i p i j ( 1 - p i ) + ij p i p i j ( 1 - p i j ) = L ( { p i } ) + i p i L ( ρ i ) , $$ \begin{array}{ll}L(\rho )& \le \sum_{{ij}} {p}_i{p}_i^j(1-{p}_i)+\sum_{{ij}} {p}_i{p}_i^j(1-{p}_i^j)\\ & =L(\{{p}_i\})+\sum_i {p}_iL({\rho }_i),\end{array} $$(56)where we used the orthogonality of the set of vectors {   | e i j   } j $ \{\enspace \left.|{e}_i^j\right\rangle\enspace {\}}_j $ for each i $ i$. □

Proposition 3.10 (Joint convexity of logical divergence).

The logical divergence d ( ρ | | σ ) $ d(\rho ||\sigma )$ is jointly convex.

Proof. We start by defining ρ = λρ1 + (1 − λ)ρ2 and σ = λσ1 + (1 − λ)σ2. Then, it follows by direct computation that d ( ρ | | σ ) = tr ( ρ - σ ) 2 = tr [ λ ( ρ 1 - σ 1 ) + ( 1 - λ ) ( ρ 2 - σ 2 ) ] 2 λ tr ( ρ 1 - σ 1 ) 2 + ( 1 - λ ) tr ( ρ 2 - σ 2 ) 2 = λ d ( ρ 1 | | σ 1 ) + ( 1 - λ ) d ( ρ 2 | | σ 2 ) , $$ \begin{array}{ll}d(\rho ||\sigma )& =\mathrm{tr}(\rho -\sigma {)}^2\\ & =\mathrm{tr}[\lambda ({\rho }_1-{\sigma }_1)+(1-\lambda )({\rho }_2-{\sigma }_2){]}^2\\ & \le \lambda \mathrm{tr}({\rho }_1-{\sigma }_1{)}^2+(1-\lambda )\mathrm{tr}({\rho }_2-{\sigma }_2{)}^2\\ & =\lambda d({\rho }_1||{\sigma }_1)+(1-\lambda )d({\rho }_2||{\sigma }_2),\end{array} $$(57)where the inequality is due to the convexity of tr ( x 2 ) $ \mathrm{tr}({x}^2)$. □

Proposition 3.11 (Concavity of conditional entropy).

The conditional logical entropy L(A/B) is a concave function of ρAB.

Proof. It follows direct from the relation given by (17) and Proposition 3.10. □

The next proposition states the fact that the divergence behaves as a metric. Tracing out a subspace can only reduce its value.

Proposition 3.12 (Monotonicity of logical divergence).

Let ρAB and σAB be two density matrices, then d ( ρ A I / b | | σ A I / b ) d ( ρ AB | | σ AB ) , $$ d({\rho }_A\otimes I/b||{\sigma }_A\otimes I/b)\le d({\rho }_{{AB}}||{\sigma }_{{AB}}), $$(58) where b is the dimension of B.

Proof. As can be seen in Chapter 11 of [16], there exist a set of unitary matrices Uj over B and a probability distribution pj such that ρ A I b = j p j U j ρ AB U j , $$ {\rho }_A\otimes \frac{I}{b}=\sum_j {p}_j{U}_j{\rho }_{{AB}}{U}_j^{\dagger }, $$(59)for every ρAB. Then, because the logical divergence is jointly convex on both entries, we can write d ( ρ A I / b | | σ A I / b ) j p j d ( U j ρ AB U j | | U j σ AB U j ) . $$ d({\rho }_A\otimes I/b||{\sigma }_A\otimes I/b)\le \sum_j {p}_jd({U}_j{\rho }_{{AB}}{U}_j^{\dagger }||{U}_j{\sigma }_{{AB}}{U}_j^{\dagger }). $$(60)

Finally, since the divergence is invariant under unitary transformations, the above sum gives d(ρAB||σAB). □

4 Quantum logical entropy of post-selected systems

We shall now propose extensions of the notion of quantum logical entropy to the class of pre- and post-selected quantum systems. Like before, we use the fact that every PVM defines a partition π in a Hilbert space given by a basis {|bi〉} associated with it since every PVM is characterized by parts Bi = |bi〉〈bi| such that ∑i Bi = I. Suppose that a system is prepared in a state |ψ〉 and a PVM is performed. Moreover, assume that it is known that the system is later found in a state |ϕ〉 non-orthogonal to |ψ〉. In scenarios like that, with a pre-selected |ψ〉 and a post-selected |ϕ〉, the system can be represented by the generalized density matrix ρ ψ | ϕ = | ψ ϕ | ϕ | ψ . $$ {\rho }_{\psi |\phi }=\frac{\left.|\psi \right\rangle\left\langle \phi |\right.}{\left\langle \phi |\psi \right\rangle}. $$(61)

The probability of measuring a result associated with B i $ {B}_i$ can be computed as | tr ( B i ρ ψ | ϕ ) | 2 = tr ( B i ρ ψ | ϕ B i ρ ψ | ϕ ) $ |\mathrm{tr}({B}_i{\rho }_{\psi |\phi }){|}^2=\mathrm{tr}({B}_i{\rho }_{\psi |\phi }{B}_i{\rho }_{\psi |\phi }^{\dagger })$. This result is known as the ABL rule [35]. Then, in analogy with the definition of logical entropy for classical and non-post-selected quantum systems, we may define L π ( ρ ψ | ϕ ) i | tr ( B i ρ ψ | ϕ ) | 2 | 1 - tr ( B i ρ ψ | ϕ ) | 2 . $$ {L}_{\pi }({\rho }_{\psi |\phi })\equiv \sum_i |\mathrm{tr}({B}_i{\rho }_{\psi |\phi }){|}^2|1-\mathrm{tr}({B}_i{\rho }_{\psi |\phi }){|}^2. $$(62)

Since | 1 - tr ( B i ρ ψ | ϕ ) | 2 = 1 - tr ( B i ρ ψ | ϕ B i ) - tr ( B i ρ ψ | ϕ B i ) + tr ( B i ρ ψ | ϕ B i ρ ψ | ϕ B i ) $ |1-\mathrm{tr}({B}_i{\rho }_{\psi |\phi }){|}^2=1-\mathrm{tr}({B}_i{\rho }_{\psi |\phi }{B}_i)-\mathrm{tr}({B}_i{\rho }_{\psi |\phi }^{\dagger }{B}_i)+\mathrm{tr}({B}_i{\rho }_{\psi |\phi }{B}_i{\rho }_{\psi |\phi }^{\dagger }{B}_i)$, it holds that L π ( ρ ψ | ϕ ) = i tr | B i ρ ψ | ϕ B i ( I - B i ρ ψ | ϕ B i ) | 2 = | tr [ ρ ψ | ϕ ( I - ρ ψ | ϕ ) ] | 2 , $$ \begin{array}{ll}{L}_{\pi }({\rho }_{\psi |\phi })& =\sum_i \mathrm{tr}|{B}_i{\rho }_{\psi |\phi }{B}_i(I-{B}_i{\rho }_{\psi |\phi }{B}_i){|}^2\\ & =|\mathrm{tr}[{\rho {\prime}}_{\psi |\phi }(I-{\rho {\prime}}_{\psi |\phi })]{|}^2,\end{array} $$(63)where ρ ψ | ϕ = i B i ρ ψ | ϕ B i $ {\rho {\prime}_{\psi |\phi }}={\sum }_i {B}_i{\rho }_{\psi |\phi }{B}_i $.

Differently from the case without post-selection, where a PVM-independent logical entropy could be defined by minimizing over non-degenerate PVMs, this is not possible here. The reason is that we are interested in cases where the pre- and post-selections are pure states. Then, although the minimum Lπψ|ϕ) corresponds to the case where tr ( ρ ψ | ϕ ) = tr ( ρ ψ | ϕ ) $ {\mathrm{tr}(\rho \mathrm{\prime}}_{\psi |\phi })=\mathrm{tr}({\rho }_{\psi |\phi }) $, it also corresponds to the null function. This is the case because in such scenarios the intermediate PVM is associated with an orthonormal basis that contains either |ψ〉 or |ϕ〉. As a result, these PVMs allow us to see ρψ|ϕ as a partition with a single part.

Observe that, by construction, the logical entropy for post-selected systems is always positive. This fact contrasts with a generalization of the von Neumann entropy proposed in [36], which can assume negative values for some post-selections.

Another approach for this problem employs directly the notion of weak values [37], which has been shown to be very constructive in the study of pre- and post-selected systems [3850]. This approach consists of replacing the intermediate PVM by the inference of the weak values of orthogonal projectors associated with a specific PVM. Given a pre-selection |ψ〉, a post-selection |ϕ〉, and a PVM characterized by parts (projectors) Bi, the weak value associated with each operator Bi is defined as tr(Bi ρψ|ϕ). Since ∑itr(Bi ρψ|ϕ), the weak values of this set of projectors can be seen as a quasi-probability distribution. It is possible, then, to use these quasi-probabilities to construct a weak logical entropy for post-selected quantum systems, which gives L π w ( ρ ψ | ϕ ) i tr ( B i ρ ψ | ϕ ) tr [ ( I - B i ) ρ ψ | ϕ ] = i tr ( B i ρ ψ | ϕ ) [ 1 - tr ( B i ρ ψ | ϕ ) ] = tr [ ρ ψ | ϕ ( I - ρ ψ | ϕ ) ] . $$ \begin{array}{cc}{L}_{\pi }^w({\rho }_{\psi |\phi })& \equiv \sum_i \mathrm{tr}({B}_i{\rho }_{\psi |\phi })\mathrm{tr}[(I-{B}_i){\rho }_{\psi |\phi }]\\ & =\sum_i \mathrm{tr}({B}_i{\rho }_{\psi |\phi })[1-\mathrm{tr}({B}_i{\rho }_{\psi |\phi })]\\ & =\mathrm{tr}\left[{{\rho }^\mathrm{\prime}}_{\psi |\phi }\left(I-{{\rho }^\mathrm{\prime}}_{\psi |\phi }\right)\right].\end{array} $$(64)

Here, again, a PVM-independent logical entropy cannot be defined. However, just like the extension of the von Neumman entropy to post-selected systems [36], the weak logical entropy can be negative. Even further, it may be complex-valued.

While the operational meaning of Lπ (ρψ|ϕ) is clear, it is not straightforward to interpret L π w ( ρ ψ | ϕ ) $ {L}_{\pi }^w({\rho }_{\psi |\phi })$. One of the reasons is that weak values are associated with weak measurements, and weak measurements do not divide the Hilbert space into disjoints parts. In fact, weak measurements can be seen as an interpolation between the identity map, which does not disturb the system but does not provide any information about it, and the associated PVM map, which provides maximum information about the system [51]. The weaker the measurement, the bigger the contribution of the identity map in this interpolation. It is clear from the discussion in this work that, while the PVM introduces partitions to the Hilbert space, the indetity map does not make distinction between any part. This fact in itself can be seen as the reason we end up with a quasiprobability (and not a probability) distribution.

However, consider the following game: A system is prepared in a state |ψ〉. Later, a weak measurement of one of the blocks Bi of a given PVM is performed. The system is then post-selected in the state |ϕ〉. If this process is repeated and the PVM is composed by n blocks, then the probability that the weak measurement of a different block is performed is simply (n − 1)/n. However, suppose that, instead, we wanted to refer to the obtained weak value as the quasiprobability associated with each block. In this case, keeping an algebraic analogy to the classical logical entropy defined in (2) and the PVM-dependent quantum logical entropy defined in (5), we obtain the weak logical entropy L π w ( ρ ψ | ϕ ) $ {L}_{\pi }^w({\rho }_{\psi |\phi })$.

Interestingly, even though the logical and the weak logical entropies of post-selected systems are associated with different operational meanings, it should be noted that L π ( ρ ψ | ϕ ) = | L π w ( ρ ψ | ϕ ) | 2 . $$ {L}_{\pi }({\rho }_{\psi |\phi })=|{L}_{\pi }^w({\rho }_{\psi |\phi }){|}^2. $$(65)

Thus, in this sense, the weak logical entropy is mathematically more fundamental than the logical entropy of post-selected systems first introduced in this section.

5 Discussion

In this work we presented a new approach to the introduction of quantum logical entropy that maintains a clear connection to the notion of partition. As we discussed, the PVM-independent definition of this entropy is equivalent to other already known quantities. However, the perspective of partitions that comes with its introduction may bring new insights into its use in the study of quantum systems within different scenarios. Furthermore, it may lead to the discovery of new implications for the logical structure, uncertainty, and information processing associated with quantum systems.

In fact, the properties of logical entropies proved here may be relevant to quantum information science and technology as well as other areas of quantum mechanics, as already shown, e.g., in [11, 5255]. In particular, a special case of Proposition 3.6 concerns quantum systems in the presence of noise. In this case, the logical entropy of the system of interest after its interaction with the noise is bounded by correlations between this system and its environment. This scenario, including an example of amplitude damping, was considered in [56]. Moreover, since the logical entropy is a measure of distinction, it seems natural to use it in the investigation of channel capacity in terms of distinctions. We expect that the language of quantum distinctions may simplify the proofs of channel properties. It would be also interesting to examine the use of logical entropy in the context of entanglement quantification in discrete/continuous systems.

Notably, while some of the properties of logical entropies are similar to the ones held by the more widely studied von Neumann entropy, many of them also evidence major differences. It stands out, for instance, that the logical entropy does not fulfill the strong subadditivity property. Even though it still satisfies subadditive properties, as stated in Propositions 3.2 and 3.3, the “breakdown” of strong additivity implies an unmistakable departure from the von Neumann entropy. The lack of this property might at first appear somewhat troublesome. However, it may also place the quantum logical entropy as a fundamental player in various special scenarios, e.g., when discussing the breakdown of subadditivity in black holes [57].

Finally, we briefly extended the notion of logical entropy to post-selected quantum systems in two ways. In the first, there exists a clear operational meaning associated with the logical entropy. In the second (weak logical entropy), there exists only a loose connection with experimental procedures associated with weak measurements. It would be important to study better the significance of the latter. Furthermore, properties of these logical entropies need yet to be investigated. Also, a relation between both definitions can be established as seen in (65), where the square of the magnitude of the weak logical entropy gives the logical entropy of post-selected systems. This raises a question about the phase of the weak logical entropy. While we could not find a direct meaning to it, it seems to us that this problem deserves further analysis.

Conflict of interest

The authors declare that they have no conflict of interest.

Acknowledgments

We thank D. Ellerman and Y. Neuman for many discussions on the topic as well as their helpful comments. We are also thankful to T. Landsberger and J. Kupferman for helpful feedback regarding a previous version of this manuscript. E.C. was supported by Grant No. FQXi-RFP-CPW-2006 from the Foundational Questions Institute and Fetzer Franklin Fund, a donor advised fund of Silicon Valley Community Foundation, the Israel Innovation Authority under Projects No. 70002 and No. 73795, the Quantum Science and Technology Program of the Israeli Council of Higher Education, and the Pazy Foundation.

References

  1. Brukner Č, Zeilinger A (1999), Operationally invariant information in quantum measurements. Phys Rev Lett 83, 3354. [CrossRef] [Google Scholar]
  2. Brukner Č, Zeilinger A (2001), Conceptual inadequacy of the Shannon information in quantum measurements. Phys Rev A 63, 022113. [CrossRef] [Google Scholar]
  3. Giraldi F, Grigolini P (2001), Quantum entanglement and entropy. Phys Rev A 64, 032310. [CrossRef] [Google Scholar]
  4. Cover TM, Thomas JA (2006), Elements of Information Theory, 2nd edn., Wiley, Hoboken, New Jersey. [Google Scholar]
  5. Ellerman D (2013), Information as distinctions: New foundations for information theory, arXiv:1301.5607. [Google Scholar]
  6. Tsallis C (1988), Possible generalization of Boltzmann-Gibbs statistics. J Stat Phys 52, 479. [CrossRef] [Google Scholar]
  7. Manfredi G, Feix MR (2000), Entropy and Wigner functions. Phys Rev E 62, 4665. [CrossRef] [PubMed] [Google Scholar]
  8. Ellerman D (2013), An introduction to logical entropy and its relation to Shannon entropy. Int J Semant Comput 7, 121. [CrossRef] [Google Scholar]
  9. Ellerman D (2014), An introduction to partition logic. Log J IGPL 22, 94. [CrossRef] [Google Scholar]
  10. Tamir B, Cohen E (2014), Logical entropy for quantum states, arXiv:1412.0616. [Google Scholar]
  11. Tamir B, Cohen E (2015), A Holevo-type bound for a Hilbert Schmidt distance measure. J Quantum Inf Sci 5, 127. [CrossRef] [Google Scholar]
  12. Gini C (1912), Variabilità e Mutabilità: Contributo allo Studio delle Distribuzioni e delle Relazioni Statistiche. [Fasc. I.]. Studi economico-giuridici pubblicati per cura della facoltà di Giurisprudenza della R. Università di Cagliari. Tipogr. di P. Cuppini. [Google Scholar]
  13. Rejewski M (1981), How Polish mathematicians broke the Enigma Cipher. Ann Hist Comput 3, 213. [CrossRef] [Google Scholar]
  14. Patil GP, Taillie C (1982), Diversity as a concept and its measurement. J Am Stat Assoc 77, 548. [CrossRef] [Google Scholar]
  15. Good IJ (1982), Comment: Diversity as a concept and its measurement. J Am Stat Assoc 77, 561. [Google Scholar]
  16. Nielsen MA, Chuang IL (2010), Quantum information and quantum computation, University Press, Cambridge. [CrossRef] [Google Scholar]
  17. Ellerman D (2014), Partitions and objective indefiniteness in quantum mechanics, arXiv:1401.2421. [Google Scholar]
  18. Ellerman D (2016), On classical and quantum logical entropy. Available at SSRN 2770162. https://doi.org/10.2139/ssrn.2770162 . [Google Scholar]
  19. Ellerman D (2016), On classical and quantum logical entropy: The analysis of measurement, arXiv:1604.04985. [Google Scholar]
  20. Ellerman D (2017), New logical foundations for quantum information theory: Introduction to quantum logical information theory, arXiv:1707.04728. [Google Scholar]
  21. Ellerman D (2017), Introduction to quantum logical information theory, Available at SSRN 3003279. https://doi.org/10.2139/ssrn.3003279 . [Google Scholar]
  22. Ellerman D (2018), Introduction to quantum logical information theory: Talk. EPJ Web Conf 182, 02039. [CrossRef] [EDP Sciences] [Google Scholar]
  23. Ellerman D (2018), Logical entropy: Introduction to classical and quantum logical information theory. Entropy 20, 679. [CrossRef] [Google Scholar]
  24. Ellerman D (2016), The quantum logic of direct-sum decompositions. Available at SSRN 2770163. https://doi.org/10.2139/ssrn.2770163. [Google Scholar]
  25. Ellerman D (2018), The quantum logic of direct-sum decompositions: the dual to the quantum logic of subspaces. Log J IGPL 26, 1. [CrossRef] [Google Scholar]
  26. Auletta G, Fortunato M, Parisi G (2009), Quantum mechanics, Cambridge University Press, New York. [Google Scholar]
  27. Lüders G (1950), Über die Zustandsänderung durch den Meßprozeß. Ann Phys 443, 322. [CrossRef] [Google Scholar]
  28. von Neumann J (1932), Mathematische Grundlagen der Quantenmechanik, Springer, Berlin. [Google Scholar]
  29. Jaeger G (2007), Quantum information: an overview, Springer, New York. [Google Scholar]
  30. Coles PJ (2011), Non-negative discord strengthens the subadditivity of quantum entropy functions, arXiv:1101.1717. [Google Scholar]
  31. Audenaert KMR (2007), Subadditivity of q-entropies for q > 1. J Math Phys 48, 083507. [CrossRef] [Google Scholar]
  32. Streltsov A, Kampermann H, Wolk S, Gessner M, Brub D (2018), Maximal coherence and the resource theory of purity. New J Phys 20, 053058. [CrossRef] [Google Scholar]
  33. Uhlmann A (1970), On the Shannon entropy and related functionals on convex sets. Rep Math Phys 1, 147. [CrossRef] [Google Scholar]
  34. Nielsen MA (2002), An introduction to majorization and its applications to quantum mechanics (Lecture Notes), Department of Physics, University of Queensland, Australia. [Google Scholar]
  35. Aharonov Y, Bergmann PG, Lebowitz JL (1964), Time symmetry in the quantum process of measurement. Phys Rev 134, B1410. [CrossRef] [Google Scholar]
  36. Salek S, Schubert R, Wiesner K (2014), Negative conditional entropy of postselected states. Phys Rev A 90, 022116. [CrossRef] [Google Scholar]
  37. Aharonov Y, Albert DZ, Vaidman L (1988), How the result of a measurement of a component of the spin of a spin-1/2 particle can turn out to be 100. Phys Rev Lett 60, 1351. [CrossRef] [PubMed] [Google Scholar]
  38. Dixon PB, Starling DJ, Jordan AN, Howell JC (2009), Ultrasensitive beam deflection measurement via interferometric weak value amplification. Phys Rev Lett 102, 173601. [CrossRef] [PubMed] [Google Scholar]
  39. Jacobs K (2009), Second law of thermodynamics and quantum feedback control: Maxwell’s demon with weak measurements. Phys Rev A 80, 012322. [CrossRef] [Google Scholar]
  40. Turner MD, Hagedorn CA, Schlamminger S, Gundlach JH (2011), Picoradian deflection measurement with an interferometric quasi-autocollimator using weak value amplification. Opt Lett 36, 1479. [CrossRef] [PubMed] [Google Scholar]
  41. Aharonov Y, Cohen E, Elitzur AC (2014), Foundations and applications of weak quantum measurements. Phys Rev A 89, 052105. [CrossRef] [Google Scholar]
  42. Dressel J, Malik M, Miatto FM, Jordan AN, Boyd RW (2014), Colloquium: Understanding quantum weak values: Basics and applications. Rev Mod Phys 86, 307. [CrossRef] [Google Scholar]
  43. Alves GB, Escher BM, de Matos Filho RL, Zagury N, Davidovich L (2015), Weak-value amplification as an optimal metrological protocol. Phys Rev A 91, 062107. [CrossRef] [Google Scholar]
  44. Cortez L, Chantasri A, Garca-Pintos LP, Dressel J, Jordan AN (2017), Rapid estimation of drifting parameters in continuously measured quantum systems. Phys Rev A 95, 012314. [CrossRef] [Google Scholar]
  45. Kim Y, Kim Y-S, Lee S-Y, Han S-W, Moon S, Kim Y-H, Cho Y-W (2018), Direct quantum process tomography via measuring sequential weak values of incompatible observables. Nat Commun 9, 192. [CrossRef] [PubMed] [Google Scholar]
  46. Naghiloo M, Alonso JJ, Romito A, Lutz E, Murch KW (2018), Information gain and loss for a quantum Maxwell’s demon. Phys Rev Lett 121, 030604. [CrossRef] [PubMed] [Google Scholar]
  47. Hu M-J, Zhou Z-Y, Hu X-M, Li C-F, Guo G-C, Zhang Y-S (2018), Observation of non-locality sharing among three observers with one entangled pair via optimal weak measurement. npj Quantum Inf 4, 63. [CrossRef] [Google Scholar]
  48. Pfender M, Wang P, Sumiya H, Onoda S, Yang W, Dasari DBR, Neumann P, Pan X-Y, Isoya J, Liu R-B, Wrachtrup J (2019), High-resolution spectroscopy of single nuclear spins via sequential weak measurements. Nat Commun 10, 594. [CrossRef] [PubMed] [Google Scholar]
  49. Cujia KS, Boss JM, Herb K, Zopes J, Degen CL (2019), Tracking the precession of single nuclear spins by weak measurements. Nature 571, 230. [CrossRef] [PubMed] [Google Scholar]
  50. Paiva IL, Aharonov Y, Tollaksen J, Waegell M (2021), Aharonov-Bohm effect with an effective complex-valued vector potential, arXiv:2101.11914. [Google Scholar]
  51. Dieguez PR, Angelo RM (2018), Information-reality complementarity: The role of measurements and quantum reference frames. Phys. Rev. A 97, 022107. [CrossRef] [Google Scholar]
  52. Zurek WH, Habib S, Paz JP (1993), Coherent states via decoherence. Phys Rev Lett 70, 1187. [CrossRef] [PubMed] [Google Scholar]
  53. Buscemi F, Bordone P, Bertoni A (2007), Linear entropy as an entanglement measure in two-fermion systems. Phys Rev A 75, 032301. [CrossRef] [Google Scholar]
  54. Nayak AS, Devi ARU, Rajagopal AK, et al. (2017), Biseparability of noisy pseudopure, W and GHZ states using conditional quantum relative Tsallis entropy. Quantum Inf Process 16, 51. [CrossRef] [Google Scholar]
  55. Khordad R, Sedehi HRR (2017), Application of non-extensive entropy to study of decoherence of RbCl quantum dot qubit: Tsallis entropy. Superlattices Microstruct 101, 559. [CrossRef] [Google Scholar]
  56. Tamir B (2017), Tsallis entropy is natural in the formulation of quantum noise, arXiv:1702.07864. [Google Scholar]
  57. Almheiri A, Marolf D, Polchinski J, Sully J (2013), Black holes: complementarity or firewalls? J High Energy Phys 2, 62. [CrossRef] [Google Scholar]

Cite this article as: Tamir B, Paiva I, Schwartzman-Nowik Z & Cohen E 2022. Quantum logical entropy: fundamentals and general properties. 4open, 5, 2.

All Figures

thumbnail Figure 1

Pictorial representation of a set with a partition and its distinction. (a) A set U with six elements ui is divided in a partition π with three blocks (green areas). (b) The set U × U has each of its points characterized by a square in a 2D mesh. Points that belong to dit(π), i.e., pairs whose components belong to different blocks are colored in green.

In the text
thumbnail Figure 2

Representation of the probability distribution used in the definition of quantum logical entropy. Any state ρ, together with a partition characterized by a PVM, which is associated with an orthonormal basis in the corresponding Hilbert space, defines a probability distribution. This is the case even in case of coarse (i.e., degenerate) PVMs, as illustrated by the yellow partition. These distributions allow the introduction of (PVM-dependent) logical entropies for quantum systems.

In the text

Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.

Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.

Initial download of the metrics may take a while.