Skip to main content
IBM Quantum Platform

Equivalence of the representations

We've now discussed three different ways to represent channels in mathematical terms, namely Stinespring representations, Kraus representations, and Choi representations. We also have the definition of a channel, which states that a channel is a linear mapping that always transforms density matrices into density matrices, even when the channel is applied to just part of a compound system. The remainder of the lesson is devoted to a mathematical proof that the three representations are equivalent and precisely capture the definition.


Overview of the proof

Our goal is to establish the equivalence of a collection of four statements, and we'll begin by writing them down precisely. All four statements follow the same conventions that have been used throughout the lesson, namely that Φ\Phi is a linear mapping from square matrices to square matrices, the rows and columns of the input matrices have been placed in correspondence with the classical states of a system X\mathsf{X} (the input system), and the rows and columns of the output matrices have been placed in correspondence with the classical states of a system Y\mathsf{Y} (the output system).

  1. Φ\Phi is a channel from X\mathsf{X} to Y.\mathsf{Y}. That is, Φ\Phi always transforms density matrices to density matrices, even when it acts on one part of a larger compound system.

  2. The Choi matrix J(Φ)J(\Phi) is positive semidefinite and satisfies the condition TrY(J(Φ))=IX.\operatorname{Tr}_{\mathsf{Y}}(J(\Phi)) = \mathbb{I}_{\mathsf{X}}.

  3. There is a Kraus representation for Φ.\Phi. That is, there exist matrices A0,,AN1A_0,\ldots,A_{N-1} for which the equation Φ(ρ)=k=0N1AkρAk\Phi(\rho) = \sum_{k = 0}^{N-1} A_k \rho A_k^{\dagger} is true for every input ρ,\rho, and that satisfy the condition k=0N1AkAk=IX.\sum_{k = 0}^{N-1} A_k^{\dagger} A_k = \mathbb{I}_{\mathsf{X}}.

  4. There is a Stinespring representation for Φ.\Phi. That is, there exist systems W\mathsf{W} and G\mathsf{G} for which the pairs (W,X)(\mathsf{W},\mathsf{X}) and (G,Y)(\mathsf{G},\mathsf{Y}) have the same number of classical states, along with a unitary matrix UU representing a unitary operation from (W,X)(\mathsf{W},\mathsf{X}) to (G,Y),(\mathsf{G},\mathsf{Y}), such that Φ(ρ)=TrG(U(00ρ)U).\Phi(\rho) = \operatorname{Tr}_{\mathsf{G}}\bigl( U (\vert 0\rangle\langle 0 \vert \otimes \rho) U^{\dagger} \bigr).

The way the proof works is that a cycle of implications is proved: the first statement in our list implies the second, the second implies the third, the third implies the fourth, and the fourth statement implies the first. This establishes that all four statements are equivalent — which is to say that they're either all true or all false for a given choice of Φ\Phi — because the implications can be followed transitively from any one statement to any other.

This is a common strategy when proving that a collection of statements are equivalent, and a useful trick to use in such a context is to set up the implications in a way that makes them as easy to prove as possible. That is the case here — and in fact we've already encountered two of the four implications.


Channels to Choi matrices

Referring to the statements listed above by their numbers, the first implication to be proved is 1 \Rightarrow 2. This implication was already discussed in the context of the Choi state of a channel. Here we'll summarize the mathematical details.

Assume that the classical state set of the input system X\mathsf{X} is Σ\Sigma and let n=Σ.n = \vert\Sigma\vert. Consider the situation in which Φ\Phi is applied to the second of two copies of X\mathsf{X} together in the state

ψ=1naΣaa,\vert \psi \rangle = \frac{1}{\sqrt{n}} \sum_{a \in \Sigma} \vert a \rangle \otimes \vert a \rangle,

which, as a density matrix, is given by

ψψ=1na,bΣabab.\vert \psi \rangle \langle \psi \vert = \frac{1}{n} \sum_{a,b \in \Sigma} \vert a\rangle\langle b \vert \otimes \vert a\rangle\langle b \vert.

The result can be written as

(IdΦ)(ψψ)=1na,b=0n1abΦ(ab)=J(Φ)n,(\operatorname{Id}\otimes \,\Phi) \bigl(\vert \psi \rangle \langle \psi \vert\bigr) = \frac{1}{n} \sum_{a,b = 0}^{n-1} \vert a\rangle\langle b \vert \otimes \Phi\bigl(\vert a\rangle\langle b \vert\bigr) = \frac{J(\Phi)}{n},

and by the assumption that Φ\Phi is a channel this must be a density matrix. Like all density matrices it must be positive semidefinite, and multiplying a positive semidefinite matrix by a positive real number yields another positive semidefinite matrix, and therefore J(Φ)0.J(\Phi) \geq 0.

Moreover, under the assumption that Φ\Phi is a channel, it must preserve trace, and therefore

TrY(J(Φ))=a,bΣTr(Φ(ab))ab=a,bΣTr(ab)ab=aΣaa=IX.\begin{aligned} \operatorname{Tr}_{\mathsf{Y}} (J(\Phi)) & = \sum_{a,b\in\Sigma} \operatorname{Tr}\bigl(\Phi( \vert a\rangle\langle b \vert)\bigr) \, \vert a\rangle\langle b \vert\\ & = \sum_{a,b\in\Sigma} \operatorname{Tr}\bigl(\vert a\rangle\langle b \vert\bigr) \, \vert a\rangle\langle b \vert\\ & = \sum_{a\in\Sigma} \vert a\rangle\langle a \vert\\ & = \mathbb{I}_{\mathsf{X}}. \end{aligned}

Choi to Kraus representations

The second implication, again referring to the statements in our list by their numbers, is 2 \Rightarrow 3. To be clear, we're ignoring the other statements — and in particular we cannot make the assumption that Φ\Phi is a channel. All we have to work with is that Φ\Phi is a linear mapping whose Choi representation satisfies J(Φ)0J(\Phi) \geq 0 and TrY(J(Φ))=IX.\operatorname{Tr}_{\mathsf{Y}} (J(\Phi)) = \mathbb{I}_{\mathsf{X}}.

This, however, is all we need to conclude that Φ\Phi has a Kraus representation

Φ(ρ)=k=0N1AkρAk\Phi(\rho) = \sum_{k = 0}^{N-1} A_k \rho A_k^{\dagger}

for which the condition

k=0N1AkAk=IX\sum_{k = 0}^{N-1} A_k^{\dagger} A_k = \mathbb{I}_{\mathsf{X}}

is satisfied.

We begin with the critically important assumption that J(Φ)J(\Phi) is positive semidefinite, which means that it is possible to express it in the form

J(Φ)=k=0N1ψkψk(1)J(\Phi) = \sum_{k = 0}^{N-1} \vert \psi_k \rangle \langle \psi_k \vert \tag{1}

for some way of choosing the vectors ψ0,,ψN1.\vert\psi_0\rangle,\ldots,\vert\psi_{N-1}\rangle. In general there will be multiple ways to do this — and in fact this directly mirrors the freedom one has in choosing a Kraus representation for Φ.\Phi.

One way to obtain such an expression is to first use the spectral theorem to write

J(Φ)=k=0N1λkγkγk,J(\Phi) = \sum_{k = 0}^{N-1} \lambda_k \vert \gamma_k \rangle \langle \gamma_k \vert,

in which λ0,,λN1\lambda_0,\ldots,\lambda_{N-1} are the eigenvalues of J(Φ)J(\Phi) (which are necessarily nonnegative real numbers because J(Φ)J(\Phi) is positive semidefinite) and γ0,,γN1\vert\gamma_0\rangle,\ldots,\vert\gamma_{N-1}\rangle are unit eigenvectors corresponding to the eigenvalues λ0,,λN1.\lambda_0,\ldots,\lambda_{N-1}.

Note that, while there's no freedom in choosing the eigenvalues (except for how they're ordered), there is freedom in the choice of the eigenvectors, particularly when there are eigenvalues with multiplicity larger than one. So, this is not a unique expression of J(Φ)J(\Phi) — we're just assuming we have one such expression. Regardless, because the eigenvalues are nonnegative real numbers, they have nonnegative square roots, and so we can select

ψk=λkγk\vert\psi_k\rangle = \sqrt{\lambda_k} \vert \gamma_k\rangle

for each k=0,,N1k = 0,\ldots,N-1 to obtain an expression of the form (1).(1).

It is, however, not essential that the expression (1)(1) comes from a spectral decomposition in this way, and in particular the vectors ψ0,,ψN1\vert\psi_0\rangle,\ldots,\vert\psi_{N-1}\rangle need not be orthogonal in general. It is noteworthy, though, that we can choose these vectors to be orthogonal if we wish — and moreover we never need NN to be larger than nmnm (recalling that nn and mm denote the numbers of classical states of X\mathsf{X} and Y,\mathsf{Y}, respectively).

Next, each of the vectors ψ0,,ψN1\vert\psi_0\rangle,\ldots,\vert\psi_{N-1}\rangle can be further decomposed as

ψk=aΣaϕk,a,\vert\psi_k\rangle = \sum_{a\in\Sigma} \vert a\rangle \otimes \vert \phi_{k,a}\rangle,

where the vectors {ϕk,a}\{ \vert \phi_{k,a}\rangle \} have entries corresponding to the classical states of Y\mathsf{Y} and can be explicitly determined by the equation

ϕk,a=(aIY)ψk\vert \phi_{k,a}\rangle = \bigl( \langle a \vert \otimes \mathbb{I}_{\mathsf{Y}}\bigr) \vert \psi_k\rangle

for each aΣa\in\Sigma and k=0,,N1.k=0,\ldots,N-1. Although ψ0,,ψN1\vert\psi_0\rangle,\ldots,\vert\psi_{N-1}\rangle are not necessarily unit vectors, this is the same process we would use to analyze what would happen if a standard basis measurement was performed on the system X\mathsf{X} given a quantum state vector of the pair (X,Y).(\mathsf{X},\mathsf{Y}).

And now we come to the trick that makes this part of the proof work. We define our Kraus matrices A0,,AN1A_0,\ldots,A_{N-1} according to the following equation.

Ak=aΣϕk,aaA_k = \sum_{a\in\Sigma} \vert \phi_{k,a}\rangle\langle a \vert

We can think about this formula purely symbolically: a\vert a\rangle effectively gets flipped around to form a\langle a\vert and moved to right-hand side, forming a matrix. For the purposes of verifying the proof, the formula is all we need.

There is, however, a simple and intuitive relationship between the vector ψk\vert\psi_k\rangle and the matrix Ak,A_k, which is that by vectorizing AkA_k we get ψk.\vert\psi_k\rangle. What it means to vectorize AkA_k is that we stack the columns on top of one another (with the leftmost column on top proceeding to the rightmost on the bottom), in order to form a vector. For instance, if X\mathsf{X} and Y\mathsf{Y} are both qubits, and for some choice of kk we have

ψk=α0000+α0101+α1010+α1111=(α00α01α10α11),\begin{aligned} \vert\psi_k\rangle & = \alpha_{00} \vert 0\rangle \otimes \vert 0\rangle + \alpha_{01} \vert 0\rangle \otimes \vert 1\rangle + \alpha_{10} \vert 1\rangle \otimes \vert 0\rangle + \alpha_{11} \vert 1\rangle \otimes \vert 1\rangle\\[2mm] & = \begin{pmatrix} \alpha_{00} \\[1mm] \alpha_{01} \\[1mm] \alpha_{10} \\[1mm] \alpha_{11} \end{pmatrix}, \end{aligned}

then

Ak=α0000+α0110+α1001+α1111=(α00α10α01α11).\begin{aligned} A_k & = \alpha_{00} \vert 0\rangle\langle 0\vert + \alpha_{01} \vert 1\rangle\langle 0\vert + \alpha_{10} \vert 0\rangle\langle 1\vert + \alpha_{11} \vert 1\rangle\langle 1\vert\\[2mm] & = \begin{pmatrix} \alpha_{00} & \alpha_{10}\\[1mm] \alpha_{01} & \alpha_{11} \end{pmatrix}. \end{aligned}

(Beware: sometimes the vectorization of a matrix is defined in a slightly different way, which is that the rows of the matrix are transposed and stacked on top of one another to form a column vector.)

First we'll verify that this choice of Kraus matrices correctly describes the mapping Φ,\Phi, after which we'll verify the other required condition. To keep things straight, let's define a new mapping Ψ\Psi as follows.

Ψ(ρ)=k=0N1AkρAk\Psi(\rho) = \sum_{k = 0}^{N-1} A_k \rho A_k^{\dagger}

Thus, our goal is to verify that Ψ=Φ.\Psi = \Phi.

The way we can do this is to compare the Choi representations of these mappings. Choi representations are faithful, so we have Ψ=Φ\Psi = \Phi if and only if J(Φ)=J(Ψ).J(\Phi) = J(\Psi). At this point we can simply compute J(Ψ)J(\Psi) using the expressions

ψk=aΣaϕk,aandAk=aΣϕk,aa\vert\psi_k\rangle = \sum_{a\in\Sigma} \vert a\rangle \otimes \vert \phi_{k,a}\rangle \quad\text{and}\quad A_k = \sum_{a\in\Sigma} \vert \phi_{k,a}\rangle\langle a \vert

together with the bilinearity of tensor products to simplify.

J(Ψ)=a,bΣabk=0N1AkabAk=a,bΣabk=0N1ϕk,aϕk,b=k=0N1(aΣaϕk,a)(bΣbϕk,b)=k=0N1ψkψk=J(Φ)\begin{aligned} J(\Psi) & = \sum_{a,b\in\Sigma} \vert a\rangle \langle b \vert \otimes \sum_{k = 0}^{N-1} A_k \vert a\rangle \langle b \vert A_k^{\dagger}\\[2mm] & = \sum_{a,b\in\Sigma} \vert a\rangle \langle b \vert \otimes \sum_{k = 0}^{N-1} \vert \phi_{k,a} \rangle \langle \phi_{k,b} \vert \\[2mm] & = \sum_{k = 0}^{N-1} \biggl(\sum_{a\in\Sigma} \vert a\rangle \otimes \vert \phi_{k,a} \rangle\biggr) \biggl(\sum_{b\in\Sigma} \langle b\vert \otimes \langle \phi_{k,b} \vert\biggr)\\[2mm] & = \sum_{k = 0}^{N-1} \vert \psi_k \rangle \langle \psi_k \vert \\[2mm] & = J(\Phi) \end{aligned}

So, our Kraus matrices correctly describe Φ.\Phi.

It remains to check the required condition on A0,,AN1,A_0,\ldots,A_{N-1}, which turns out to be equivalent to the assumption TrY(J(Φ))=IX\operatorname{Tr}_{\mathsf{Y}}(J(\Phi)) = \mathbb{I}_{\mathsf{X}} (which we haven't used yet). What we'll show is this relationship:

(k=0N1AkAk)T=TrY(J(Φ))(2)\Biggl( \sum_{k = 0}^{N-1} A_k^{\dagger} A_k \Biggr)^{T} = \operatorname{Tr}_{\mathsf{Y}}(J(\Phi)) \tag{2}

(in which we're referring the matrix transpose on the left-hand side).

Starting on the left, we can first observe that

(k=0N1AkAk)T=(k=0N1a,bΣbϕk,bϕk,aa)T=k=0N1a,bΣϕk,bϕk,aab.\begin{aligned} \Biggl(\sum_{k = 0}^{N-1} A_k^{\dagger} A_k\Biggr)^T & = \Biggl(\sum_{k = 0}^{N-1} \sum_{a,b\in\Sigma} \vert b \rangle \langle \phi_{k,b} \vert \phi_{k,a} \rangle \langle a \vert\Biggr)^T\\ & = \sum_{k = 0}^{N-1} \sum_{a,b\in\Sigma} \langle \phi_{k,b} \vert \phi_{k,a} \rangle \vert a \rangle \langle b \vert. \end{aligned}

The last equality follows from the fact that the transpose is linear and maps ba\vert b\rangle\langle a \vert to ab.\vert a\rangle\langle b \vert.

Moving to the right-hand side of our equation, we have

J(Φ)=k=0N1ψkψk=k=0N1a,bΣabϕk,aϕk,bJ(\Phi) = \sum_{k = 0}^{N-1} \vert \psi_k\rangle\langle\psi_k \vert = \sum_{k = 0}^{N-1} \sum_{a,b\in\Sigma} \vert a\rangle \langle b \vert \otimes \vert\phi_{k,a}\rangle\langle \phi_{k,b} \vert

and therefore

TrY(J(Φ))=k=0N1a,bΣTr(ϕk,aϕk,b)ab=k=0N1a,bΣϕk,bϕk,aab.\begin{aligned} \operatorname{Tr}_{\mathsf{Y}}(J(\Phi)) & = \sum_{k = 0}^{N-1} \sum_{a,b\in\Sigma} \operatorname{Tr}\bigl(\vert\phi_{k,a}\rangle\langle \phi_{k,b} \vert \bigr)\, \vert a\rangle \langle b \vert\\ & = \sum_{k = 0}^{N-1} \sum_{a,b\in\Sigma} \langle \phi_{k,b} \vert \phi_{k,a} \rangle \vert a \rangle \langle b \vert. \end{aligned}

We've obtained the same result, and therefore the equation (2)(2) has been verified. It follows, by the assumption TrY(J(Φ))=IX,\operatorname{Tr}_{\mathsf{Y}} (J(\Phi)) = \mathbb{I}_{\mathsf{X}}, that

(k=0N1AkAk)T=IX\Biggl(\sum_{k = 0}^{N-1} A_k^{\dagger} A_k\Biggr)^T = \mathbb{I}_{\mathsf{X}}

and therefore, because the identity matrix is its own transpose, the required condition is true.

k=0N1AkAk=IX\sum_{k = 0}^{N-1} A_k^{\dagger} A_k = \mathbb{I}_{\mathsf{X}}

Kraus to Stinespring representations

Now suppose that we have a Kraus representation of a mapping

Φ(ρ)=k=0N1AkρAk\Phi(\rho) = \sum_{k = 0}^{N-1} A_k \rho A_k^{\dagger}

for which

k=0N1AkAk=IX.\sum_{k = 0}^{N-1} A_k^{\dagger} A_k = \mathbb{I}_{\mathsf{X}}.

Our goal is to find a Stinespring representation for Φ.\Phi.

What we'd like to do first is to choose the garbage system G\mathsf{G} so that its classical state set is {0,,N1}.\{0,\ldots,N-1\}. For (W,X)(\mathsf{W},\mathsf{X}) and (G,Y)(\mathsf{G},\mathsf{Y}) to have the same size, however, it must be that nn divides mN,m N, allowing us to take W\mathsf{W} to have classical states {0,,d1}\{0,\ldots,d-1\} for d=mN/n.d = mN/n.

For an arbitrary choice of n,n, m,m, and N,N, it may not be the case that mN/nmN/n is an integer, so we're not actually free to choose G\mathsf{G} so that it's classical state set is {0,,N1}.\{0,\ldots,N-1\}. But we can always increase NN arbitrarily in the Kraus representation by choosing Ak=0A_k = 0 for however many additional values of kk that we wish.

And so, if we tacitly assume that mN/nmN/n is an integer, which is equivalent to NN being a multiple of m/gcd(n,m),m/\operatorname{gcd}(n,m), then we're free to take G\mathsf{G} so that its classical state set is {0,,N1}.\{0,\ldots,N-1\}. In particular, if it is the case that N=nm,N = nm, then we may take W\mathsf{W} to have m2m^2 classical states.

It remains to choose U,U, and we'll do this by matching the following pattern.

U=(A0??A1??AN1??)U = \begin{pmatrix} A_{0} & \fbox{?} & \cdots & \fbox{?} \\[1mm] A_{1} & \fbox{?} & \cdots & \fbox{?} \\[1mm] \vdots & \vdots & \ddots & \vdots\\[1mm] A_{N-1} & \fbox{?} & \cdots & \fbox{?} \end{pmatrix}

To be clear, this pattern is meant to suggest a block matrix, where each block (including A0,,AN1A_{0},\ldots,A_{N-1} as well as the blocks marked with a question mark) has mm rows and nn columns. There are NN rows of blocks, which means that there are d=mN/nd = mN/n columns of blocks.

Expressed in more formulaic terms, we will define UU as

U=k=0N1j=0d1kjMk,j=(M0,0M0,1M0,d1M1,0M1,1M1,d1MN1,0MN1,1MN1,d1)\begin{aligned} U & = \sum_{k=0}^{N-1} \sum_{j=0}^{d-1} \vert k \rangle \langle j \vert \otimes M_{k,j} \\[4mm] & = \begin{pmatrix} M_{0,0} & M_{0,1} & \cdots & M_{0,d-1} \\[1mm] M_{1,0} & M_{1,1} & \cdots & M_{1,d-1} \\[1mm] \vdots & \vdots & \ddots & \vdots\\[1mm] M_{N-1,0} & M_{N-1,1} & \cdots & M_{N-1,d-1} \end{pmatrix} \end{aligned}

where each matrix Mk,jM_{k,j} has mm rows and nn columns, and in particular we shall take Mk,0=AkM_{k,0} = A_k for k=0,,N1.k = 0,\ldots,N-1.

This must be a unitary matrix, and the blocks labeled with a question mark, or equivalently Mk,jM_{k,j} for j>0,j>0, must be selected with this in mind — but aside from allowing UU to be unitary, the blocks labeled with a question mark won't have any relevance to the proof.

Let's momentarily disregard the concern that UU is unitary and focus on the expression

TrG(U(00Wρ)U)\operatorname{Tr}_{\mathsf{G}} \bigl( U (\vert 0\rangle \langle 0 \vert_{\mathsf{W}} \otimes \rho)U^{\dagger}\bigr)

that describes the output state of Y\mathsf{Y} given the input state ρ\rho of X\mathsf{X} for our Stinespring representation. We can alternatively write

U(00ρ)U=U(0IW)ρ(0IW)U,U(\vert 0\rangle\langle 0 \vert \otimes \rho)U^{\dagger} = U(\vert 0\rangle\otimes\mathbb{I}_{\mathsf{W}}) \rho (\langle 0\vert \otimes \mathbb{I}_{\mathsf{W}}) U^{\dagger},

and we see from our choice of UU that

U(0IW)=k=0N1kAk.U(\vert 0\rangle\otimes\mathbb{I}_{\mathsf{W}}) = \sum_{k = 0}^{N-1} \vert k\rangle \otimes A_k.

We therefore find that

U(00ρ)U=j,k=0N1kjAkρAj,U(\vert 0\rangle\langle 0 \vert \otimes \rho)U^{\dagger} = \sum_{j,k = 0}^{N-1} \vert k\rangle\langle j\vert \otimes A_k \rho A_j^{\dagger},

and so

TrG(U(00Wρ)U)=j,k=0N1Tr(kj)AkρAj=k=0N1AkρAk=Φ(ρ).\begin{aligned} \operatorname{Tr}_{\mathsf{G}} \bigl( U (\vert 0\rangle \langle 0 \vert_{\mathsf{W}} \otimes \rho) U^{\dagger}\bigr) & = \sum_{j,k = 0}^{N-1} \operatorname{Tr}\bigl(\vert k\rangle\langle j\vert\bigr) \, A_k \rho A_j^{\dagger} \\ & = \sum_{k = 0}^{N-1} A_k \rho A_k^{\dagger} \\ & = \Phi(\rho). \end{aligned}

We therefore have a correct representation for the mapping Φ,\Phi, and it remains to verify that we can choose UU to be unitary.

Consider the first nn columns of UU when it's selected according to the pattern above. Taking these columns alone, we have a block matrix

(A0A1AN1).\begin{pmatrix} A_0\\[1mm] A_1\\[1mm] \vdots\\[1mm] A_{N-1} \end{pmatrix}.

There are nn columns, one for each classical state of X,\mathsf{X}, and as vectors let us name the columns as γa\vert \gamma_a \rangle for each aΣ.a\in\Sigma. Here's a formula for these vectors that can be matched to the block matrix representation above.

γa=k=0N1kAka\vert \gamma_a\rangle = \sum_{k = 0}^{N-1} \vert k\rangle \otimes A_k \vert a \rangle

Now let's compute the inner product between any two of these vectors, meaning the ones corresponding to any choice of a,bΣ.a,b\in\Sigma.

γaγb=j,k=0N1kjaAkAjb=a(k=0N1AkAk)b\langle \gamma_a \vert \gamma_b \rangle = \sum_{j,k = 0}^{N-1} \langle k \vert j \rangle \, \langle a \vert A_k^{\dagger} A_j \vert b\rangle = \langle a \vert \Biggl( \sum_{k = 0}^{N-1} A_k^{\dagger} A_k \Biggr) \vert b\rangle

By the assumption

k=0m1AkAk=IX\sum_{k = 0}^{m-1} A_k^{\dagger} A_k = \mathbb{I}_{\mathsf{X}}

we conclude that the nn column vectors {γa:aΣ}\{\vert\gamma_a\rangle\,:\,a\in\Sigma\} form an orthonormal set:

γaγb={1a=b0ab\langle \gamma_a \vert \gamma_b \rangle = \begin{cases} 1 & a = b\\ 0 & a\neq b \end{cases}

for all a,bΣ.a,b\in\Sigma.

This implies that it is possible to fill out the remaining columns of UU so that it becomes a unitary matrix. In particular, the Gram-Schmidt orthogonalization process can be used to select the remaining columns. Something similar was done in the Quantum circuits lesson of "Basics of quantum information" in the context of the state discrimination problem.


Stinespring representations back to the definition

The final implication is 4 \Rightarrow 1. That is, we assume that we have a unitary operation transforming a pair of systems (W,X)(\mathsf{W},\mathsf{X}) into a pair (G,Y),(\mathsf{G},\mathsf{Y}), and our goal is to conclude that the mapping

Φ(ρ)=TrG(U(00Wρ)U)\Phi(\rho) = \operatorname{Tr}_{\mathsf{G}} \bigl( U (\vert 0\rangle \langle 0 \vert_{\mathsf{W}} \otimes \rho)U^{\dagger}\bigr)

is a valid channel. From its form, it is evident that Φ\Phi is linear, and it remains to verify that it always transforms density matrices into density matrices. This is pretty straightforward and we've already discussed the key points.

In particular, if we start with a density matrix σ\sigma of a compound system (Z,X),(\mathsf{Z},\mathsf{X}), and then add on an additional workspace system W,\mathsf{W}, we will certainly be left with a density matrix. If we reorder the systems (W,Z,X)(\mathsf{W},\mathsf{Z},\mathsf{X}) for convenience, we can write this state as

00Wσ.\vert 0\rangle\langle 0\vert_{\mathsf{W}} \otimes \sigma.

We then apply the unitary operation U,U, and as we already discussed this is a valid channel, and hence maps density matrices to density matrices. Finally, the partial trace of a density matrix is another density matrix.

Another way to say this is to observe first that each of these things is a valid channel:

  1. Introducing an initialized workspace system.
  2. Performing a unitary operation.
  3. Tracing out a system.

And finally, any composition of channels is another channel — which is immediate from the definition, but is also a fact worth observing in its own right.

This completes the proof of the final implication, and therefore we've established the equivalence of the four statements listed at the start of the section.

Was this page helpful?
Report a bug or request content on GitHub.