Equivalence of the representations
We've now discussed three different ways to represent channels in mathematical terms, namely Stinespring representations, Kraus representations, and Choi representations. We also have the definition of a channel, which states that a channel is a linear mapping that always transforms density matrices into density matrices, even when the channel is applied to just part of a compound system. The remainder of the lesson is devoted to a mathematical proof that the three representations are equivalent and precisely capture the definition.
Overview of the proof
Our goal is to establish the equivalence of a collection of four statements, and we'll begin by writing them down precisely. All four statements follow the same conventions that have been used throughout the lesson, namely that is a linear mapping from square matrices to square matrices, the rows and columns of the input matrices have been placed in correspondence with the classical states of a system (the input system), and the rows and columns of the output matrices have been placed in correspondence with the classical states of a system (the output system).
-
is a channel from to That is, always transforms density matrices to density matrices, even when it acts on one part of a larger compound system.
-
The Choi matrix is positive semidefinite and satisfies the condition
-
There is a Kraus representation for That is, there exist matrices for which the equation is true for every input and that satisfy the condition
-
There is a Stinespring representation for That is, there exist systems and for which the pairs and have the same number of classical states, along with a unitary matrix representing a unitary operation from to such that
The way the proof works is that a cycle of implications is proved: the first statement in our list implies the second, the second implies the third, the third implies the fourth, and the fourth statement implies the first. This establishes that all four statements are equivalent — which is to say that they're either all true or all false for a given choice of — because the implications can be followed transitively from any one statement to any other.
This is a common strategy when proving that a collection of statements are equivalent, and a useful trick to use in such a context is to set up the implications in a way that makes them as easy to prove as possible. That is the case here — and in fact we've already encountered two of the four implications.
Channels to Choi matrices
Referring to the statements listed above by their numbers, the first implication to be proved is 1 2. This implication was already discussed in the context of the Choi state of a channel. Here we'll summarize the mathematical details.
Assume that the classical state set of the input system is and let Consider the situation in which is applied to the second of two copies of together in the state
which, as a density matrix, is given by
The result can be written as
and by the assumption that is a channel this must be a density matrix. Like all density matrices it must be positive semidefinite, and multiplying a positive semidefinite matrix by a positive real number yields another positive semidefinite matrix, and therefore
Moreover, under the assumption that is a channel, it must preserve trace, and therefore
Choi to Kraus representations
The second implication, again referring to the statements in our list by their numbers, is 2 3. To be clear, we're ignoring the other statements — and in particular we cannot make the assumption that is a channel. All we have to work with is that is a linear mapping whose Choi representation satisfies and
This, however, is all we need to conclude that has a Kraus representation
for which the condition
is satisfied.
We begin with the critically important assumption that is positive semidefinite, which means that it is possible to express it in the form
for some way of choosing the vectors In general there will be multiple ways to do this — and in fact this directly mirrors the freedom one has in choosing a Kraus representation for
One way to obtain such an expression is to first use the spectral theorem to write
in which are the eigenvalues of (which are necessarily nonnegative real numbers because is positive semidefinite) and are unit eigenvectors corresponding to the eigenvalues
Note that, while there's no freedom in choosing the eigenvalues (except for how they're ordered), there is freedom in the choice of the eigenvectors, particularly when there are eigenvalues with multiplicity larger than one. So, this is not a unique expression of — we're just assuming we have one such expression. Regardless, because the eigenvalues are nonnegative real numbers, they have nonnegative square roots, and so we can select
for each to obtain an expression of the form
It is, however, not essential that the expression comes from a spectral decomposition in this way, and in particular the vectors need not be orthogonal in general. It is noteworthy, though, that we can choose these vectors to be orthogonal if we wish — and moreover we never need to be larger than (recalling that and denote the numbers of classical states of and respectively).
Next, each of the vectors can be further decomposed as
where the vectors have entries corresponding to the classical states of and can be explicitly determined by the equation
for each and Although are not necessarily unit vectors, this is the same process we would use to analyze what would happen if a standard basis measurement was performed on the system given a quantum state vector of the pair
And now we come to the trick that makes this part of the proof work. We define our Kraus matrices according to the following equation.
We can think about this formula purely symbolically: effectively gets flipped around to form and moved to right-hand side, forming a matrix. For the purposes of verifying the proof, the formula is all we need.
There is, however, a simple and intuitive relationship between the vector and the matrix which is that by vectorizing we get What it means to vectorize is that we stack the columns on top of one another (with the leftmost column on top proceeding to the rightmost on the bottom), in order to form a vector. For instance, if and are both qubits, and for some choice of we have
then
(Beware: sometimes the vectorization of a matrix is defined in a slightly different way, which is that the rows of the matrix are transposed and stacked on top of one another to form a column vector.)
First we'll verify that this choice of Kraus matrices correctly describes the mapping after which we'll verify the other required condition. To keep things straight, let's define a new mapping as follows.
Thus, our goal is to verify that
The way we can do this is to compare the Choi representations of these mappings. Choi representations are faithful, so we have if and only if At this point we can simply compute using the expressions
together with the bilinearity of tensor products to simplify.
So, our Kraus matrices correctly describe
It remains to check the required condition on which turns out to be equivalent to the assumption (which we haven't used yet). What we'll show is this relationship:
(in which we're referring the matrix transpose on the left-hand side).
Starting on the left, we can first observe that
The last equality follows from the fact that the transpose is linear and maps to
Moving to the right-hand side of our equation, we have
and therefore
We've obtained the same result, and therefore the equation has been verified. It follows, by the assumption that
and therefore, because the identity matrix is its own transpose, the required condition is true.
Kraus to Stinespring representations
Now suppose that we have a Kraus representation of a mapping
for which
Our goal is to find a Stinespring representation for
What we'd like to do first is to choose the garbage system so that its classical state set is For and to have the same size, however, it must be that divides allowing us to take to have classical states for
For an arbitrary choice of and it may not be the case that is an integer, so we're not actually free to choose so that it's classical state set is But we can always increase arbitrarily in the Kraus representation by choosing for however many additional values of that we wish.
And so, if we tacitly assume that is an integer, which is equivalent to being a multiple of then we're free to take so that its classical state set is In particular, if it is the case that then we may take to have classical states.
It remains to choose and we'll do this by matching the following pattern.
To be clear, this pattern is meant to suggest a block matrix, where each block (including as well as the blocks marked with a question mark) has rows and columns. There are rows of blocks, which means that there are columns of blocks.
Expressed in more formulaic terms, we will define as
where each matrix has rows and columns, and in particular we shall take for
This must be a unitary matrix, and the blocks labeled with a question mark, or equivalently for must be selected with this in mind — but aside from allowing to be unitary, the blocks labeled with a question mark won't have any relevance to the proof.
Let's momentarily disregard the concern that is unitary and focus on the expression
that describes the output state of given the input state of for our Stinespring representation. We can alternatively write
and we see from our choice of that
We therefore find that
and so
We therefore have a correct representation for the mapping and it remains to verify that we can choose to be unitary.
Consider the first columns of when it's selected according to the pattern above. Taking these columns alone, we have a block matrix
There are columns, one for each classical state of and as vectors let us name the columns as for each Here's a formula for these vectors that can be matched to the block matrix representation above.
Now let's compute the inner product between any two of these vectors, meaning the ones corresponding to any choice of
By the assumption
we conclude that the column vectors form an orthonormal set:
for all
This implies that it is possible to fill out the remaining columns of so that it becomes a unitary matrix. In particular, the Gram-Schmidt orthogonalization process can be used to select the remaining columns. Something similar was done in the Quantum circuits lesson of "Basics of quantum information" in the context of the state discrimination problem.
Stinespring representations back to the definition
The final implication is 4 1. That is, we assume that we have a unitary operation transforming a pair of systems into a pair and our goal is to conclude that the mapping
is a valid channel. From its form, it is evident that is linear, and it remains to verify that it always transforms density matrices into density matrices. This is pretty straightforward and we've already discussed the key points.
In particular, if we start with a density matrix of a compound system and then add on an additional workspace system we will certainly be left with a density matrix. If we reorder the systems for convenience, we can write this state as
We then apply the unitary operation and as we already discussed this is a valid channel, and hence maps density matrices to density matrices. Finally, the partial trace of a density matrix is another density matrix.
Another way to say this is to observe first that each of these things is a valid channel:
- Introducing an initialized workspace system.
- Performing a unitary operation.
- Tracing out a system.
And finally, any composition of channels is another channel — which is immediate from the definition, but is also a fact worth observing in its own right.
This completes the proof of the final implication, and therefore we've established the equivalence of the four statements listed at the start of the section.