Quantum state discrimination and tomography
Halaman ini belum diterjemahkan. Anda sedang melihat versi asli dalam bahasa Inggris.
In the last part of the lesson, we'll briefly consider two tasks associated with measurements: quantum state discrimination and quantum state tomography.
-
Quantum state discrimination
For quantum state discrimination, we have a known collection of quantum states along with probabilities associated with these states. A succinct way of expressing this is to say that we have an ensemble
of quantum states.
A number is chosen randomly according to the probabilities and the system is prepared in the state The goal is to determine, by means of a measurement of alone, which value of was chosen.
Thus, we have a finite number of alternatives, along with a prior — which is our knowledge of the probability for each to be selected — and the goal is to determine which alternative actually happened. This may be easy for some choices of states and probabilities, and for others it may not be possible without some chance of making an error.
-
Quantum state tomography
For quantum state tomography, we have an unknown quantum state of a system — so unlike in quantum state discrimination there's typically no prior or any information about possible alternatives.
This time, however, it's not a single copy of the state that's made available, but rather many independent copies are made available. That is, identical systems are each independently prepared in the state for some (possibly large) number The goal is to find an approximation of the unknown state, as a density matrix, by measuring the systems.
Discriminating between two states
The simplest case for quantum state discrimination is that there are two states, and that are to be discriminated.
Imagine a situation in which a bit is chosen randomly: with probability and with probability A system is prepared in the state meaning or depending on the value of and given to us. Our goal is to correctly guess the value of by means of a measurement on To be precise, we shall aim to maximize the probability that our guess is correct.
An optimal measurement
An optimal way to solve this problem begins with a spectral decomposition of a weighted difference between and where the weights are the corresponding probabilities.
Notice that we have a minus sign rather than a plus sign in this expression: this is a weighted difference not a weighted sum.
We can maximize the probability of a correct guess by selecting a projective measurement as follows. First let's partition the elements of into two disjoint sets and depending upon whether the corresponding eigenvalue of the weighted difference is nonnegative or negative.
We can then choose a projective measurement as follows.
(It doesn't actually matter in which set or we include the values of for which Here we're choosing arbitrarily to include these values in )
This is an optimal measurement in the situation at hand that minimizes the probability of an incorrect determination of the selected state.
Correctness probability
Now we will determine the probability of correctness for the measurement
To begin we don't really need to be concerned with the specific choice we've made for and though it may be helpful to keep it in mind. For any measurement (not necessarily projective) we can write the correctness probability as follows.
Using the fact that is a measurement, so we can rewrite this expression as follows.
On the other hand, we could have made the substitution instead. That wouldn't change the value but it does give us an alternative expression.
The two expressions have the same value, so we can average them to give yet another expression for this value. (Averaging the two expressions is just a trick to simplify the resulting expression.)
Now we can see why it makes sense to choose the projections and (as specified above) for and respectively — because that's how we can make the trace in the final expression as large as possible. In particular,
So, when we take the trace, we obtain the sum of the absolute values of the eigenvalues — which is equal to what's known as the trace norm of the weighted difference.
Thus, the probability that the measurement leads to a correct discrimination of and given with probabilities and respectively, is as follows.
The fact that this is the optimal probability for a correct discrimination of and given with probabilities and is commonly referred to as the Helstrom–Holevo theorem (or sometimes just Helstrom's theorem).
Discriminating three or more states
For quantum state discrimination when there are three or more states, there is no known closed-form solution for an optimal measurement, although it is possible to formulate the problem as a semidefinite program — which allows for efficient numerical approximations of optimal measurements with the help of a computer.
It is also possible to verify (or falsify) optimality of a given measurement in a state discrimination task through a condition known as the Holevo-Yuen-Kennedy-Lax condition. In particular, for the state discrimination task defined by the ensemble
the measurement is optimal if and only if the matrix
is positive semidefinite for every
For example, consider the quantum state discrimination task in which one of the four tetrahedral states is selected uniformly at random. The tetrahedral measurement succeeds with probability
This is optimal by the Holevo-Yuen-Kennedy-Lax condition, as a calculation reveals that
for
Quantum state tomography
Finally, we'll briefly discuss the problem of quantum state tomography. For this problem, we're given a large number of independent copies of an unknown quantum state and the goal is to reconstruct an approximation of To be clear, this means that we wish to find a classical description of a density matrix that is as close as possible to
We can alternatively describe the set-up in the following way. An unknown density matrix is selected, and we're given access to quantum systems each of which has been independently prepared in the state Thus, the state of the compound system is
The goal is to perform measurements on the systems and, based on the outcomes of those measurements, to compute a density matrix that closely approximates This turns out to be a fascinating problem and there is ongoing research on it.
Different types of strategies for approaching the problem may be considered. For example, we can imagine a strategy where each of the systems is measured separately, in turn, producing a sequence of measurement outcomes. Different specific choices for which measurements are performed can be made, including adaptive and non-adaptive selections. In other words, the choice of what measurement is performed on a particular system might or might not depend on the outcomes of prior measurements. Based on the sequence of measurement outcomes, a guess for the state is derived — and again there are different methodologies for doing this.
An alternative approach is to perform a single joint measurement of the entire collection, where we think about as a single system and select a single measurement whose output is a guess for the state This can lead to an improved estimate over what is possible for separate measurements of the individual systems, although a joint measurement on all of the systems together is likely to be much more difficult to implement.
Qubit tomography using Pauli measurements
We'll now consider quantum state tomography in the simple case where is a qubit density matrix. We assume that we're given qubits that are each independently in the state and our goal is to compute an approximation that is close to
Our strategy will be to divide the qubits into three roughly equal-size collections, one for each of the three Pauli matrices and Each qubit is then measured independently as follows.
-
For each of the qubits in the collection associated with we perform a measurement. This means that the qubit is measured with respect to the basis which is an orthonormal basis of eigenvectors of and the corresponding measurement outcomes are the eigenvalues associated with the two eigenvectors: for the state and for the state By averaging together the outcomes over all of the states in the collection associated with we obtain an approximation of the expectation value