Classical information
To describe quantum information and how it works, we will begin with an overview of classical information. It is natural to wonder why so much attention is paid to classical information in a course on quantum information, but there are good reasons.
For one, although quantum and classical information are different in some spectacular ways, their mathematical descriptions are actually quite similar. Classical information also serves as a familiar point of reference when studying quantum information, as well as a source of analogy that goes a surprisingly long way. It is common that people ask questions about quantum information that have natural classical analogs, and often those questions have simple answers that can provide both clarity and insight into the original questions about quantum information. Indeed, it is not at all unreasonable to claim that one cannot truly understand quantum information without understanding classical information.
Some readers may already be familiar with the material to be discussed in this section, while others may not — but the discussion is meant for both audiences. In addition to highlighting the aspects of classical information that are most relevant to an introduction to quantum information, this section introduces the Dirac notation, which is often used to describe vectors and matrices in quantum information and computation. As it turns out, the Dirac notation is not specific to quantum information; it can equally well be used in the context of classical information, as well as for many other settings in which vectors and matrices arise.
Classical states and probability vectors
Suppose that we have a system that stores information. More specifically, we shall assume that this system can be in one of a finite number of classical states at each instant. Here, the term classical state should be understood in intuitive terms, as a configuration that can be recognized and described unambiguously.
The archetypal example, which we will come back to repeatedly, is that of a bit, which is a system whose classical states are and Other examples include a standard six-sided die, whose classical states are and (represented by the corresponding number of dots on whatever face is on top); a nucleobase in a strand of DNA, whose classical states are A, C, G, and T; and a switch on an electric fan, whose classical states are (commonly) high, medium, low, and off. In mathematical terms, the specification of the classical states of a system are, in fact, the starting point: we define a bit to be a system that has classical states and and likewise for systems having different classical state sets.
For the sake of this discussion, let us give the name to the system being considered, and let us use the symbol to refer to the set of classical states of In addition to the assumption that is finite, which was already mentioned, we naturally assume that is nonempty — for it is nonsensical for a physical system to have no states at all. And while it does make sense to consider physical systems having infinitely many classical states, we will disregard this possibility, which is certainly interesting but is not relevant to this course. For these reasons, and for the sake of convenience and brevity, we will hereafter use the term classical state set to mean any finite and nonempty set.
Here are a few examples:
- If is a bit, then In words, we refer to this set as the binary alphabet.
- If is a six-sided die, then
- If is an electric fan switch, then
When thinking about as a carrier of information, the different classical states of could be assigned certain meanings, leading to different outcomes or consequences. In such cases, it may be sufficient to describe as simply being in one of its possible classical states. For instance, if is a fan switch, we might happen to know with certainty that it is set to high, which might then lead us to switch it to medium.
Often in information processing, however, our knowledge is uncertain. One way to represent our knowledge of the classical state of a system is to associate probabilities with its different possible classical states, resulting in what we shall call a probabilistic state.
For example, suppose is a bit. Based on what we know or expect about what has happened to in the past, we might perhaps believe that is in the classical state with probability and in the state with probability We may represent these beliefs by writing this:
A more succinct way to represent this probabilistic state is by a column vector.
The probability of the bit being is placed at the top of the vector and the probability of the bit being is placed at the bottom, because this is the conventional way to order the set
In general, we can represent a probabilistic state of a system having any classical state set in the same way, as a vector of probabilities. The probabilities can be ordered in any way we choose, but it is typical that there is a natural or default way to do this. To be precise, we can represent any probabilistic state through a column vector satisfying two properties:
- All entries of the vector are nonnegative real numbers.
- The sum of the entries is equal to
Conversely, any column vector that satisfies these two properties can be taken as a representation of a probabilistic state. Hereafter, we will refer to vectors of this form as probability vectors.
Alongside the succinctness of this notation, identifying probabilistic states as column vectors has the advantage that operations on probabilistic states are represented through matrix–vector multiplication, as will be discussed shortly.
Measuring probabilistic states
Next let us consider what happens if we measure a system when it is in a probabilistic state. In this context, by measuring a system we simply mean that we look at the system and recognize whatever classical state it is in without ambiguity. Intuitively speaking, we can't "see" a probabilistic state of a system; when we look at it, we just see one of the possible classical states.
By measuring a system, we may also change our knowledge of it, and therefore the probabilistic state we associate with it can change. That is, if we recognize that is in the classical state then the new probability vector representing our knowledge of the state of becomes the vector having a in the entry corresponding to and for all other entries. This vector indicates that is in the classical state with certainty — which we know having just recognized it — and we denote this vector by which is read as "ket " for a reason that will be explained shortly. Vectors of this sort are also called standard basis vectors.
For example, assuming that the system we have in mind is a bit, the standard basis vectors are given by
Notice that any two-dimensional column vector can be expressed as a linear combination of these two vectors. For example,