Probability theory
Classical probability
In a classical probabilistic space you have events and a probability measure:
But you want numerical data, so you study random variables: functions
such that
for every open set
They constitute the algebra
plays the role of a linear functional
In fact, it is not only an algebra, but a commutative von Neumann algebra, and
There is a theorem called the Gelfand-Naimark theorem which applied to a von Neumann algebra with a normal, faithful state let us recover the original probabilistic space. That is to say: all the data is inside the von Neumann algebra and the linear functional.
Sketch of the construction: if we begin with an algebra
Quantum probability
I have written two times about quantum probability. I have to re-read and merge both text (I don't know which is better)
Text 1:
Coming from above. And now, here comes the key idea: making a mental effort you can see any
simply defined like the diagonal matrix whose entries are the values of
This way we distinguish a subset
Let's try with an example. Consider
such that
But we can think in a more active visualization of all this stuff. Imagine that
I can express a player with three numbers
What is the role of our function
- The function
applied to a player by means of returns a new player with improved (or worsened) features. - A state
(a class of players) acts as a functional on the algebra and is the average improvement of the treatment over the player of type .
If several treatments changes features in a isolated way, like the previous one, you can apply them to the player in the order you want.That is, they commute. But we can admit other treatments which act in a more complicated way, i. e., no diagonal matrices than may not commute. Our
That is, we are identifying
such that
But, in fact,
And the evaluation process
corresponds to
This operation has a name, is the Frobenius inner product, related to the trace:
Even
being
Moreover, the probability of an event, for example
Matrices such that
For example, consider the matrix
We can try to evaluate our
and we obtain 2'5. Something between
Another key idea to develop: as subspaces and even as linear maps, events and states are the same. I compute the probability of an event given a state by projecting the state vector over the event subspace: the squared length is the probability and the resulting vector is the new state, or by means of the trace if I consider them operators. They should be treated on the same footing.
Text 2:
With classical probability
We are going to analyse, from the beginning, the Stern-Gerlach experiment from a mathematical viewpoint, and try to see why is natural the formulation of QM. Imagine that electrons have an internal configuration that can be observed with a Stern-Gerlach device (
From the point of view of set theory, improved with basic probability theory, our first thought is: "ok, I have a set of electrons
Then, we observe that if we have a copy of this machine and rotate it (or keep it fixed and rotate the gun that shoot the electrons in the opposite direction) to the
with a new probability function. Our new set appears like a Cartesian product of the previous "sets of possibilities".
Observe that the joint distribution is not necessarily
being this the case only when the variables are independent. For example, maybe there is a correlation between being red and being big.
A different approach
All of this could have been translated into math in a very different way, far more complicated, but that it will pay off later.
A finite set
with inner product
This way we have "enriched" the set:
- we conserve the original elements of the set, codified in the rays through the canonical basis; but we get new objects, the superposition of the elements. This new objects may not have an interpretation for us, at a first glance.
- And we also have a measure of "how independent" this objects are: the inner product. For example, the original elements are totally independent, since their inner product is 0.
Let's come back to Stern-Gerlach. Instead of thinking as before in the set
with the usual inner product.
The canonical basis elements of
represents the probability of obtaining the output
So far, two questions can arise:
-
We have chosen in
the usual inner product (equivalent to ). Why not other inner product or even why not a simpler mathematical object like an absolute value norm? That is, why don't we take, for example, and encode probabilities in without the square? Well, the only -norm satisfying the parallelogram rule is , and this rule is needed to form an inner product from the norm. The finer approach of inner product will be needed later because it let us use further mathematics concepts (orthogonality, unitary Lie groups,...). Moreover, even intuitively the inner product approach is desirable because it gives us a measure of the independence of the elements represented by the rays in the Hilbert space. -
Why don't we use real numbers and take as probabilities the absolute values of the component instead of the square? Once we fix the use of the 2-norm, we need to deal with amplitudes (i.e., numbers whose square give probabilities) because we want to keep with us the addition of probabilities for incompatible events, from the classical setup.
-
And also, Why complex numbers in QM?
Within this approach to sets as Hilbert spaces, subsets are encoded as subspaces, the union of sets is translated as the direct sum of subspaces, intersection of sets as intersections of subspaces and the complement of a set as the orthogonal complement of the subspace.
Random variables vs operators
Let's continue with
such that
i.e.,
The random variable registers the idea of a measurement. This new approach to them may look very artificial, but it retains the same information that the classical probability approach:
- We recuperate the values, for example for
, with
- The expected value of
, provided that probabilities are given by the vector , would be
When we take our second machine
If we assume, as before, a classical behavior of the states, we can model this with a new Hilbert space
whose basis will be denoted by
We can think that the state describing the system would be
but in fact this is only an special case when the two machines are yielding the equivalent to "independent variables" (the joint probability is the product of the probabilities). In general, it is valid any
are such that
where the second and the fourth equations could be deduced, so we can remove it.
The random variable represented by the operator