Quantum Entanglement
It’s a popular myth that identical twins, it's said, can
sometimes sense when one of the pair is in danger, even if they're oceans apart.
Tales of telepathy abound. Scientists cast a skeptical eye over such claims,
largely because it isn't clear how these weird connections could possibly work.
Yet they've had to come to terms with something that's no less strange in the
world of physics: an instantaneous link between particles that remains strong,
secure, and undiluted no matter how far apart the particles may be – even if
they're on opposite sides of the universe. It's a link that Einstein went to
his grave denying, yet its existence is now beyond dispute. This quantum
equivalent of telepathy is demonstrated daily in laboratories around the world.
It holds the key to future hyperspeed computing and underpins the science of
teleportation. Its name is entanglement.
The discovery of entanglement
The concept, but not the name of entanglement, was first put
under the scientific spotlight on May 15, 1935, when a paper by Einstein and
two younger associates, Boris Podolosky and Nathen Rosen, appeared in the
journal Physical Review.[1]
Its title – "Can a Quantum-Mechanical Description of Physical Reality Be Considered Complete?" – leaves no doubt that the paper was a challenged to Niels Bohr and his vision of the subatomic world. On June 7, Erwin Schrödinger, himself no lover of quantum weirdness, wrote to Einstein, congratulating him on the paper and using in his letter the word entanglement – or, rather, its German equivalent verschränkung – for the first time. This new term soon found its way into print in an article – sent to the Cambridge Philosophical Society on August 14 that was published a couple of months later.[2]
In it he wrote:
Its title – "Can a Quantum-Mechanical Description of Physical Reality Be Considered Complete?" – leaves no doubt that the paper was a challenged to Niels Bohr and his vision of the subatomic world. On June 7, Erwin Schrödinger, himself no lover of quantum weirdness, wrote to Einstein, congratulating him on the paper and using in his letter the word entanglement – or, rather, its German equivalent verschränkung – for the first time. This new term soon found its way into print in an article – sent to the Cambridge Philosophical Society on August 14 that was published a couple of months later.[2]
In it he wrote:
When two systems ... enter into temporary physical
interaction ... and when after a time of mutual influence the systems separate
again, then they can no longer be described in the same way as before, viz. by
endowing each of them with a representative of its own. I would not call that
one but rather the characteristic trait of quantum mechanics, the one that
enforces its entire departure from classical lines of thought. By the
interaction the two representatives [the quantum states] have become entangled.
The characteristic trait of quantum mechanics ... the one
that enforces its entire departure from classical lines of thought – here was
an early sign of the importance attached to this remarkable effect.
Entanglement lay at the very heart of quantum reality – its most startling and defining feature and Einstein would have none of it.
Entanglement lay at the very heart of quantum reality – its most startling and defining feature and Einstein would have none of it.
For the best part of a decade, the man who revealed the
particle nature of light (see Einstein and the photoelectric effect) had been
trying to undermine Bohr's interpretation of quantum theory. Einstein couldn't
stomach the notion that particles didn't have properties, such as momentum and
position, with real, determinable (if only we knew how), preexisting values.
Yet that notion was spelled out in a relationship discovered in 1927 by Werner
Heisenberg.
Known as the uncertainty principle, it stems from the rule that the
result of multiplying together two matrices representing certain pairs of
quantum properties, such as position and momentum, depends on the order of
multiplication. The same oddball math that says X times Y doesn't have to equal
Y times X implies that we can never know simultaneously the exact values of
both position and momentum. Heisenberg proved that the uncertainty in momentum
can never be smaller than a particular number that involves Planck's constant.
In one sense, this relationship quantifies wave-particle duality. Momentum is a
property that waves can have (related to their wavelength); position is a
particlelike property because it refers to a localization in space.
Heisenberg's formula reveals the extent to which one of these aspects fades out
as the other becomes the focus of attention.
In a different but related sense,
the uncertainty principle tells how much the complementary descriptions of a
quantum object overlap. Position and momentum are complimentary properties
because to pin down one is to lose track of the other; they coexist but are
mutually exclusive, like the opposite sides of the same object. Heisenberg's
formula quantifies the extent to which knowledge of one limits knowledge of the
other.(For More see my article on the life and work of Werner Heisenberg which has a more detailed description of the uncertainty principle and the context of wave particle duality)
Einstein didn't buy this. He believed that a particle does
have a definite position and momentum all the time, whether we're watching it
or not, despite what quantum theory says. From his point of view, the
Heisenberg uncertainty principle isn't a basic rule of nature; it's just an
artifact of our inadequate understanding of the subatomic realm. In the same
way, he though, wave-particle duality isn't grounded in reality but instead
arises from a statistical description of how large numbers of particles behave.
Given a better theory, there'd be no wave-particle duality or uncertainty
principle to worry about. The problem was, as Einstein saw it, that quantum
mechanics wasn't telling the whole story: it was incomplete.
Einstein versus Bohr
Intent on exposing this fact to the world and championing a
return to a more classical pragmatic view of nature, Einstein devised several
thought experiments between the late 1920s and the mid-1930s. Targeted
specifically at the idea of complementarity, these experiments were designed to
point out ways to simultaneously measure a particle's position and momentum, or
its precise energy at a precise time (another complementary pair), thus pulling
the rug from under the uncertainty principle and wave-particle duality.
The first of these experiments was talked about informally
in 1927, in hallway discussions at the fifth Solvay Conference in Brussels.
Einstein put to Bohr a modified version of the famous double-slit experiment in
which quantum objects – electrons, say – emerging from the twin slits are
observed by shining light onto them. Coherent Photons bouncing off a particle would have
their momenta changed by an amount that would reveal the particle's trajectory
and, therefore, which slit it had passed through. The particle would then go on
to strike the detector screen and contribute to the buildup of an interference
pattern. Wave-particle duality would be circumvented, Einstein argued, because
we would have simultaneously measured particlelike behavior (the trajectory the
particle took) and wavelike behavior (the interference pattern on the screen).
But Bohr spotted something about this thought experiment
that Einstein had overlooked. To be able to tell which slit a particle went
through, you'd have to fix its position with an accuracy better than the
distance between the slits. Bohr then applied Heisenberg's uncertainty
principle, which demands that if you pin down the particle's position to such
and such a precision, you have to give up a corresponding amount of knowledge
of its momentum. Bohr said that this happens because the photons deliver random
kicks as they bounce off the particle. The result of these kicks is to inject
uncertainty into the whereabouts of the particle when it strikes the screen.
And here's the crucial caveat: the uncertainty turns out to be roughly as large as the
spacing between the interference bands. The pattern is smeared out and lost as the quantum mechanical wavefunction becomes decoherent. With this it disappears Einstein's hoped-for contradiction.
On several other occasions, Einstein confronted Bohr with
thought experiments cunningly contrived to blow duality out of the water. Each
time, Bohr used the uncertainty principle to exploit a loophole and win the say
against his arch rival (and, incidentally, good friend). In the battle for the
future of quantum physics, Bohr defeated Einstein and, in the process, showed
just how important Heisenberg's little formula was in the quantum scheme of
things.
These arguments between Bohr and Einstein were never truly resolved and got evermore technical. At the sixth Congress of Solvay, in 1930, the indeterminacy relation was Einstein's target of criticism. His idea contemplates the existence of an experimental apparatus which was subsequently designed by Bohr in such a way as to emphasize the essential elements and the key points which he would use in his response.
In this Einstein considers a box, sometime's called "Einstein's Box" or "Einstein's Box of Light". With this thought experiment, which was designed with Bohr's assistance, Einstein's was supposed to prove the violation of the indeterminacy relation between time and energy. The schematic of Einetin and Bohr's apparatus is shown below:
"Einstein's Box of Light" - Einstein's secret weapon to destroy quantum mechanics?
Einstein described a box full of light and said that it was possible to measure both the energy 'E' of a single photon and the time 't' when it was emitted. This was not allowed by a variant on Heisenberg’s uncertainty principle, namely .
Einstein said that the box could be weighed at first and then a single photon be allowed to escape through a shutter controlled by a clock inside the box. The box would then be weighed again and the mass difference 'm' determined. The energy of the photon 'E' is simply E = mc^2.
It appeared that both the photon's energy and its time of emission could be determined! This caused a bit of a shock when first seen by Bohr, he genuinely did not see the solution at once and Einstein seemed at first sight to have one the battle this time, meaning that the uncertainty of quantum mechanics was going to be finally wiped out!
Bohr, after sleeping on the problem, finally realized that there was a flaw in Einstein's reasoning. When the photon is released, the box will recoil (to conserve momentum) and the position of the box in the earth's gravitational field will be uncertain. Einstein's very own general theory of relativity said that this would cause a corresponding uncertainty in the time recorded.
An illustration of this, in context of Einstein's light box is shown on the left - critically it depends on the presence of a clock in the device recording the time at which particles are measured and weighted at precise time intervals by recoil from the light source; clocks are effected by gravity in special relativity and run slower in a gravitational field than in zero gravity. Hence, gravitation affects the measurement, and thus induces uncertainties in the weight of quanta - as quantum mechanics predicts. *
*(although this may not be a successful tool for disproving quantum mechanics, as Einstein intended, this may be an important way however to measure certain weakly interacting affects of gravity in context of the predictions of general relativity -such as local gravitational waves and instances of spatial warping.)
So Bohr had been saved by defeat by Einstein forgetting his own theory of general relativity!
This was to be the last serious assault – approximately 28 years after its inception at the hands of Planck, the foundations of quantum mechanics seemed to be complete and depended wholly on this uncertainty business which Einstein never accepted and always saw it as a kind of a magician's curtain, hiding the true mechanism of what appears to be a trick of nature itself.
These arguments between Bohr and Einstein were never truly resolved and got evermore technical. At the sixth Congress of Solvay, in 1930, the indeterminacy relation was Einstein's target of criticism. His idea contemplates the existence of an experimental apparatus which was subsequently designed by Bohr in such a way as to emphasize the essential elements and the key points which he would use in his response.
In this Einstein considers a box, sometime's called "Einstein's Box" or "Einstein's Box of Light". With this thought experiment, which was designed with Bohr's assistance, Einstein's was supposed to prove the violation of the indeterminacy relation between time and energy. The schematic of Einetin and Bohr's apparatus is shown below:
"Einstein's Box of Light" - Einstein's secret weapon to destroy quantum mechanics?
Einstein described a box full of light and said that it was possible to measure both the energy 'E' of a single photon and the time 't' when it was emitted. This was not allowed by a variant on Heisenberg’s uncertainty principle, namely .
Einstein said that the box could be weighed at first and then a single photon be allowed to escape through a shutter controlled by a clock inside the box. The box would then be weighed again and the mass difference 'm' determined. The energy of the photon 'E' is simply E = mc^2.
It appeared that both the photon's energy and its time of emission could be determined! This caused a bit of a shock when first seen by Bohr, he genuinely did not see the solution at once and Einstein seemed at first sight to have one the battle this time, meaning that the uncertainty of quantum mechanics was going to be finally wiped out!
Bohr, after sleeping on the problem, finally realized that there was a flaw in Einstein's reasoning. When the photon is released, the box will recoil (to conserve momentum) and the position of the box in the earth's gravitational field will be uncertain. Einstein's very own general theory of relativity said that this would cause a corresponding uncertainty in the time recorded.
An illustration of this, in context of Einstein's light box is shown on the left - critically it depends on the presence of a clock in the device recording the time at which particles are measured and weighted at precise time intervals by recoil from the light source; clocks are effected by gravity in special relativity and run slower in a gravitational field than in zero gravity. Hence, gravitation affects the measurement, and thus induces uncertainties in the weight of quanta - as quantum mechanics predicts. *
*(although this may not be a successful tool for disproving quantum mechanics, as Einstein intended, this may be an important way however to measure certain weakly interacting affects of gravity in context of the predictions of general relativity -such as local gravitational waves and instances of spatial warping.)
So Bohr had been saved by defeat by Einstein forgetting his own theory of general relativity!
This was to be the last serious assault – approximately 28 years after its inception at the hands of Planck, the foundations of quantum mechanics seemed to be complete and depended wholly on this uncertainty business which Einstein never accepted and always saw it as a kind of a magician's curtain, hiding the true mechanism of what appears to be a trick of nature itself.
Such is the version of this clash of 20th-century titans
that's been dutifully repeated in textbooks and spoon fed to physics students
for many years. But evidence has recently come to light that Bohr had
unwittingly hoodwinked Einstein with arguments that were fundamentally unsound.
This disclosure doesn't throw quantum mechanics back into the melting pot, but
it does mean that the record needs setting straight, and that the effect that
really invalidates Einstein's position should be given proper credit.
The revisionist picture of the Bohr-Einstein debates stems
partly from a suggestion made in 1991 by Marlan Scully, Berthold-Georg Englert,
and Herbert Walther of the Max Planck Institute for Quantum Optics in Garching,
Germany.[3] These researchers proposed using atoms as quantum objects in a
version of Young's two-slit experiment.
Atoms have an important advantage over simpler particles, such as photons or electrons: they have a variety of internal states, including a coherent ground state (lowest energy state) and a series of decoherent excited states. These different states, the German team reckoned, could be used to track the atom's path.
Atoms have an important advantage over simpler particles, such as photons or electrons: they have a variety of internal states, including a coherent ground state (lowest energy state) and a series of decoherent excited states. These different states, the German team reckoned, could be used to track the atom's path.
This two-state example of coherence and decoherence is what allowed this formulation of a quantum version of the famous Two-Slit Experiment.
We can label the probability-amplitude wave function passing through the left hand slit in the figure ψleft and the waves passing through the right-hand slit ψright. These are coherent and show the characteristic quantum interference fringes on the detector screen (a photographic plate or CCD array). This is the case even if the intensity of particles is so low that only one particle at a time arrives at the screen.
In a dramatic experimental proof of decoherence, physicist Gerhard Rempe sent matter waves of heavy Rubidium atoms through two slits. He then irradiated the left slit with microwaves that could excite the hyperfine structure in Rb atoms passing through that slit. As he turned up the intensity, the interference fringes diminished in proportion to the number of photons falling on the left slit. The photons decohere the otherwise coherent wave functions.[4]
The crucial factor in this version of the double-slit
experiment is that the microwaves have hardly any momentum of their own, so
they can cause virtually no change to the atom's momentum – nowhere near enough
to smear out the interference pattern.
Heisenberg's uncertainty principle can't possibly play a significant hand in the outcome. Yet with the microwaves turned on so that we can tell which way the atoms went, the interference pattern suddenly vanishes. Bohr had argued that when such a pattern is lost, it happens because a measuring device gives random kicks to the particles. But there aren't any random kicks to speak of in the rubidium atom experiment; at most, the microwaves deliver momentum taps ten thousand times too small to destroy the interference bands. Yet, destroyed the bands are. It isn't that the uncertainty principle is proved wrong, but there's no way it can account for the results.
Heisenberg's uncertainty principle can't possibly play a significant hand in the outcome. Yet with the microwaves turned on so that we can tell which way the atoms went, the interference pattern suddenly vanishes. Bohr had argued that when such a pattern is lost, it happens because a measuring device gives random kicks to the particles. But there aren't any random kicks to speak of in the rubidium atom experiment; at most, the microwaves deliver momentum taps ten thousand times too small to destroy the interference bands. Yet, destroyed the bands are. It isn't that the uncertainty principle is proved wrong, but there's no way it can account for the results.
The only reason momentum kicks seemed to explain the classic
double slit experiment discussed by Bohr and Einstein turns out to be a
fortunate conspiracy of numbers. There's a mechanism at work far deeper than
random jolts and uncertainty. What destroys the interference pattern is the
very act of trying to get information about which paths is followed. The effect
at work is entanglement.
Nonlocality
Ordinarily, we think of separate objects as being
independent of one another. They live on their own terms, and anything tying
them together has to be forged by some tangible particles, A and B, which have
come into contact, interacted for a brief while, and then flown apart. Each
particle is described by (among other properties) its own position and
momentum. The uncertainty principle insists that one of these can't be measured
precisely without destroying knowledge of the other. However, because A and B
have interacted and, in the eyes of quantum physics, have effectively merged to
become one interconnected system, it turns out that the momentum of both
particles taken together and the distance between them can be measured as
precisely as we like. Suppose we measure the momentum of A, which we'll assume
has remained behind in the lab where we can keep an eye on it. We can then
immediately deduce the momentum of B without having to do any measurement on it
at all. Alternatively, if we choose to observe the position of A, we would
know, again without having to measure it, the position of B. This is true
whether B is in the same room or a great distance away.
From Heisenberg's relationship, we know that measuring the
position of, say, A will lead to an uncertainty in its momentum, Einstein,
Podolosky, and Rosen pointed out, however, that by measuring the position of A,
we gain precise knowledge of the position of B. Therefore, if we take quantum
mechanics at face value, by gaining precise knowledge of its position, an
uncertainty in momentum has been introduced for B. In other words, the state of
B depends on what we choose to do with A in our lab. And, again, this is true
whatever the separation distance may be. EPR considered such a result patently
absurd. How could B possibly know whether it should have a precisely defined
position or momentum? The fact that quantum mechanics led to such an
unreasonable conclusion, they argued, showed that it was flawed – or, at best,
that it was only a halfway house toward some more complete theory.
At the core of EPR's challenge is the notion of locality:
the common sense idea that things can only be affected directly if they're
nearby. To change something that's far away, there's a simple choice: you can
either go there yourself or send some kind of signal. Either way, information
or energy has to pass through the intervening space to the remote site in order
to affect it. The fastest this can happen, according to Einstein's special
theory of relativity, is the speed of light.
The trouble with entanglement is that it seems to ride
roughshod over this important principle. It's fundamentally nonlocal. A
measurement of particle A affects its entangled partner B instantaneously,
whatever the separation distance, and without signal or influence passing
between the two locations. This bizarre quantum connection isn't mediated by
fields of force, like gravity or electromagnetism. It doesn't weaken as the
particles move apart, because it doesn't actually stretch across space. As far
as entanglement is concerned, it's as if the particles were right next to one
another: the effect is as potent at a million light-years as it is at a
millimeter. And because the link operates outside space, it also operates
outside time. What happens at A is immediately known at B. No wonder Einstein
used words such as "spook" and "telepathic" to describe –
and deride – it. No wonder that as the author of relativity he argued that the
tie that binds entangled particles is a physical absurdity. Any claim that an
effect could work at faster-than-light speeds, that it could somehow serve to
connect otherwise causally isolated objects, was to Einstein an intellectual
outrage.
A close look at the EPR scenario reveals that it doesn't
actually violate causality, because no information passes between the entangled
particles. The information is already, as it were, built into the combined
system, and no measurement can add to it. But entanglement certainly does throw
locality out the window, and that development is powerfully counterintuitive.
It was far too much for Einstein and his colleagues to accept, and they were
firmly convinced that quantum mechanics, as it stood, couldn't be the final
word. It was, they suggested, a mere approximation of some as yet undiscovered
description of nature. This description would involve variables that contain
missing information about a system that quantum mechanics doesn't reveal, and
that tell particles how to behave before a measurement is carried out. A theory
along these lines – a theory of so-called local hidden variables – would
restore determinism and mark a return to the principle of locality.
The shock waves from the EPR paper quickly reached the
shores of Europe. In Copenhagen, Bohr was once again cast into a fever of excitement
and concern as he always was by Einstein's attacks on his beloved quantum
worldview. He suspended all other work in order to prepare a counterstrike.
Three months later, Bohr's rebuttal was published in the same American journal
that had run the EPR paper. Basically, it argued that the nonlocality objection
to the standard interpretation of quantum theory didn't represent a practical
challenge. It wasn't yet possible to test it, and so physicists should just get
on with using the mathematics of the subject, which worked so well, and not
fret about the more obscure implications.
David Bohm's View on The EPR Paradox
Most scientists, whose interest was simply in using quantum
tools to probe the structure of atoms and molecules were happy to follow Bohr's
advice. But a few theorists continued to dig away at the philosophical roots.
In 1952, American Physicist David Bohm, at Birkbeck College, London, who had been
hounded out of his homeland during the McCarthy "Red Scare" inquisitions, came up with a
variation on the EPR experiment that paved the way for further progress in the
matter.[5] Instead of using two properties, position and momentum, as in the
original version, Bohm focused on just one: the property known as spin.
The spin of subatomic particles, such as electrons, is
analogous to spin in the everyday world but with a few important differences.
Crudely speaking, an electron can be thought of as spinning around the way a
basketball does on top of an athlete's finger. But whereas spinning basketballs
eventually slow down, all electrons in the universe, whatever their
circumstances, spin all the time and at exactly the same rate. What's more,
they can only spin in one of two directions, clockwise or counterclockwise,
referred to as spin-up and spin-down.
Bohm's revised EPR thought experiment starts with the
creation, in a single event, of two particles with opposite spin. This means
that if we measure particle A and find that its spin-up, then, from that point
on, B must be spin-down. The only other possible result is that A is measured
to be spin-up, which forces B to be spin-down. Taking this second case as an
example, we're not to infer, says quantum mechanics, that A was spin-up before
we measured it and therefore that B was spin-down, in a manner similar to a
coin being heads or tails. Quantum interactions always produce superpositions.
The state of each particle in Bohm's revised EPR scenario is a mixed
superposition that we can write as: psi = (A spin-up and B spin-down) + (A
spin-down + B spin-up). A measurement to determine A's spin causes this wave
function to collapse and a random choice to be made of spin-up or spin-down. At
that very same moment, B also ceases to be in a superposition of states and
assumes the opposite spin.
This is the standard quantum mechanical view of the
situation and it leads to the same kind of weird conclusion that troubled
Einstein and friends. No matter how widely separated the spinning pair of
particles may be, measuring the spin of one causes the wave function of the
combined system to collapse instantaneously so that the unmeasured twin assumes
a definite (opposite) spin state, too. The mixed superposition of states, which
is the hallmark of entanglement, ensures nonlocality. Set against this is the
Einsteinian view that "spooky action at a distance" stems not from
limitations about what the universe is able to tell us but instead from
limitations in our current knowledge of science. At a deeper, more basic level
than that of wave functions and complementary properties, are hidden variables
that will restore determinism and locality to physics.
John Bell's inequality
Bohm's new version of the EPR paradox didn't in itself offer
a way to test these radically different worldviews, but it set the scene for
another conceptual breakthrough that did eventually lead to a practical
experiment. This breakthrough came in 1964 from Irish physicist,
John S. Bell, who worked at CERN, the European center for high-energy particle
research in Switzerland. Colleagues considered Bell to be the only physicist of
his generation to rank with the pioneers of quantum mechanics, such as Niels
Bohr and Max Born, in the depth of his philosophical understanding of the
implications of the theory. What Bell found is that it makes an experimentally
observable difference whether the particles described in the EPR experiment
have definite properties before measurement, or whether they're entangled in a
ghostlike hybrid reality that transcends normal ideas of space and time.
Bell's test hinges on the fact that a particle's spin can be
measured independently in three directions, conventionally called x, y, and z,
at right angles to one another. If you measure the spin of particle A along the
x direction, for example, this measurement also affects the spin of entangled
particle B in the x direction, but not in the y and z directions. In the same
way, you can measure the spin of B in, say, the y direction without affecting
A's spin along x or z. Because of these independent readings, it's possible to
build up a picture of the complementary spin states of both particles. Being a
statistical effect, lots of measurements are needed in order to reach a
definite conclusion. What Bell showed is that measurements of the spin states
in the x, y, and z directions on large numbers of real particles could in principle
distinguish between the local hidden variable hypothesis championed by the
Einstein-Bohm camp and the standard nonlocal interpretation of quantum
mechanics.
If Einstein was right and particles really did always have a
predetermined spin, then, said Bell, a Bohm-type EPR experiment ought to
produce a certain result. If the experiment were carried out on many pairs of
particles, the number of pairs of particles in which both are measured to be
spin-up, in both the x and y directions ("xy up"), is always less
than the combined total of measurements showing xz up and yz up. This statement
became known as Bell's inequality. Standard quantum theory, on the other hand,
in which entanglement and nonlocality are facts of life, would be upheld if the
inequality worked the other way around. The decisive factor is the degree of
correlation between the particles, which is significantly higher if quantum
mechanics rules.
This was big news. Bell's inequality, although the subject
of a modest little paper and hardly a poplar rival to the first Beatles tour of
America going on at the same time, provided a way to tell by actual experiment
which of the two major, opposing visions of subatomic reality was closer to the
truth.[6] Bell made no bones about what his analysis revealed: Einstein;s ideas
about locality and determinism were incompatible with the predictions of
orthodox quantum mechanics. Bell's paper offered a clear alternative that lay
between the EPR/Bohemian local hidden variables viewpoint and Bohrian, nonlocal
weirdness. The way that Bell's inequality was set up, its violation would mean
that the universe was inherently nonlocal*, allowing particles to form and
maintain mysterious connections with each other no matter how far apart they
were. All that was needed now was for someone to come along and set up an
experiment to see for whom Bell's inequality tolled.
But that was easier said than done. Creating, maintaining,
and measuring individual entangled particles is a delicate craft, and any
imperfection in the laboratory setup masks the subtle statistical correlations
being sought. Several attempts were made in the 1970s to measure Bell's
inequality but none was completely successful. Then a young French graduate
student, Alain Aspect, at the Institute of Optics in Orsay, took up the
challenge for his doctoral research.
* If we want to study the applicability of Classical Probability Theory, i.e. Bayes theorem, to all of quantum mechanics, beyond the Copenhagen interpretation, we have to recognize that quantum entanglement states, such as basic singlet states, are nonfactorizable and therefore will not follow the simple factorizable relations used in proving the consistency of Bayes theorem with quantum probability functions.
This is again a result of the paradox that entanglement is a non-local effect and therefore there isn't any possibility of decomposing/factorizing the density of
states of such an entangled quantum state locally.
Even by applying a locality condition to the Bell inequalities, in the stochastic Clauser-Horne Model say, it can be shown that this (local) model, as far as applied to the singlet-state and without using quantum mechanical formalism, is not completely stochastic (i.e. there are possible configurations for which the model is deterministic). However as soon as you apply quantum mechanical formalism it becomes non-local again.
Even in experiment, where the so-called Clauser-Horne inequalities correspond to the fixed conditional probabilities of light polariser orientations for entangled photon ensembles which have not been identified show that unless they were exactly identical, the different conditional probability values of the photons themselves could not be factorized.
So the paradox is based on non-locality which is in direct contradiction to the pure local cause and effect framework of special relativity. Any theory combined with quantum mechanics becomes non-local, it has to whenever any form of quantum formalism is used. EPR is purely a paradox of relativity, not of quantum mechanics.
The first efforts to relate theory and thought experiment with actual experiment came from the pioneering work of Australian physicist John Clive Ward working with British physicist Maurice Pryce along with the work of one of the greatest experimental physicists of the 20th century, Chinese-American Physicist Chien-Shiung Wu.
Their work was on formulating and experimentally verifying the probability amplitude for quantum entanglement was the first attempt to develop a way to find such "spooky actions" in an EPR apparatus.
In a 1947 paper, published in Nature[7], Ward and Pryce were the first to calculate, and use, the probability amplitudes for the polarisation of two entangled photons moving in opposite directions.
For polarisations x and y, Ward derived this probability amplitude to be
which once normalised can be expressed as
where 1 and 2 refer to the two quanta propagating in different directions. Ward's probability amplitude is then applied to derive the correlation of the quantum polarisations of the two photons propagating in opposite directions.
This prediction was experimentally confirmed by Wu and Shaknov in 1950.[8] In current terminology this result corresponds to a pair of entangled photons and is directly relevant to a typical Einstein-Podolsky-Rosen (EPR) paradox
Chien-Shiung Wu – often referred to as Madame Wu or the First Lady of Physics – from the University of Columbia was first to give indirect evidence of entanglement in the laboratory.
She showed an Einstein-type correlation between the polarisation of two well-separated photons, which are tiny localised particles of light.
Chien-Shiung Wu at her Columbia University Physics Lab, 1963.
Madame Wu's work led to the confirmation of the Pryce and Ward calculations on the correlation of the quantum polarizations of two photons propagating in opposite directions. This was the first experimental confirmation of quantum results relevant to a pair of entangled photons as applicable to the Einstein-Podolsky-Rosen (EPR) paradox.
However, direct evidence of the EPR paradox, one which would include a complete isolation of local effects and test the effect of nonlocality of quantum phenomina would require a few more decades, until the laser was invented, which allowed the French physicist Alain Aspect to form an experiment that we would recognize today as a quantum entanglement circuit.
* If we want to study the applicability of Classical Probability Theory, i.e. Bayes theorem, to all of quantum mechanics, beyond the Copenhagen interpretation, we have to recognize that quantum entanglement states, such as basic singlet states, are nonfactorizable and therefore will not follow the simple factorizable relations used in proving the consistency of Bayes theorem with quantum probability functions.
This is again a result of the paradox that entanglement is a non-local effect and therefore there isn't any possibility of decomposing/factorizing the density of
states of such an entangled quantum state locally.
Even by applying a locality condition to the Bell inequalities, in the stochastic Clauser-Horne Model say, it can be shown that this (local) model, as far as applied to the singlet-state and without using quantum mechanical formalism, is not completely stochastic (i.e. there are possible configurations for which the model is deterministic). However as soon as you apply quantum mechanical formalism it becomes non-local again.
Even in experiment, where the so-called Clauser-Horne inequalities correspond to the fixed conditional probabilities of light polariser orientations for entangled photon ensembles which have not been identified show that unless they were exactly identical, the different conditional probability values of the photons themselves could not be factorized.
So the paradox is based on non-locality which is in direct contradiction to the pure local cause and effect framework of special relativity. Any theory combined with quantum mechanics becomes non-local, it has to whenever any form of quantum formalism is used. EPR is purely a paradox of relativity, not of quantum mechanics.
First Indirect Evidence of EPR Entanglement
The first efforts to relate theory and thought experiment with actual experiment came from the pioneering work of Australian physicist John Clive Ward working with British physicist Maurice Pryce along with the work of one of the greatest experimental physicists of the 20th century, Chinese-American Physicist Chien-Shiung Wu.
Their work was on formulating and experimentally verifying the probability amplitude for quantum entanglement was the first attempt to develop a way to find such "spooky actions" in an EPR apparatus.
In a 1947 paper, published in Nature[7], Ward and Pryce were the first to calculate, and use, the probability amplitudes for the polarisation of two entangled photons moving in opposite directions.
For polarisations x and y, Ward derived this probability amplitude to be
which once normalised can be expressed as
where 1 and 2 refer to the two quanta propagating in different directions. Ward's probability amplitude is then applied to derive the correlation of the quantum polarisations of the two photons propagating in opposite directions.
This prediction was experimentally confirmed by Wu and Shaknov in 1950.[8] In current terminology this result corresponds to a pair of entangled photons and is directly relevant to a typical Einstein-Podolsky-Rosen (EPR) paradox
Chien-Shiung Wu – often referred to as Madame Wu or the First Lady of Physics – from the University of Columbia was first to give indirect evidence of entanglement in the laboratory.
She showed an Einstein-type correlation between the polarisation of two well-separated photons, which are tiny localised particles of light.
Chien-Shiung Wu at her Columbia University Physics Lab, 1963.
Madame Wu's work led to the confirmation of the Pryce and Ward calculations on the correlation of the quantum polarizations of two photons propagating in opposite directions. This was the first experimental confirmation of quantum results relevant to a pair of entangled photons as applicable to the Einstein-Podolsky-Rosen (EPR) paradox.
However, direct evidence of the EPR paradox, one which would include a complete isolation of local effects and test the effect of nonlocality of quantum phenomina would require a few more decades, until the laser was invented, which allowed the French physicist Alain Aspect to form an experiment that we would recognize today as a quantum entanglement circuit.
Alain Aspect's experiment
Aspect was set upon his way by his supervising professor,
Bernard d'Espagnat, whose career centered around gathering experimental
evidence to uncover the deep nature of reality. "I had the luck,"
said d'Espagnat, "to discover in my university a young French physicist, Alain
Aspect, who was looking for a thesis subject and I suggested that testing the
Bell inequalities might be a good idea. I also suggested that he go and talk to
Bell, who convinced him it was a good idea and the outcome of this was that
quantum mechanics won.”
Aspect's experiment used particles of light – photons –
rather than material particles such as electrons or protons. Then, as now,
photons are by far the easiest quantum objects from which to produce entangled
pairs. There is, however, a minor complication concerning the property that is
actually recorded in a photon-measuring experiments such as Aspect's or those
of other researchers we'll be talking about later. Both Bell and Bohm presented
their theoretical arguments in terms of the particle spin. Photons do have a
spin (they're technically known as spin-1 particles), but because they travel at
the speed of light, their spin axes always lie exactly along their direction of
motion, like that of a spinning bullet shot from a rifle barrel. You can
imagine photons to be right-handed or left-handed depending on which way they
rotate as you look along their path of approach. What's actually measured in
the lab isn't spin, however, but the very closely related property of
polarization.
Effectively, polarization is the wavelike property of light
that corresponds to the particlelike property of spin. Think of polarization in
terms of Maxwell's equations, which tell us that the electric and magnetic
fields of a light wave oscillate at right angles to each other and also to the
direction in which the light is traveling. The polarization of a photon is the
direction of the oscillation of its electric field: up and down, side to side,
or any orientation in between. Ordinarily, light consists of photons polarized
every which way. But if light is passed through a polarizing filter, like that
used in Polaroid sunglasses, only photons with a particular polarization – the
one that matches the slant of the filter – can get through. (The same happens
if two people make waves by flicking one end of a rope held between them. If
they do this through a gap between iron railings only waves that vibrate in the
direction of the railings can slip through to the other side.)
Aspect designed his experiment to examine correlations in
the polarization of photons produced by calcium atoms – a technique that had
already been used by other researchers. He shone laser light onto the calcium
atoms, which caused the electrons to jump from the ground state to a higher
energy level. As the electrons tumbled back down to the ground state, they
cascaded through two different energy states, like a two-step waterfall,
emitting a pair of entangled photons – one photon per step – in the process.
Illustration of the Aspect Experiment
The photons passed through a slit, known as a collimator,
designed to reduce and guide the light beam. Then they fell into an automatic
switching device that randomly sent them in one of two directions before
arriving, in each case, at a polarization analyzer – a device that recorded
their polarization state.
An important consideration in Aspect's setup was the
possibility, however small, that information might leak from one photon to its
partner. It was important to rule out a scenario in which a photon arrived at a
polarization analyzer, found that polarization was being measured along say the
vertical direction, and then somehow communicated this information to the other
photon. (How this might happen doesn't matter: the important thing was to
exclude it as an option.) By carefully setting up the distances through which
the photons traveled and randomly assigning the direction in which the polarization
would be measured while the photons were in flight, Aspect ensured that under
no circumstances could such a communicating signal be sent between photons. The
switches operated within 10 nanoseconds, while the photons took 20 nanoseconds
to travel the 6.6 meters to the analyzers. Any signal crossing from one
analyzer to the other at the speed of light would have taken 40 nanoseconds to
complete the journey – much too long to have any effect on the measurement.
In a series of these experiments in the 1980s, Aspect's team
showed what most quantum theorists expected all along: Bell's inequality was
violated.[9] The result agreed completely with the predictions of standard
quantum mechanics and discredited any theories based on local hidden variables.
More recent work had backed up this conclusion. What's more, these newer
experiments have included additional refinements designed to plug any remaining
loopholes in the test. For example, special crystals have enabled experimenters
to produce entangled photons that are indistinguishable, because each member of
the pair has the same wavelength. Such improvements have allowed more accurate
measurements of the correlation between the photons. In all cases, however, the
outcomes have upheld Aspect's original discovery. Entanglement and nonlocality
are indisputable facts of the world in which we live, even if we may find it uncomfortable or "spooky" as Einstein himself did.
Einstein found it "Spooky" because the experiment is a paradox of special relativity, which is based on the idea of local causality, i.e. in a given event in space-time the distance between cause and effect must be separated by a time lapse which depends on the speed of light. In quantum entanglement a measurement, of spin say, of one particle allows you to know the spin of the other particle without disturbing it, such that the measurement of one particle is imposing an equal, but opposite, property, i.e. spin, on the other.
This is a paradox of special relativity because no events can affect another in a non-local way and must be mediated by signals. However the entangled particles, separated in space, are not separated in time. They both share a common time, however mathematically the time of one particle is real and the other is complex (i.e.using complex or imaginary numbers). So in unifying quantum mechanics with special relativity we get scenarios where, instead of the states being predetermined, one state measures time in a negative frame relative to the other and so cancel each other out.
This is why, in drawing Feynman diagrams in space-time, it appears that antimatter particles move backwards in time relative to the matter particles in pair production events for example. The particles do not travel backwards in time, this is just a consequence of unifying a local theory, special relativity, and a non-local theory, quantum mechanics. The paradox lies within this, as why should the local nature vanish? we know "how" to interpret it but we cannot really know the "why".
Niels Bohr felt as if we have no real right to know the "why", which displeased Einstein as all of his theories relied on a local space-time and to accept quantum correlations means having to violate the cosmic speed limit. Setting up the quantum systems however is determined by local events, the particles must be after all carried under local cause and effect events limited by the speed of light, so the paradox can be ironed out by what sets up the system, i.e. me bringing an entangled partner to the moon and leaving its partner on earth but as for the determination of the states of the particles themselves, in special relativity alone the paradox exists so we must interpret it using quantum mechanical non-locality.
Its kind of like saying "we see the world is flat around us, in a local frame, but by travelling around it, in a non-local frame, we know it is round" - so experimental determination of this effect, which has been done for almost 30 years now, is our equivalent of determining the roundness of the world despite it locally appearing flat in this analogy.
Applications of Quantum Entanglement
The phenomenon of entanglement has already begun to be
exploited for practical purposes. In the late 1980s, theoreticians started to
see entanglement not just as a puzzle and a way to penetrate more deeply into
the mysteries of the quantum world, but also as a resource. Entanglement could
be exploited to yield new forms of communication and computing. It was a vital
missing link between quantum mechanics and another field of explosive growth:
information theory. The proof of nonlocality and the quickly evolving ability
to work with entangled particles in the laboratory were important factors in
the birth of a new science. Out of the union of quantum mechanics and
information theory sprang quantum information science – the fast-developing
field whose most important fields of development are quantum cryptography,
quantum teleportation, and quantum computers.
To see some of the applications of quantum entanglement in the context of quantum computer technology, see my article on quantum computer physics and architecture.
To see some of the applications of quantum entanglement in the context of quantum computer technology, see my article on quantum computer physics and architecture.
To see a nice example of merging quantum information theory with game theory see my short article here
References
Einstein, A., B. Podolsky, and N. Rosen. "Can a
quantum-mechanical description of physical reality be considered incomplete?
Physical Review 47 (1935): 777-80.
Schrödinger, E. "Discussion of probability relations
between separated systems." Proceedings of the Cambridge Philosophical
Society 31 (1935): 555-63.
Scully, M. O., B. G. Englert, and H. Walther. "Quantum
optical tests of complimentarity." Nature 351 (1991): 111-16.
Dürr, S., T. Nonn, and G. Rempe. "Origin of QM
complementarity probed by a 'which-way' experiment in an atom interferometer."
Nature 395 (1998): 33.
Bohm, D. "A suggested reinterpretation of quantum
theory in terms of hidden variables." Physical Review 85 (1952): 611-23.
Bell, J. S. "On the Einstein-Podolsky-Rosen
paradox." Physics 1 (1964): 195-200.
M. H. L. Pryce and J. C. Ward, Angular correlation effects with annihilation radiation, Nature 160, 435 (1947).
C. S. Wu and I. Shaknov, The angular correlation of scattered annihilation radiation, Phys. Rev. 77, 136 (1950).
M. H. L. Pryce and J. C. Ward, Angular correlation effects with annihilation radiation, Nature 160, 435 (1947).
C. S. Wu and I. Shaknov, The angular correlation of scattered annihilation radiation, Phys. Rev. 77, 136 (1950).
Aspect, A. P., P. Grangier, and G. Roger. "Experimental
tests of relaistic local theories via Bell's theorem." Physical Review
Letters 47 (1981): 460.
No comments:
Post a Comment
Note: only a member of this blog may post a comment.