All posts by StenBlog

The Cosmological Redshift

Galaxy Redshifts Reconsidered

Written by Sten Odenwald 
Copyright (C) 1993 Sky Publishing Corporation. Reprinted by permission. See February 1993 issue

Since its discovery nearly 65 years ago, the cosmological redshift has endured as one of the most persuasive ‘proofs’ that our universe is expanding. The steps leading to its discovery are well known. Soon after Christian Doppler discovered that motion produces frequency shifts in 1842, astronomers began an aggressive spectroscopic program to measure the velocities of stars and planets using their Doppler shifts. This continued through the first few decades of the 20th century ‘culminating’ in the work by Vesto Slipher, Edwin Hubble and Milton Humason on the so-called spiral nebulae — distinctly non- stellar objects that also seemed to display star-like Doppler shifts. So long as velocities of only a few hundred kilometers per second were measured, no one questioned that the frequency shifts for the spiral nebulae indicated relative motion just as they had for stars and planets.
But, during the 1920’s and 30’s spiral nebulae with Doppler shifts of over 34,000 kilometers per second were discovered. In a letter by Hubble to the Dutch cosmologist Willem De Sitter in 1931, he stated his concerns about these velocities by saying “… we use the term ‘apparent velocities’ in order to emphasize the empirical feature of the correlation. The interpretation, we feel, should be left to you and the very few others who are competent to discuss the matter with authority.” Dispite this cautionary note, the fact of the matter was that the redshifts measured for the distant galaxies LOOKED like Doppler shifts. The terms ‘recession velocity’ and ‘expansion velocity’ were quickly brought into service by astronomers at the telescope, and by popularizers, to describe the physical basis for the redshift.

As astronomers explored the universe to greater depths, galaxies and quasars appeared to be rushing away at faster and faster speeds. It seems to be a completely natural consequence of the outrushing of matter from the big bang. Like a sparkling display of fireworks on a warm summer evening, we imagine ourselves standing on one of those galactic ‘cinders’, watching the others rush past us into the dark void of infinite space. Upon closer examination, however, this intuitively-compelling and seductive mental image is both inadequate and misleading.

The Mysteries of Relativity

Big bang cosmology is based on Einstein’s general theory of relativity. It is a theory transcending both Newton’s mechanics and Einstein’s special theory of relativity, introducing us to concepts that do not exist within the older theories. Nor are these concepts easily comprehensible by our common sense which has been honed by organic evolution to see the world only through a narrow set of glasses.

For example, special relativity is based on the difficult-to-fathom postulate that the speed of light is absolutely constant when measured in reference frames moving at a constant speed. From this emerges the concept of ‘spacetime’ which then becomes the arena for all phenomena involving time dilation, length contraction and the Twin Paradox. Beyond special relativity lies the incomparably more alien landscape of general relativity. Gravitational fields now become geometric curvatures of spacetime. This has no analog in special relativity based as it is on a perfectly flat spacetime that remains aloof from any influence on it by matter or energy.

Just as the constancy of the speed of light led to the Twin Paradox, the curvature of spacetime leads to its own menageri of peculiar phenomena. One of these involves the slowing-down of clocks in the presence of a strong gravitational field. Related to this is the “gravitational redshift” which occurs when the frequency of light sent from the surface of a body is shifted to lower frequencies during the journey to the observer. This redshift is not related to the famous Doppler shift since the observer is not in motion relative to the body emitting the light signal!

A second phenomenon predicted by general relativity that also has no analog in special relativity is the cosmological redshift. Simply stated, the cosmological redshift occurs because the curvature of spacetime was smaller in the past when the universe was younger than it is now. Light waves become stretched en route between the time they were emitted long ago, and the time they are detected by us today.

The Doppler shift and cosmology

It is tempting to refer to cosmological redshifts as Doppler shifts. This choice of interpretation has in the years since Hubble’s work led to an unfortunate misunderstanding of big bang cosmology, obscurring one of its most mysterious beauties. As noted with a hint of frustration by cosmologists such as Steven Weinberg and Jaylant Narlikar and John Wheeler, “The frequency of light is also affected by the gravitational field of the universe, and it is neither useful nor strictly correct to interpret the frequency shifts of light…in terms of the special relativistic Doppler effect.”.

By refering to cosmological redshifts as Doppler shifts, we are insisting that our Newtonian intuition about motion still applies without significant change to the cosmological arena. A result of this thinking is that quasars now being detected at redshifts of Z = 4.0 would have to be interpreted as traveling a speeds of more than V = Z x c or 4 times the speed of light. This is, of course, quite absurd, because we all know that no physical object may travel faster than the speed of light.

To avoid such apparently nonsensical speeds, many popularizers use the special relativistic Doppler formula to show that quasars are really not moving faster than light. The argument being that for large velocities, special relativity replaces Newtonian physics as the correct framework for interpreting the world. By using a special relativistic velocity addition formula the quasar we just discussed has a velocity of 92 percent the speed of light. Although we now have a feeling that Reason has returned to our description of the universe, in fact, we have only replaced one incomplete explanation for another. The calculation of the quasar’s speed now presupposes that special relativity ( a theory of flat spacetime) is applicable even at cosmological scales where general relativity predicts that spacetime curvature becomes important. This is equivalent to a surveyor making a map of the state of California, and not allowing for the curvature of the earth!

The adoption of the special relativistic Doppler formula by many educators has led to a peculiar ‘hybrid’ cosmology which attempts to describe big bang cosmology using general relativity, but which is still firmly mired in the ruberik of special relativity. For instance, under the entry ‘redshift’ in the Cambridge Encyclopedia of Astronomy it is explicitly acknowledged that the redshift is not a Doppler shift, but less than two paragraphs later, the special relativistic Doppler formula is introduced to show how quasars are moving slower than the speed of light! It is also common for popularizers of cosmology to describe how ‘space itself stretches’ yet continue to describe the expansion of the universe as motion governed by the restrictions of special relativity. What’s going on here?

General relativity to the rescue

By adopting general relativity as the proper guide, such contradictions are eliminated. General relativity leads us to several powerful conclusions about our cosmos: 1) special relativity is inapplicable for describing the larger universe; 2) the concepts of distance and motion are not absolutely defined and 3) Preexisting spacetime is undefined. Each of these conclusions are as counter-intuitive as the Twin Paradox or as the particle/wave dualism of quantum mechanics. As Nobel Physicist John Wheeler once put it “If you are not completely confused by quantum mechanics, you do not understand it” The same may be said for general relativity.

The first conclusion means that we cannot trust even the insights hard won from special relativity to accurately represent the ‘big picture’ of the universe. General relativity must replace special relativity in cosmology because it denies a special role to observers moving at constant velocity, extending special relativity into the arena of accelerated observers. It also denies a special significance to special relativity’s flat spacetime by relegating it to only a microscopic domain within a larger geometric possibility. Just as Newtonian physics gave way to special relativity for describing high speed motion, so too does special relativity give way to general relativity. This means that the special relativistic Doppler formula should not, in fact cannot, be used to quantify the velocity of distant quasars. We have no choice in this matter if we want to maintain the logical integrity of both theories.

Distance and motion

The second conclusion is particularly upsetting because if we cannot define what we mean by distance, how then can we discuss in meaningful terms the ‘motion’ of distant quasars, or a Hubble Law interpreted as a distance versus velocity relation? In a small region of spacetime, we can certainly define motion as we always have because space has a static, flat geometry. When a body moves from point x to point y in a time interval, T, we say it is moving with a speed of S = (x – y)/T. There are also specific experimental ways of measuring x, y and T to form the quotent S by using clocks and rulers. The crucil feature behind these measurements is that nothing happens to the geometry of space during the experiment to change the results of the measuring process.

In the cosmological setting which we believe is accurately described by general relativity, we have none of these luxuries! Astronomers cannot wait millions of years to measure quasar proper motions. They cannot, like Highway Patrol officers, bounce radar beams off distant galaxies to establish their relative distances or speeds. Unlike all other forms of motion that have been previously observed, cosmological ‘motion’ cannot be directly observed. It can only be INFERRED from observations of the cosmological redshift, which general relativity then TELLS US means that the universe is expanding.

In big bang cosmology, galaxies are located at fixed positions in space. They may perform small dances about these positions in accordance with special relativity and local gravitational fields, but the real ‘motion’ is in the literal expansion of space between them! This is not a form of movement that any human has ever experienced. It is, therefore, not surprising that our intuition reels at its implication and seeks other less radical interpretations for it including special relativity. But even the exotic language and conundrums of special relativity cannot help us. Instead we are forced to interrogate the mathematics of general relativity itself for whatever landmarks it can provide. In doing so, we are left, however, with a riddle as profound as that of the Twin Paradox, and equally challenging to explain.

Two galaxies permanently located at positions (x1 , y1 , z1 ) and ( x2 , y2 , z2 ) at one time find themselves one billion light years apart. Then a few billion years later while located at the same coordinates, they find themselves 3 billion light years apart. The galaxies have not ‘moved’, nevertheless, their separations have increased. In fact, when the universe was only one year old, the separations between these galaxies were increasing at 300 times the speed of light! Space can expand faster than the speed of light in general relativity because space does not represent matter or energy. The displacements that arise from its dilation produce an entirely new kind of motion for which even our special relativistically-trained intuitions remain profoundly silent. Like that gentleman from Main once said “You can’t get there [to general relativity] from here [special relativity]”. To the extent that general relativity has been tested and found correct, we have no choice but to accept its consequences at face value.

Space, time and matter

The last conclusion drawn from general relativistic cosmology is that, unlike special relativity, it is not physically meaningful to speak of spacetime existing independently of matter and energy. In big bang cosmology, both space and time came into existence along side matter and energy at ‘time zero’. If our universe contains more than a critical density of matter and energy, its spacetime is forever finite and bounded, in a shape analogous to a sphere. Beyond this boundary, space and time simply do not exist. In fact, general relativity allows the Conservation of Energy to be suspended so that matter and energy may be created quite literally from the nothingness of curved spacetime. General relativity provides a means for ‘jump-starting’ Creation!

Big bang cosmology is both a profoundly beautiful, and disturbing, model for our universe, its shape and its destiny. It contains many surprises which have yet to be completely worked-out. But one feature of the evolving universe seems absolutly clear, the big bang was not some grand fireworks display, but an event of a completely different order. It resembled more an expanding soap bubble film upon which galactic dust motes are carried along for the ride. This film represents the totality of all the space and matter in our universe, and it expands into a mysterious primordial void which is itself empty of space, dimension, time or matter.

In the future it is hoped that a death knell will finally have sounded for the last vestage of the older thinking. With the Doppler interpretation of the cosmological redshift at last reconsidered, and rejected, we will finally be able to embrace the essential beauty and mystery of cosmic expansion as it was originally envisioned by its discoverers.

Einstein’s Fudge

Einstein’s Cosmic Fudge Factor

Written by Sten Odenwald
Copyright (C) 1991. Sky Publishing Corporation. Reprinted by permission. See April, 1991 issue

Black holes…quarks…dark matter. It seems like the cosmos gets a little stranger every year. Until recently, the astronomical universe known to humans was populated by planets, stars, galaxies, and scattered nebulae of dust and gas. Now, theoretists tell us it may also be inhabited by objects such as superstrings, dark matter and massive neutrinos — objects that have yet to be discovered if they exist at all!
As bizarre as these new constituents may sound, you don’t have to be a rocket scientist to appreciate the most mysterious ingredient of them all. It is the inky blackness of space itself that commands our attention as we look at the night sky; not the sparse points of light that signal the presence of widely scattered matter.

During the last few decades, physicists and astronomers have begun to recognize that the notion of empty space presents greater subtleties than had ever before been considered. Space is not merely a passive vessel to be filled by matter and radiation, but is a dynamic, physical entity in its own right.

One chapter in the story of our new conception of space begins with a famous theoretical mistake made nearly 75 years ago that now seems to have taken on a life of its own.

In 1917, Albert Einstein tried to use his newly developed theory of general relativity to describe the shape and evolution of the universe. The prevailing idea at the time was that the universe was static and unchanging. Einstein had fully expected general relativity to support this view, but, surprisingly, it did not. The inexorable force of gravity pulling on every speck of matter demanded that the universe collapse under its own weight.

His remedy for this dilemma was to add a new ‘antigravity’ term to his original equations. It enabled his mathematical universe to appear as permanent and invariable as the real one. This term, usually written as an uppercase Greek lambda, is called the ‘cosmological constant’. It has exactly the same value everywhere in the universe, delicately chosen to offset the tendency toward gravitational collapse at every point in space.

A simple thought experiment may help illustrate the nature of Lambda. Take a cubic meter of space and remove all matter and radiation from it. Most of us would agree that this is a perfect vacuum. But, like a ghost in the night, the cosmological constant would still be there. So, empty space is not really empty at all — Lambda gives it a peculiar ‘latent energy’. In other words, even Nothing is Something!

Einstein’s fudged solution remained unchallenged until 1922 when the Russian mathematician Alexander Friedmann began producing compelling cosmological models based on Einstein’s equations but without the extra quantity. Soon thereafter, theorists closely examining Einstein’s model discovered that, like a pencil balanced on its point, it was unstable to collapse or expansion. Later the same decade, Mount Wilson astronomer Edwin P. Hubble found direct observational evidence that the universe is not static, but expanding.

All this ment that the motivation for introducing the cosmological constant seemed contrived. Admitting his blunder, Einstein retracted Lambda in 1932. At first this seemed to end the debate about its existence. Yet decades later, despite the great physicist’s disavowal, Lambda keeps turning up in cosmologists’ discussions about the origin, evolution, and fate of the universe.

THEORY MEETS OBSERVATION

Friedmann’s standard ‘Big Bang’ model without a cosmological constant predicts that the age of the universe, t0, and its expansion rate (represented by the Hubble parameter, H0) are related by the equation t0 = 2/3H0. Some astronomers favor a value of H0 near 50 kilometers per second per megaparsec (one megaparsec equals 3.26 million light years). But the weight of the observational evidence seems to be tipping the balance towards a value near 100. In the Friedmann model, this implies that the cosmos can be no more than 7 billion years old. Yet some of our galaxy’s globular clusters have ages estimated by independent methods of between 12 and 18 billion years!

In what’s called the Einstein-DeSitter cosmology, the Lambda term helps to resolve this discrepancy. Now a large value for the Hubble parameter can be attributed in part to “cosmic repulsion”. This changes the relationship between t0 and H0, so that for a given size, the universe is older than predicted by the Friedmann model.

In one formulation of Einstein’s equation, Lambda is expressed in units of matter density. This means we can ask how the cosmological constant, if it exists at all, compares with the density of the universe in the forms of stars and galaxies.

So far, a careful look at the available astronomical data has produced only upper limits to the magnitude of Lambda. These vary over a considerable range – from about 10 percent of ordinary matter density to several times that density.

The cosmological constant can also leave its mark on the properties of gravitational lenses and faint galaxies. One of the remarkable features of Einstein’s theory of general relativity is its prediction that space and time become deformed or ‘warped’ in the vicinity of a massive body such as a planet, star or even a galaxy. Light rays passing through such regions of warped “space-time” have their paths altered. In the cosmological arena, nearby galaxies can deflect and distort the images of more distant galaxies behind them. Sometimes, the images of these distant galaxies can appear as multiple images surrounding the nearby ‘lensing’ galaxy.

At Kyoto University M. Fukugita and his coworkers predicted that more faint galaxies and gravitational lenses will be detected than in a Friedmann universe if Lambda is more than a few times the matter density. Edwin Turner, an astrophysicist at Princeton University also reviewed the existing, scant, data on gravitational lenses and found that they were as numerous as expected for Lambda less that a few times the matter density. By the best astronomical reconning, Lambda is probably not larger than the observed average matter density of the universe. For that matter, no convincing evidence is available to suggest that Lambda is not exactly equal to zero. So why not just dismiss it as an unnecessary complication? Because the cosmological constant is no longer, strictly, a construct of theoretical cosmology.

NOTHING AND EVERYTHING

To understand how our universe came into existence, and how its various ingredients have evolved, we must delve deeply into the fundamental constituents of matter and the forces that dictate how it will interact. This means that the questions we will have to ask will have more to do with physics than astronomy. Soon after the big bang, the universe was at such a high temperature and density that only the details of matter’s composition (quarks, electrons etc) and how they interact via the four fundamental forces of nature were important. They represented the most complex collections of matter in existence, long before atoms, planets, stars and galaxies had arrived on the scene.

For two decades now, physicists have been attempting to unify the forces and particles that make up our world – to find a common mathematical description that encompasses them all. Some think that such a Theory of Everything is just within reach. It would account not only for the known forms of matter, but also for the fundamental interactions among them: gravity, electromagnetism, and the strong and weak nuclear forces.

These unification theories are known by a variety of names: grand unification theory, supersymmetry theory and superstring theory. Their basic claim is that Nature operates according to a small set of simple rules called symmetries.

The concept of symmetry is at least as old as the civilization of ancient Greece, whos art and archetecture are masterworks of simplicity and balance. Geometers have known for a long time that a simple cube can be rotated 90 degrees without changing its outward appearance. In two dimensions, equalateral triangles look the same when they are rotated by 120 degrees. These are examples of the geometric concept of Rotation Symmetry.

There are parallels to geometric symmetry in the way that various physical phenomena and qualities of matter express themselves as well. For example, the well-known principle of the Conservation of Energy is a consequence of the fact that when some collections of matter and energy are examined at different times, they each have precisely the same total energy, just as a cube looks the same when it is rotated in space by a prescribed amount. Symmetry under a ‘shift in time’ is as closely related to the Conservation of Energy as is the symmetry of a cube when rotated by 90 degrees.

Among other things, symmetries of Nature dictate the strengths and ranges of the natural forces and the properties of the particles they act upon. Although Nature’s symmetries are hidden in today’s cold world, they reveal themselves at very high temperatures and can be studied in modern particle accelerators.

The real goal in unification theory is actually two-fold: not only to uncover and describe the underlying symmetries of the world, but to find physical mechanisms for ‘breaking’ them at low energy. After all, we live in a complex world filled with a diversity of particles and forces, not a bland world with one kind of force and one kind of particle!

Theoreticians working on this problem are often forced to add terms to their equations that represent entirely new fields in Nature. The concept of a field was invented by mathematicians to express how a particular quantity may vary from point to point in space. Physicists since the 18th century have adopted this idea to describe quantitatively how forces such as gravity and magnetism change at different distances from a body.

The interactions of these fields with quarks, electrons and other particles cause symmetries to break down. These fields are usually very different than those we already know about. The much sought after Higgs boson field, for example, was introduced by Sheldon Glashow, Abdus Salam and Steven Weinberg in their unified theory of the electromagnetic and weak nuclear forces.

Prior to their work, the weak force causing certain particles to decay, and the electromagnetic force responsible for the attraction between charged particles and the motion of compass needles, were both considered to be distinct forces in nature. By combining their mathematical descriptions into a common language, they showed that this distinction was not fundamental to the forces at all! A new field in nature called the Higgs field makes these two forces act differently at low temperature. But at temperatures above 1000 trillion degrees, the weak and electromagnetic forces become virtually identical in the way that they affect matter. The corresponding particles called the Higgs Boson not only cause the symmetry between the electromagnetic and weak forces to be broken at low temperature, but they are also responsible for confiring the property of mass on particles such as the electrons and the quarks!

There is, however a price that must be paid for introducing new fields into the mathematical machinery. Not only do they break symmetries, but they can also give the vacuum state an enormous latent energy that, curiously, behaves just like Lambda in cosmological models.

The embarrassment of having to resurrect the obsolete quantity Lambda is compounded when unification theories are used to predict its value. Instead of being at best a vanishingly minor ingredient to the universe, the predicted values are in some instances 10 to the power of 120 times greater than even the most generous astronomical upper limits!

It is an unpleasant fact of life for physicists that the best candidates for the Theory of Everything always have to be fine-tuned to get rid of their undesirable cosmological consequences. Without proper adjustment, these candidates may give correct predictions in the microscopic world of particle physics, but predict a universe which on its largest scales looks very different from the one we inhabit.

Like a messenger from the depths of time, the smallness – or absence – of the cosmological constant today is telling us something important about how to craft a correct Theory of Everything. It is a signpost of the way Nature’s symmetries are broken at low energy, and a nagging reminder that our understanding of the physical world is still incomplete in some fundamental way.

A LIKELY STORY

Most physicists expect the Theory of Everything will describe gravity the same way we now describe matter and the strong, weak and electromagnetic forces – in the language of quantum mechanics. Gravity is, after all, just another force in Nature. So far this has proven elusive, due in part to the sheer complexity of the equations of general relativity. Scientists since Einstein have described gravity ( as well as space and time) in purely geometric terms. Thus we speak of gravity as the “curvature of space-time”.

To acheive complete unification, the dialects of quantum matter and geometric space have to be combined into a single language. Matter appears to be rather precisely described in terms of the language of quantum mechanics. Quarks and electrons exchange force-carrying particles such as photons and gluons and thereby feel the electromagnetic and strong nuclear forces. But, gravity is described by Einstein’s theory of general relativity as a purely geometric phenomenon. These geometric ideas of curvature and the dimensionality of space have nothing to do with quantum mechanics.

To unify these two great foundations of physics, a common language must be found. This new language will take some getting used to. In it, the distinction between matter and space dissolves away and is lost completely; matter becomes a geometric phenomenon, and at the same time, space becomes an exotic form of matter.

Beginning with work on a quantum theory of gravity by John Wheeler and Bryce DeWitt in the 1960’s, and continuing with the so-called superstring theory of John Schwartz and Michael Green in the 1980’s, a primitive version of such a ‘quantum-geometric’ language is emerging. Not surprisingly, it borrows many ideas from ordinary quantum mechanics.

A basic concept in quantum mechanics is that every system of elementary particles is defined by a mathematical quantity called a wave function. This function can be used, for example, to predict the probability of finding an electron at a particular place and time within an atom. Rather than a single quantity, the wave function is actually a sum over an infinite number of factors or ‘states’, each representing a possible measurement outcome. Only one of these states can be observed at a time.

By direct analogy, in quantum gravitation, the geometry of space-time, whether flat or curved, is only one of an infinite variety of geometric shapes for space-time, and therefore the universe. All of these possibilities are described as separate states in the wave function for the universe.

But what determines the probability that the universe will have the particular geometry we now observe out of the infinitude of others? In quantum mechanics, the likelihood that an electron is located somewhere within an atom is determined by the external electric field acting on it. That field is usually provided by the protons in the atomic nucleus. Could there be some mysterious field ‘outside’ our universe that determines its probability?

According to Cambridge University theorist Stephen Hawking, this is the wrong way to look at the problem. Unlike the electron acted upon by protons, our universe is completely self-contained. It requires no outside conditions or fields to help define its probability. The likelihood that our universe looks the way it does depends only on the strengths of the fields within it.

Among these internal fields, there may even be ones that we haven’t yet discovered. Could the cosmological constant be the fingerprint in our universe of a new ‘hidden’ field in Nature? This new field could affect the likelihood of our universe just as a kettle of soup may contain unknown ingredients although we can still precisely determine the kettle’s mass.

A series of mathematical considerations led Hawking to deduce that the weaker the hidden field becomes, the smaller will be the value we observe for the cosmological constant, and surprisingly, the more likely will be the current geometry of the universe.

This, in turn, implies that if Lambda were big enough to measure by astronomers in the first place, our universe would be an improbable one. Philosophically, this may not trouble those who see our cosmos as absolutely unique, but in a world seemingly ruled by probability, a counter view is also possible. There may, in fact, exist an infinite number of universes, but only a minority of them have the correct blend of physical laws and physical conditions resembling our life-nurturing one.

Hawking continued his line of speculation by suggesting that, if at the so-called Planck scale of 10 to the power of -33 centimeters the cosmos could be thought of as an effervescent landscape, or “space-time foam”, then perhaps a natural mechanism could exist for eliminating the cosmological constant for good.

One of the curiosities of combining the speed of light and Newton’s constant of gravitation from general relativity, with Planck’s constant from quantum mechanics, is that they can be made to define unique values for length, time and energy. Physicists believe that at these Planck scales represented by 10 to the power of -33 centimeters and 10 to the power of -43 seconds, general relativity and quantum mechanics blend together to become a single, comprehensive theory of the physical world: The Theory Of Everything. The energy associated with this unification, 10 to the power of 19 billion electron volts, is almost unimaginably big by the standards of modern technology.

The universe itself, soon after the Big Bang, must also have passed through such scales of space, time and energy during its first instants of existence. Cosmologists refer to this period as the Planck Era. It marks the earliest times that physicists are able to explore the universe’s physical state without having a complete Theory of Everything to guide them.

WORMHOLES

Harvard University physicist Sidney Coleman has recently pursued this thought to a possible conclusion. Instead of some mysterious new field in Nature, maybe the Lambda term appears in our theories because we are using the wrong starting model for the geometry of space at the Planck scale.

Previous thinking on the structure of space-time had assumed that it behaved in some sense like a smooth rubber sheet. Under the action of matter and energy, space-time could be deformed into a variety of shapes, each a possible geometric state for the universe. Nearly all candidates for the Theory of Everything’s embed their fields and symmetries in such a smooth geometrical arena.

But what if space-time were far more complicated? One possibility is that ‘wormholes’ exist, filling space-time with a network of tunnels. The fabric of space-time may have more in common with a piece of Swiss cheese than with a smooth rubber sheet.

According to Coleman, the addition of wormholes to space-time means that, like the ripples from many stones tossed into a pond, one geometric state for the universe could interfere with another. The most likely states ( or the biggest ripples) would win out. The mathematics suggest that quantum wormhole interference at the Planck scale makes universes with cosmological constants other than zero exceedingly unlikely.

How big would wormholes have to be to have such dramatic repurcussions? Surprisingly, the calculations suggest that small is beautiful. Wormholes the size of dogs and planets would be very rare. Universes containing even a few of them would exist with a vanishingly low probability. But wormholes smaller than 10 to the power of -33 centimeters could be everywhere. A volume the size of a sugar cube might be teeming with uncounted trillions of them flashing in and out of existence!

Coleman proposes that the action of these previously ignored mini- wormholes upon the geometric fabric of the universe that forces Lambda to be almost exactly zero. Like quantum ‘Pac Men’, they gobble up all the latent energy of space-time that would otherwise have appeared to us in the form of a measureable cosmological constant!

The addition of wormholes to the description of space-time admits the possibility that our universe did not spring into being aloof and independent, but was influenced by how other space-times had already evolved – ghostly mathematical universes with which we can never communicate directly.

The most likely of these universes had Lambda near zero, and it is these states that beat out all other contenders. In a bizarre form of quantum democracy, our universe may have been forced to follow the majority, evolving into the high probability state we now observe, without a detectable cosmological constant.

EPILOG

Wormholes? Wave functions? Hidden fields? The answer to the cosmological constant’s smallness, or absence, seems to recede into the farthest reaches of abstract thinking, faster than most of us can catch up.

As ingenious as these new ideas may seem, the final pages in this unusual story have probably not been written, especially since we can’t put any of these ideas to a direct test. It is a tribute to Einstein’s genius that even his ‘biggest blunder’ made near the beginning of this century still plagues physicists and astronomers as we prepare to enter the 21st century. Who would ever have thought that something that may not even exist would lead to such enormous problems!

The Planck Era

The Planck Era

Written by Sten Odenwald. Copyright (C) 1984 Kalmbach Publishing. Reprinted by permission

The Big Bang theory says that the entire universe was created in a tremendous explosion about 20 billion years ago. The enormity of this event is hard to grasp and it seems natural to ask ourselves ‘What was it like then?’ and ‘What happened before the Big Bang?’. To try to answer these queries, lets take a brief journey backwards in time.
We first see the formation of our own sun about 15 billion years after the Big Bang and then by 5 billion years, the formation of the first galaxies. By 700,000 years, the universe is awash with the fireball radiation that keeps all matter at a temperature of 4,000 degrees. Because of this, darkness is completely absent since every point in the sky glows with the brilliance of the sun. No stars, planets or even dust grains exist, just a hot dense plasma of electrons, protons and helium nuclei. By 3 minutes, we see helium form from the fusion of hydrogen atoms while the universe seeths at a temperature of nearly 1 billion degrees. The average density of matter is that of lead. By 1 second, the Lepton Era ends and the ratio of neutrons to protons has become fixed at 1 neutron for every 5 protons. The temperature is now 5 billion degrees everywhere. At about .0001 second, we watch as the Quark Era ends and the temperature of the fireball radiation rises to an incredable 1 trillion degrees. Quarks, for the first time, can combine in groups of two and three to become neutrons, protons and other types of heavy particles. The universe is now packed with matter as densly as the nucleus of an atom. A mountain like Mt. Everest could be squeezed into a volume no greater than the size of a golf ball!

By 1 billionth of a second, the temperature is 1 thousand trillion degrees and we see the electromagnetic and weak forces merge into one force. The density of the universe has increased to the point where the entire earth could be contained in a thimble. Quarks and anti-quarks are no longer confined inside of particles like neutrons and protons but are now part of a superheated plasma of unbound particles. As the remaining history of the universe unfolds, a long period seems to pass when nothing really new happens. Then, at a time 10(-35) second after the Big Bang, a spectac ular change in the size of the universe occurs. This is the GUT Era when the strong nuclear force becomes distinguishable from the weak and electromagnetic forces. The temperature is an incredable 10 thousand trillion trillion degrees and the density of matter has sored to nearly 10(75) gm/cm3. This number is so enormous that even our analogies are almost beyond comprehension. At these densities, the entire Milky Way galaxy could easily be stuffed into a volume no larger than a single hydrogen atom! Electrons and quarks together with their anti-particles, were the major constituents of matter and very massive particles called Leptoquark Bosons caused the quarks to decay into electrons and vice versa. If we now move forward in time we would witness the vacuum of space undergoing a ‘phase transition’ from a higher energy state to a lower energy state. This is analogous to a ball rolling down the side of a mountain and coming to rest in the lowest valley. As the universe ‘rolls down hill’ it begins a brief but stupendous period of expansion. The universe swells to billions of times its former size in almost no time at all.

In addition to this, a slight excess of matter over anti-matter appears becaus of the decay of massive particles called X Higgs Bosons. As we continue to watch the universe age, the remaining pairs of particles and anti-particles find themselves and vanish in a tremendous burst of annihilation. From this paroxysm, the bulk of the fireball radiation that we now observe is born.

The GUT Era is the last stop in our fanciful journey through time. If we had asked what it was like before the GUT Era, we would immediately have entered a vast no mans land where few indisputable facts would serve to gui de us. What does seem clear is that gravity is destined to grow in importance, eventually becoming the dominant force acting between parti cles, even at the microscopic level.

G R A V I T Y

According to theories developed since the 1930’s, what we call a ‘force’ is actually a collective phenomenon caused by the exchange of innumerable, force-carrying particles called gauge bosons. The electromagnetic force, which causes like charges to attract and dissimilar ones to repel, is transmitted by gauge bosons called photons, the strong force that binds nucleii together is transmitted by gluons and the weak force which causes particles to decay is transmitted by the, recently discovered, W and Z Intermediate Vector Bosons. In an analogous way, physicists believe that gravity is transmitted by particles called Gravitons. If gravity really does have such a quantum property, its effects should appear once quarks and electrons can be forced to within 10(-33) centimeter of one another, a distance called the Planck length. To acheive these conditions, quarks and electrons will have to be collided at energies of 10(19) GeV. An accelerator patterned after the 2-mile, Stanford Linear Accelerator would have to be 1 light-year in length to push particles to these incredable energies! Fortunatly, what humans find impossible to do, Nature with its infinite resources finds less difficult. Before the universe was 10(-43) second old, matter routinely experienced collisions at these energies. This period is what we call the Planck Era.

THROUGH A LOOKING GLASS, DARKLEY

Since our technology will not allow us to physically reproduce the conditions during these ancient times, we must use our mathematical theories of how matter behaves to mentally explore what the universe was like then. We know that the appearence of the universe before 10(-43) second can only be adequatly described by modifying the Big Bang theory because this theory is, in turn, based on the General Theory of Relativity. General Relativity tells us how gravity operates on the macroscopic scale of planets, stars and galaxies. At the Planck scale, we need to extend General Relativity so that it includes not only the macroscopic properties of gravity but also is microscopic characteristics as well. The theory of ‘Quantum Gravity’ is still far from completion but physicists tend to agree that, at the very least, Quantum Gravity must combine the conceptual elements of the two great theories of modern physics: General Relativity and Quantum Mechanics.

In the language of General Relativity, gravity is a consequence of the deformati on of space caused by the presence of matter and energy. Gravity is just another name for the amount of curvature in the geometry of 3-dimensional space. In Quantum Gravity theory, gravity is produced by massless gravitons so that gravitons now represent individual packages of curved space that travel through space at the speed of light.

The appearence and dissappearence of innumerable gravitons gives the geometry of space a very lumpy and dynamic appearance. John Wheeler at Princeton University thinks of this as a foamy, sub-structure to space where the geometry of space twists and contorts so that far flung regions of space may suddenly find themselves connected by ‘wormholes’ which constantly appear and dissappear within 10(-43) seconds. Even as you are reading this article, this frenetic activity is occurring in the hyper-microscopic domain, 100 billion billion times smaller than the nucleus of an atom. For a comparison, the size of the sun and the size of a single atom stand in about this same proportion. Although Quantum Gravity effects are completely undetectable today at the atomic and nuclear scale, during the Planck Era, macroscopic and microscopic worlds merged and the Quantum Gravity of the microcosm suddenly became the Quantum Cosmology of the macrocosm!

QUANTUM COSMOLOGY

As we approach the end of the Planck Era, the random appearance and dissappearance of innumerable gravitons will eventually force us to give up the concept of a specific geometry to 3-dimensional space. Instead, the geometry at a given moment will have to be thought of as an average over all 3-dimensional space geometries that are possible. Once again, the reason for this is that particles are squeezed so closely together that we can now see individual gravitons moving around in the space between them causing space to become curved. We can no longer get away with saying that the space between two quarks, for example, is flat. This is what we mean when we say that the gravitational force between them is insignificant when compared to the other three forces of Nature.

To make matters much worse, not only will Quantum Gravity not allow us to calculate the exact 3-dimensional geometry to space but, at the Planck scale, it will not allow us to simultaneously determine its exact geometry and precise rate of change in time. What this means is that we may never be able to calculate with any certainty exactly what the history of the universe was like before 10-43 second. Today, the large-scale geometry of space is one of three possible types: flat and infinite, negatively curved and infinite or positively curved and finite. During the Planck Era, the ‘large-scale’ geometry was contorted by wormholes and and infinite number of possibilities were possible. To probe the history of the universe then would be like trying to trace your ancestral roots if every human being on earth had a possibility of being one of your parents. Now try to trace your family tree back a few generations! The farther back in time you go, the greater are the number of possible ancestors you could have had. An entirely new conception of what we mea n by ‘a history for the universe’ will have to be developed. Even the concepts of space and time will have to be completely re-evaluated in the face of the qua ntum fluctuations of spacetime at the Planck Era!

THE BIRTH OF THE UNIVERSE

The picture that seems to emerge from using our sketchy outline of what Quantum Gravity theory might look like is that as we approach the Planck Era, gravitons are exchanged between quarks and electrons with increasingly higher energy and in greater number. By the time we reach the end of the Planck Era at 10(-43) second, gravitons will begin to carry as much energy as the other force carriers (Gluons, IVBs and Photons). At still earlier times, a period of complet e symmetry and unification between all the natural forces will ensue. Only one super-unified force exists here (gravity) and only one kind of particle dominates the activity of this age(Gravitons).

During the early 70’s, the Russian physicists Ya. Zel’dovitch and A. Starobinski of the USSR Academy of Science proposed that the rapidly changing geometry of space during the Planck Era may actually have created all the matter, anti-matter and radiation that existed soon after Creation. In their picture of Creation, the rapidly changing geometry of space created particles and anti-particles with masses of 10(19) GeV. This production of matter and anti-matter removed energy from the enormous fluctuations occuring in the geometry of space and eventually succeeded in damping them out altogether by the end of the Planck Era. They also found that the rate of particle creation increased as more and more particles were created.

Several recent studies by Physicists Edward Tryon of Hunter College, R. Brout, F. Englert and E. Gunzig of the University of Brussels and david Atkatz and Heinz Pagels of the Rockefeller University have shed additional light on what Creation may have been like. Imagine if you can, nothing at all! This is the primordial vacuum of space. There is complete darkness here, no light yet exists. The number of dimensions to space was probably not the normal 3 that we are so accustomed to but may have been as high as 11 according to Supergravity theory! In this infinite emptiness, random fluctuations occurred that ever so slightly changed the energy of the vacuum at various points in space. Eventually, one of these fluctuations attained a critical energy and began to grow. As it grew, very massive particles called leptoquarks and anti-leptoquarks were created, causing the expansion to accelerate. This is much like a ball rolling down a hill that moves slowly at first and then gains momentum. The expansion of the proto-universe, in turn, caused still more leptoquarks to be created. This furious cycle continued until, at long last, the leptoquarks decayed into quarks, leptons (electrons, muons etc) and their anti-particles and the universe emerged from the Planck Era. Particle creation stopped once the fluctuations in the geometry of space subsided.

So, we are left with the remarkable possibility that, in the beginning, there ex isted quite literally, nothing at all and from it emerged nearly all of the matter and radiation that we now see. This process has been described by the physicist Frank Wilczyk at the University of California, Santa Barbara by saying, ” The reason that there is something instead of nothing is that nothing is unstable”. A ball sitting on the summit of a steep hill needs but the slightest tap to set it in motion. A random fluctuation in space was apparently all that was required to unleash the incredable latent energy of the vacuum, thus creating matter and energy and an expanding universe from ‘nothing at all’.

The universe did not spring into being instantaneously but was created a little bit at a time in a ‘bootstrap’ process. Once a few particles were created by quantum fluctuations of the empty vacuum, it became easier for a few more to appear and so, in a rapidly escalating process, the universe gushed forth from nothingness.

How long did this take? The primordial vacuum could have existed for an eternity before the particular fluctuation that gave rise to our universe happened. Physicist Edward Tryon expresses this best by saying that ” Our universe is simply one of those things that happens from time to time”.

The principles of Quantum Gravity may ultimatly force us to reconsider questions like ‘What happened before the Big Bang?’ because they imply the existence of something (time) that may not have any meaning at all. These questions may be as empty of meaning as an explorer on the north pole asking, ‘Which way is North?’. Only the complete theory of Quantum Gravity may tell us how to ask the right questions!

What is Space? Part II

Space-Time: The Final Frontier

Written by Sten Odenwald. Copyright (C) 1995 Sky Publishing Corporation. See February 1996 issue.
THE NIGHT SKY, when you think about it, is one of the strangest sights imaginable. The pinpoint stars that catch your eye are all but swallowed up by the black nothingness of space – an entity billions of light-years deep with which we here on Earth have no direct ex- perience.
What is empty space, really? At first the question seems silly. There’s nothing to it! But look again in light of what modern physics knows and suspects, and the nature of space emerges as one of the most important “sleeper” issues growing for the last 50 years. “Nature abhors a vacuum,” proclaimed Aristotle more than 2,300 years ago. Today physicists are discovering that this is true in ways the ancient Greeks could never have imagined.

True, the cosmos consists overwhelmingly of vacuum. Yet vacuum itself is proving not to be empty at all. It is much more complex than most people would guess. “But surely,” you might ask, “if you take a container and remove everything from inside it – every atom, every photon – there will be nothing left?” Not by a long shot. Since the 1920s physicists have recognized that on a microscopic scale, the vacuum itself is alive with activity. Moreover, this network of activity may extend right down to include the very structure of space-time itself. The fine structure of the vacuum may ultimately hold the keys to some of the deepest questions facing physics – from why elementary particles have the properties they do, to the cause of the Big Bang and the likelihood of other universes outside our own.

THINGS THAT GO BUMP IN THE DARK

The state of the art in physics – our deepest current understanding of the world – is embodied in the so-called Standard Model, in which all matter and forces are accounted for by an astonishingly few types of particles (see Sky & Telescope – December 1987, page 582). Six quarks and six leptons make up all possible forms of matter. In practice just two of the quarks (the up and down) and one lepton (the electron) account for everything in the world except for a few whiffs of exotica known only to high-energy physicists. The 12 particles of matter (and their 12 corresponding particles of antimatter, or antiparticles) are acted upon by “messenger particles” that carry all the known forces. The photon mediates the electromagnetic force, including all the familiar chemical and structural forces around us on Earth. The members of the gluon family carry the strong force that binds neutrons and protons together in atomic nuclei. The W’, W-, and Zo mediate the weak nuclear force, and the as-yet-undiscovered graviton is believed to carry the force of gravity.

Every possible event involving the 12 matter particles can be completely explained as an exchange of messenger particles. During some of these events, for example when electrons accelerate in a radio-transmitter antenna, messenger particles (in this case photons) materialize and travel through space. At other times, however, the messengers remain almost entirely hidden within the interacting system. When the messengers exist in this hidden form, they are called “virtual particles.” Virtual particles may seem ghostly and unreal by everyday standards. But real they are. Moreover, they are not limited to their role of mediating interactions. Virtual particles can also pop in and out of empty space all by themselves.

Quantum mechanics, the rulebook of the Standard Model, states as a bedrock principle that you need a certain length of time to measure a particle’s energy or mass to a given degree of accuracy. The shorter the observation time, the more uncertain the measurement. If the time is very brief, the uncertainty becomes larger than the particie’s entire mass, and you cannot say whether or not the particle is there at all. The lighter the particle, the longer its uncertainty time. In the case of an electron-positron pair, the uncertainty time scale is about 10^-21″ seconds.

On time scales shorter than this, virtual electrons and positrons can, and do, pop in and out of nothingness like peas in a shell game. It’s as if, just because you can’t say a particle doesn’t exist when you look very briefly, then in a sense it does. This is not mere theorizing. In 1958 a tabletop experiment demonstrated the “Casimir effect,” measuring the force caused by virtual particles appearing and vanishing in total vacuum through the attraction they caused between two parallel metal plates. If the vacuum were truly empty the plates should not have attracted, but the incessant dance of virtual particles in the space between them produces a detectable effect.

Every particle – matter as well as messenger – seems to display a virtual form, each seething in greater or lesser abundances in what physicists call the “physical vacuum.” When it comes to affecting the ordinary world, moreover, virtual particles may do much more than just mediate forces. Some, in fact, may cause matter to have the property we call mass. The electron is the simplest of matter particles. Our knowledge of the physical world rests upon a solid understanding of its properties. Yet despite its abundance in the circuitry around us, the electron harbors an enigma. The fact that it has mass cannot be explained in the Standard Model, at least the parts of it that have been experimentally verified. More than 30 years ago particle physicist Peter Higgs suggested that the existence of mass has to do with a new ingredient of nature that is now called the Higgs field, which provides a new type of messenger particle that interacts with the electron to make it “weigh.”

The Higgs field has yet to be discovered, but many physicists expect it to exist everywhere in the physical vacuum, ensuring through its interactions with electrons and other particles that they will display mass. Even now, particle accelerators at CERN in Switzerland and at Fermilab near Chicago are straining at their maximum capabilities to cause just one “Higgs boson,” the presumed messenger particle for this field, to break loose from the vacuum and leave a detectable trace. Success would provide a triumphant completion of the Standard Model.

So to answer our question about whether a container of empty space is truly empty, the best anyone can do is remove the normal, physical particles that nature allows us to see and manipulate. The virtual particles can never be evicted. And in addition there may exist the ever-present Higgs field.

QUANTUM GRAVITY

For most of this century, physicists have struggled to bring gravity into the scheme of forces that are mediated by virtual messenger particles. To put this another way, the theory of general relativity, which shows the force of gravity to be a curvature of space-time, needs to be integrated with quantum mechanics, which shows forces to be virtual particle exchanges. Working on the assumption that such a marriage is possible, physicists named gravity’s messenger particle the graviton. But general relativity requires that gravitons be more than just quanta of gravity. In essence, gravitons define the structure of space-time itself.

The reconciliation of quantum mechanics and general relativity may lead us to dramatically new notions of the nature of space and time. Some theorists have suggested that points in space-time become defined only when a particle (such as a graviton or photon) interacts with other particles. In this view, what they are doing between interactions is a nonphysical question, since only an interaction defines a measurable time and place. Gravitational forces (and thus gravitons) exert an influence at distances much larger than the subatomic realm, as anyone who has fallen down a flight of stairs can attest. But only at an extremely small scale – the Planck length of 10^-33 centimeters – does the quantum nature of gravity become important.

Suppose you could magically look through a microscope that magnified an atomic nucleus to be some 10 light-years across. Under this magnification the smallest gravitons – that is, the most energetic and massive ones – would be about a millimeter in size. Here we might see a strange world in which space-time itself was defined by gravitons intersecting and looping around each other. In a similar vein, Roger Penrose has suggested that the gravitational field and space-time are built up from still more primitive mathematical entities called twistors, and that “ultimately the [space-time] concept may possibly be eliminated from the basis of physical theory altogether.” In essence, space and time become factored out as less- than-fundamental parts of the physical world.

In such a view, only the interactions between twistors, or perhaps gravitons, define when and where space-time is and is not. Are there gaps in the physical vacuum, voids of true and absolute nothing where space and time themselves do not exist?

Another viewpoint on the structure of space-time is offered by “superstring theory.” String theories posit that the fundamental objects of nature are one-dimensional lines rather than points; the “elementary” particles we measure are only oscillations of these strings. Superstring theory only seems to work, however, if space-time has not just four dimensions (three of space and one of time), but 10 dimensions. This hardly seems like the world we live in. To hide the extra six dimensions, mathematicians roll them up into conceptual corners that go by such cryptic names as “Calabi-Yau manifolds” and “orbifold space.” A recent textbook on the subject concludes on a wistful note that “if the string idea is correct, we may never catch more than a glimpse of the full ex- tent of reality.”

More recently, theorists Carlo Rovelli (University of Pittsburgh) and Lee Smolin (Pennsylvania State University) completed their analysis of a quantum gravity model developed by Abhay Ashtekar at Syracuse University in 1985. Unlike string theory, Ashtekar’s work applies only to gravity. However, it posits that at the Planck scale, space-time dissolves into a network of “loops” that are held together by knots. Somewhat like a chain-mail coat used by knights of yore, space-time resembles a fabric fashioned in four dimensions from these tiny one-dimensional loops and knots of energy.

Is this the way the world really is on its most fundamental level, or have mathematicians become detached from reality? Superstring theory has enticed physicists for over a decade now because it hints at a super unification of all four fundamental forces of nature. But it remains frustratingly hard to plant anchors down from these cloud castles into the real world of observation and experiment. The famous remark that superstring theory is “a piece of 21st- century physics that accidentally fell into the 20th century” captures both the excitement and frustration of workers stuck with 20th-century tools.

Surprisingly, string theory, Ashtekar’s loopy space-time, and twistors are not entirely independent ways of looking at space-time. In 1986 theorists discovered that superstrings have some things in common with twistors. A deep connection had been uncovered between two very different, independent theories. Like two teams of tunnelers starting on opposite sides of a mountain, they had met at the middle – a sign, perhaps, that they are dealing with a single real mountain, not separate ones in their own imaginations. And in 1995 Rovelli and Smolin also found that their graviton loops are very closely related to both the twistors and superstrings, though not identical in all respects.

THE COSMIC CONNECTION

Space-time could be strange in other ways too. Theorist John A. Wheeler (In- stitute for Advanced Study) has long advocated that at the Planck scale, space-time has a complex shape that changes from instant to instant. Wheeler called his picture “space-time foam” – a sea of quantum black holes and worm holes appearing and vanishing on a time scale of about 10^-43″ seconds. This is the Planck time, the time it takes light to cross the Planck length. Shorter than that, time, like space, presumably cannot exist – or, at least, our everyday notions of them cease to be valid.

Wheeler’s idea of space-time foam is a natural extrapolation from the idea of virtual particles. According to quantum mechanics, the higher the energy and mass of a particle, the smaller it must appear. A virtual particle as small as 10^-33″ cm, lasting only 10^-43 second, has so great a mass (10^-5 gram) in such a tiny volume that its own surface gravity would give it an escape velocity greater than the speed of light. In other words, it is a tiny black hole. But a black hole is not an ordinary object sitting in space- time like a particle; it is a structure of distorted, convoluted space-time itself. Although the consequences of such phe- nomena are not understood, it is rea- sonable to assume that these virtual par- ticles dramatically distort all space-time at the Planck scale.

If we take this reasoning at face value, and consider the decades-old experiments proving that the virtual particle phenomenon in a vacuum is real, it is hard to believe that space-time is smooth at or below the Planck scale. Space must be broken up and quantized. The only question is how. Wheeler’s original idea of space-time foam is especially potent because according to recent proposals by Sidney Coleman (Harvard) and Stephen Hawking (Cambridge University), its worm holes not only connect different points very close together within our space-time, but connect our space-time to other universes that, as far as we are concerned, exist only as ghostly probabilities. These connections to other universes cause the so-called cosmological constant – an annoying intrusion into the equations of cosmology ever since Einstein (see Sky & Telescope- April 1991, page 362) – to neatly vanish within our own universe.

Space-time foam has also been implicated as the spawning ground for baby universes. In several theories explaining the cause of the Big Bang and what came before, big bangs can bud off from a previously existing space-time, break away completely while still microscopic, and inflate with matter to become new universes of their own, completely disconnected (“disjoint”) from their space- time of origin. This process, proposed by Alan Guth (MIT) and others, gives a handle on what many expect to be another key issue of 21st-century physics: was our Big Bang unique? Or was it just a routine spinoff of natural processes happening all the time in some larger, outside realm? (see Sky & Telescope- September 1988, page 253).

Yet there are problems. The amount of latent energy in the quantum fluctuations of space-time foam is staggering: 10^105 ergs per cubic centimeter. This amounts to 10 billion billion times the mass of all the galaxies in the observ- able universe – packed into every cubic centimeter! Fortunately, Mother Nature seems to have devised some means of exactly canceling out this phenomenon to an accuracy of about 120 decimal places. The problem is that we haven’t a clue how.

It’s unnerving to think that in the 16 inches separating this page from your eyes, new big bangs are perhaps being spawned out and away from our quiet space-time every instant. By comparison, it seems positively dull that the photons by which you see this page might be playing a hop-scotch game to avoid gaps where space-time doesn’t exist.

REALITY CHECK

Some physicists have begun to throw cold water on these fantastic ideas. For instance, in 1993 Matt Visser (Washington University) studied the mathematical properties of quantum worm holes and discovered that, once they are formed, they become stable: they can’t foam at all. Kazuo Ghoroku (Fukuoka Institute of Technology, Japan) also found that quantum worm holes become stable even when their interactions with other fields are considered. What Wheeler called space-time foam may be something else entirely.

Among the unresolved problems facing theorists is the nature of time, which has been recognized as inextricably bound up with space ever since Einstein posited a constant speed for light. In general relativity, it isn’t always obvious how to define what we mean by time, especially at the Planck scale where time seems to lose its conventional meaning. Central to any quantum theory is the concept of measurement, but what does this imply for physics at the Planck scale, which sets an ultimate limit to the possibility of measurement? How any of these ideas about space- time can be tested is currently unknown. Some physicists believe this makes these ideas not real scientific inquiry at all. And it’s worth remembering that mathematics can sometimes introduce concepts that are only a means to an end and have no independent reality.

In the abstract world of mathematical symbolism, it isn’t always clear what is real and what’s not. For example, when we do long division on paper to divide 54,162 by 2 to get 27,081, we generate the intermediate numbers 14, 16, and 2, which we then just throw away. Are virtual particles, compact 6-dimensional manifolds, and twistors simply nonphysical means to an end – mere artifacts of how we humans do our mathematics? Particle physicists often have to deal with “ghost fields” that are simply the temporary scaffolding used for calculations, and that vanish when the calculations are complete. Nonphysical devices such as negative probability and faster- than-light tachyon particles are grudgingly tolerated so long as they disappear before the final answers. Even in super- string theory, recent work suggests that it may be possible to build consistent models entirely within ordinary four-dimensional space-time, without recourse to higher dimensions.

ANGEL FOOD CAKE

So, how should we think of the great, dark void that we gaze into at night? All clues point to space-time being a kind of layer cake of busy phenomena on the submicroscopic scale. The topmost layer contains the quarks and electrons comprising ordinary matter, scattered here and there like raisins in the frosting. These raisins can be plucked away to make a region of space appear empty. The frosting itself consists of virtual particles, primarily those carrying the electromagnetic, weak, and strong forces, filling the vacuum with incessant activity that can never be switched off. Their quantum comings and goings may completely fill space-time so that no points are ever really missing. This layer of the cake of “empty space” seems pretty well established by laboratory experiment.

Beneath this layer we have the domain of the putative Higgs field. No matter where the electron and quark “raisins” go, in this view, there is always a piece of the Higgs field nearby to affect them and give them mass. Below the Higgs layer there may exist other layers, representing fields we have yet to discover. But eventually we arrive at the lowest stratum, that of the gravitational field. There is more of this field wherever mass is present in the layers above it, but there is no place where it is entirely absent. This layer recalls the Babylonian Great Turtle that carried the universe on its back. Without it, all the other layers above would vanish into nothingness.

We know that space-time is quite smooth down to at least the scale of the electron, 10^-20 cm – 10 million times smaller than an atomic nucleus. This is the size limit set for any internal component of the electron, based on careful comparisons between experiment and the predictions of quantum electrodynamics. But near the Planck horizon of 10^-33 cm, space-time must change its structure drastically. It may be a world in which conventional notions of dimensionality, time, and space need to be redefined and possibly eliminated altogether.

The conceit of our universe’s uniqueness may disappear, with big bangs becoming viewed as run-of-the-mill events in some much larger outside realm, and with physical constants being attributed to causes in space-times forever beyond human experience.

There is much that’s spooky about the physical vacuum. This spookiness may be rooted more in the way our brains work than in some objective aspect of nature. Einstein stressed, “Space and time are not conditions in which we live, but modes in which we think.” Our understanding of space remains in its infancy. With Aristotle smiling at us down the centuries, we now see the vacuum as much more than a vacancy. It will take many decades, if not centuries, before a complete understanding of it is fashioned. In the meantime, enjoy the nighttime view!

FURTHER READING

Davies, Paul. The New Physics. Cambridge: Cambridge University Press, 1989.

Mallove, Eugene. “The Self-Reproducing Universe.” Sky & Telescope, September 1988, page 253.

Matthews, Robert. “Nothing Like a Vacuum.” New Scientist, February 25, 1995, page 30.

Pagels, Heinz. “Perfect Symmetry.” New York: Simon & Schuster, 1985.

The Next Sunspot Cycle

Forecasters are already starting to make predictions for what might be in store as our sun winds-down its current sunspot cycle (Number 24) in a few years. Are we in for a very intense cycle of solar activity, or the beginning of a century-long absence of sunspots and a rise in colder climates?

Figure showing the sunspot counts for the past few cycles. (Credit:www.solen.info)

Ever since Samuel Schwabe discovered the 11-year ebb and flow of sunspots on the sun in 1843, predicting when the next sunspot cycle will appear, and how strong it will be, has been a cottage industry among scientists and non-scientists alike. For solar physicists, the sunspot cycle is a major indicator of how the sun’s magnetic field is generated, and the evolution of various patterns of plasma circulation near the solar surface and interior. Getting these forecasts bang-on would be proof that we indeed have a ‘deep’ understanding of how the sun works that is a major step beyond just knowing it is a massive sphere of plasma heated by thermonuclear fusion in its core.

So how are we doing?

For over a century, scientists have scrutinized the shapes of dozens of individual sunspot cycles to glean features that could be used for predicting the circumstances of the next one. Basically, we know that 11-years is an average and some cycles are as short as 9 years or as long as 14. The number of sunspots during the peak year, called sunspot maximum, can vary from as few as 50 to as many as 260. The speed with which sunspot numbers rise to a maximum can be as long as 80 months for weaker sunspot cycles, and as short as 40 months for the stronger cycles. All of these features, and many other statistical rules-of-thumb, lead to predictive schemes of one kind or another, but they generally fail to produce accurate and detailed forecasts of the ‘next’ sunspot cycle.

Prior to the current sunspot cycle (Number 24), which spans the years 2008-2019, NASA astronomer Dean Pesnell collected 105 forecasts for Cycle 24 . For something as simple as how many sunspots would be present during the peak year, the predictions varied from as few as 40 to as many as 175 with an average of 106 +/-31. The actual number at the 2014 peak was 116. Most of the predictions were based on little more than extrapolating statistical patterns in older data. What we really want are forecasts that are based upon the actual physics of sunspot formation, not statistics. The most promising physics-based models we have today actually follow magnetic processes on the surface of the sun and below and are called Flux Transport Dynamo models.

Solar polar magnetic field trends (Credit: Wilcox Solar Observatory)

The sun’s magnetic field is much more fluid than the magnetic field of a toy bar magnet. Thanks to the revolutionary work by helioseismologists using the SOHO spacecraft and the ground-based GONG program, we can now see below the turbulent surface of the sun. There are vast rivers of plasma wider than a dozen Earths, which wrap around the sun from east to west. There is also a flow pattern that runs north and south from the equator to each pole. This meridional current is caused by giant convection cells below the solar surface and acts like a conveyor belt for the surface magnetic fields in each hemisphere. The sun’s north and south magnetic fields can be thought of as waves of magnetism that flow at about 60 feet/second from the equator at sunspot maximum to the poles at sunspot minimum, and back again to the equator at the base of the convection cell. At sunspot minimum they are equal and opposite in intensity at the poles, but at sunspot maximum they vanish at the poles and combine and cancel at the sun’s equator. The difference in the polar waves during sunspot minimum seems to predict how strong the next sunspot maximum will be about 6 years later as the current returns the field to the equator at the peak of the next cycle. V.V Zharkova at Northumbria University in the UK uses this to predict that Cycle 25 might continue the declining trend of polar field decrease seen in the last three sunspot cycles, and be even weaker than Cycle 24 with far fewer than 100 spots. However, a recent paper by NASA solar physicists David Hathaway and Lisa Upton  re-assessed the trends in the polar fields and predict that the average strength of the polar fields near the end of Cycle 24 will be similar to that measured near the end of Cycle 23, indicating that Cycle 25 will be similar in strength to the current cycle.

But some studies such as those by Matthew Penn and William Livingston at the National Solar Observatory seem to suggest that  sunspot magnetic field strengths have been declining since about 2000 and are already close to the minimum needed to sustain sunspots on the solar surface.  By Cycle 25 or 26, magnetic fields may be too weak to punch through the solar surface and form recognizable sunspots at all, spelling the end of the sunspot cycle phenomenon, and the start of another Maunder Minimum cooling period perhaps lasting until 2100. A quick GOOGLE search will turn up a variety of pages claiming that a new ‘Maunder Minimum’ and mini-Ice Age are just around the corner! An interesting on-the-spot assessment of these disturbing predictions was offered back in 2011 by NASA solar physicist C. Alex Young, concluding from the published evidence that these conclusions were probably ‘Much Ado about Nothing’.

What can we bank on?

The weight of history is a compelling guide, which teaches us that tomorrow will be very much like yesterday. Statistically speaking, the current Cycle 24 is scheduled to draw to a close about 11 years after the previous sunspot minimum in January 2008, which means sometime in 2019. You can eyeball the figure at the top of this blog and see that that is about right. We entered the Cycle 24 sunspot minimum period in 2016 because in February and June, we already had two spot-free days. As the number of spot-free days continues to increase in 2017-2018, we will start seeing the new sunspots of Cycle 25 appear sometime in late-2019. Sunspot maximum is likely to occur in 2024, with most forecasts predicting about half as many sunspots as in Cycle 24.

None of the current forecasts suggest Cycle 25 will be entirely absent. A few forecasts even hold out some hope that a sunspot maximum equal to or greater than Cycle 24 which was near 140 is possible, while others place the peak closer to 60 in 2025.

It seems to be a pretty sure bet that there will be yet-another sunspot cycle to follow the current one. If you are an aurora watcher, 2022-2027 would be the best years to go hunting for them. If you are a satellite operator or astronaut, this next cycle may be even less hostile than Cycle 24 was, or at least no worse!

In any event, solar cycle prediction will be a rising challenge in the next few years as scientists pursue the Holy Grail of creating a reliable theory of why the sun even has such cycles in the first place!

Check back here on Friday, February 17 for my next blog!