Skip to content
Starts With A Bang

The big theoretical problem of dark energy

The zero-point energy of empty space is not zero. Even with all the physics we know, we have no idea how to calculate what it ought to be.
regions of the universe
In a Universe that comes to be dominated by dark energy, there are four regions: one where everything within it is reachable and observable, one where everything is observable but unreachable, one where things will someday be observable but aren't today, and one where things will never be observable. The labeled numbers correspond to our consensus cosmology as of 2024, with boundaries of 18 billion light-years, 46 billion light-years, and 61 billion light-years separating the four regions.
Credit: Andrew Z. Colvin/Wikimedia Commons; annotations: E. Siegel
Key Takeaways
  • Here in our expanding Universe, ultra-distant objects aren’t just speeding away from us, the rate at which they’re speeding away is increasing: teaching us that the Universe is accelerating.
  • When we examine how the Universe is accelerating, we find that it’s behaving as though the Universe is filled with some sort of energy inherent to space: dark energy, or a cosmological constant.
  • But theoretically, we have no idea how to calculate what the value of dark energy ought to be. Its extremely small but non-zero value remains a tremendous puzzle in fundamental physics.
Sign up for Smart Faster newsletter
The most counterintuitive, surprising, and impactful new stories delivered to your inbox every Thursday.

One of the most fundamental questions we can ask about our Universe itself is “What makes it up?” For a long time, the answer seemed obvious: matter and radiation. We observe them in great abundances, everywhere and at all times throughout our cosmic history. For some ~100 years, we’ve recognized that — consistent with General Relativity — our Universe is expanding, and the way that the Universe expands is determined by all the forms of matter and radiation within it. Ever since we realized this, we’ve strived to measure how fast the Universe is expanding and how that expansion has changed over our cosmic history, as knowing both would determine the contents of our Universe.

In the 1990s, observations finally became good enough to reveal the answer: yes, the Universe contains matter and radiation, as about 30% of the Universe is made of matter (normal and dark, combined) and about ~0.01% is radiation, today. But surprisingly, about 70% of the Universe is neither of these, but rather a form of energy that behaves as though it’s inherent to space: dark energy. The way this dark energy behaves is identical to how we’d expect either a cosmological constant (in General Relativity) or the zero-point energy of space (in Quantum Field Theory) to behave. But theoretically, this is an absolute nightmare. Here’s what everyone should know.

feynman diagrams
Today, Feynman diagrams are used in calculating every fundamental interaction spanning the weak and electromagnetic forces, including in high-energy and low-temperature/condensed conditions. Including higher-order “loop” diagrams leads to more refined, more accurate approximations of the true value to quantities in our Universe. However, the strong interactions cannot be computed in this fashion, and must either be subject to non-perturbative computer calculations (Lattice QCD) or require experimental inputs (the R-ratio method) in order to account for their contributions.
Credit: V. S. de Carvalho and H. Freire, Nucl. Phys. B, 2013

From a quantum point-of-view, the way we picture our Universe is that real particles (quanta) exist atop the fabric of spacetime, and that they interact with one another through the exchange of (virtual) particles. We draw out diagrams that represent all the possible interactions that can occur between particles — Feynman diagrams — and then calculate how each such diagram contributes to the overall interaction between the multiple quanta in question. When we sum up the diagrams in increasing order of complexity — tree diagrams, one-loop diagrams, two-loop diagrams, etc. — we arrive at closer and closer approximations to our actual physical reality.

But there are other diagrams as well that we can draw out: diagrams that don’t correspond to incoming-and-outgoing particles, but diagrams that represent the “field fluctuations” that occur in empty space itself. Just like in the case of real particles, we can write down and calculate diagrams of ever-increasing complexity, and then sum up what we get to approximate the real value of the zero-point energy: or the energy inherent to empty space itself.

Of course, there are a truly infinite number of terms, but whether we calculate the first one, the first few, or the first several terms, we find that they all give extremely large contributions: contributions that are too large to be consistent with the observed Universe by over 120 orders of magnitude. (That is, a factor of over 10120.)

zero point energy contributions
A few terms contributing to the zero-point energy in quantum electrodynamics. The development of this theory, due to Feynman, Schwinger, and Tomonaga, led to them being awarded the Nobel Prize in 1965. These diagrams may make it appear as though particles and antiparticles are popping in and out of existence, but that is only a calculational tool; these particles are virtual, not real.
Credit: R. L. Jaffe, Phys. Rev. D, 2005

In general, whenever you have two large numbers and you take the difference of them, you’re going to get another large number as well. For example, imagine the net worths of two random people on one of the world’s “billionaires” lists, person A and person B. Maybe person A is worth $3.8 billion and maybe person B is worth $1.6 billion, and therefore the difference between them would be ~$2.2 billion: a large number indeed. You can imagine a scenario where the two people you randomly picked are worth almost exactly the same amount, but these instances usually only occur when there’s some relationship between the two: like they cofounded the same company or happen to be identical twins with one another.

In general, if you have two numbers that are both large, “A” and “B,” then the difference between those numbers, |A – B|, is also going to be large. Only if there’s some sort of reason — an underlying symmetry, for example, or an underlying relationship between them, or some mechanism that’s responsible for those two numbers to nearly-perfectly match — will the difference between those numbers, |A – B|, turn out to be very small compared to “A” and “B” themselves.

The alternative explanation is that these two numbers really are very close together, but entirely coincidentally: something that’s more and more unlikely the closer these two values are to each other.

unstable equilibrium
When we see something like a ball balanced precariously atop a hill, this appears to be what we call a finely-tuned state, or a state of unstable equilibrium. A much more stable position is for the ball to be down somewhere at the bottom of the valley. Whenever we encounter a finely-tuned physical situation, there are good reasons to seek a physically-motivated explanation for it; when we have hills with false minima on them, it’s possible to get caught up in one and not arrive at the “true” minimum.
Credit: L. Albarez-Gaume & J. Ellis, Nature Physics, 2011

When we attempt to calculate, using quantum field theory, the expected value of the zero-point energy of empty space, the individual terms that contribute do so with values that are proportional to a combination of fundamental constants — √(ℏc/G) — raised to the fourth power. That combination of constants is also known as the Planck mass, and has a value that’s equivalent to ~1028 eV (electron-volts) of energy when you remember that E = mc². When you raise that value to the fourth power and keep it in terms of energy, you get a value of 10112 eV4, and you get that value distributed over some region of space.

Now, in our real Universe, we actually measure the dark energy density cosmologically: by inferring what value it needs to have in order to give the Universe its observed expansion properties. The equations that we use to describe the expanding Universe allow us to translate the “energy value” from above into an energy density (an energy value over a specific volume of space), which we can then compare to the actual, observed dark energy value. Instead of 10112 eV4, we get a value that’s more like 10-10 or 10-11 eV4, which corresponds to that mismatch of more than 120 orders of magnitude mentioned earlier.

Friedmann equation
The relative importance of different energy components in the Universe at various times in the past. Note that when dark energy reaches a number near 100% in the future, the energy density of the Universe (and, therefore, the expansion rate) will remain constant arbitrarily far ahead in time. Owing to dark energy, distant galaxies are already speeding up in their apparent recession speed from us. Way off the scale of this diagram, to the left, is when the inflationary epoch ended and the hot Big Bang began. Dark energy’s energy density is ~123 orders of magnitude lower than the theoretical expectation.
Credit: E. Siegel

For many decades, people have noted this property of the Universe: that our predicted value of the zero-point energy of space is nonsensical. If it were correct, the expanding Universe would have either recollapsed or expanded into empty nothingness extremely early on: before the electroweak symmetry broke and particles even received a non-zero rest mass, much less before atoms, nuclei, or even protons and neutrons could form. We knew that “prediction” must be wrong, but which of the following reasons explained why?

  1. The sum of all of these terms, even though they’re individually large, will somehow exactly cancel, and so the real value of the zero-point energy of space is truly zero.
  2. The actual value of the zero-point energy of space takes on all possible values, randomly, and then only in locations where its value admits our existence can we arise to observe it.
  3. Or this is a calculable entity, and if we could calculate it properly, we’d discover an almost-exact but only approximate cancellation, and hence the real value of the zero-point energy is small but non-zero.

Of these options, the first is just a hunch that can’t explain the actual dark energy in the Universe, while the second basically gives up on a scientific approach to the question. Regardless of the answer, we still need to rise to the challenge of figuring out how to calculate the actual zero-point energy of empty space itself.

quantum gravity
Quantum gravity attempts to combine Einstein’s general theory of relativity with quantum mechanics. Quantum corrections to classical gravity are visualized as loop diagrams, as the one shown here in white. Alternatively, it’s possible that gravity is always classical and continuous, and that quantum field theory, not general relativity, needs to be modified.
Credit: SLAC National Accelerator Laboratory

If you’re a physicist, you might imagine that there’s some sort of miraculous cancellation of most of the possible contributions to the zero-point energy, but that a few contributions remained and do not have an equal-and-opposite contribution to cancel them out. Perhaps the contributions of all the quarks and antiquarks cancel. Perhaps the contributions of all the charged leptons (electron, muon, and tau) cancel with their antiparticle partners, and perhaps only the remaining, “uncanceled” contributions actually account for the dark energy that exists in the Universe.

If we imagine that there’s some sort of partial cancellation that occurs, what would we need to remain, left over, in order to explain the (relatively tiny) amount of dark energy that’s present in the Universe?

The answer is surprising: something that corresponds to an energy scale of only a fraction of an electron-volt, or somewhere between 0.001 and 0.01 eV. What sort of particles have a rest mass that’s the equivalent of that particular energy value? Believe it or not, we have some right here in the standard model: neutrinos.

standard model color
The particles and antiparticles of the Standard Model have now all been directly detected, with the last holdout, the Higgs boson, falling at the LHC earlier this decade. Today, only the gluons and photons are massless; everything else has a non-zero rest mass.
Credit: E. Siegel/Beyond the Galaxy

As originally formulated, the Standard Model would have all of the quarks be massive, along with the charged leptons, the W-and-Z bosons, and the Higgs boson. The other particles — neutrinos and antineutrinos, the photon, and the gluons — would all be massless. In the aftermath of the hot Big Bang, in addition to the normal matter particles (protons, neutrons, and electrons) that are produced, enormous numbers of neutrinos, antineutrinos, and photons are produced: about ~1 billion of them, each, for each and every proton that survives.

As it actually turns out, as we first suspected in the 1960s and then puzzled out in the 1990s and early 2000s, neutrinos aren’t massless at all. Rather, the species of neutrino or antineutrino (electron, muon, or tau) that’s produced, initially, isn’t always the species of neutrino you observe later on. Whether passing through the vacuum of space or whether passing through matter, neutrinos have a non-zero probability of changing their flavor, which can only occur if they have mass. (Otherwise, as massless particles, they wouldn’t experience time, and so would have no period-of-oscillation.) The fact that neutrinos have mass means, necessarily, that there’s some property about them that the original formulation of the Standard Model doesn’t account for.

neutrino oscillation
Vacuum oscillation probabilities for electron (black), muon (blue), and tau (red) neutrinos for a chosen set of mixing parameters, beginning from an initially produced electron neutrino. An accurate measurement of the mixing probabilities over different length baselines can help us understand the physics behind neutrino oscillations and could reveal the existence of any other types of particles that couple to the three known species of neutrino. For neutrinos to oscillate, they must have non-zero mass. If additional particles (such as dark matter particles) carry energy away, the overall neutrino flux will show a deficit.
Credit: Strait/Wikimedia Commons

Since we don’t know what, exactly, gives neutrinos these non-zero rest masses, we have to be very careful that we don’t prematurely rule out a scenario that connects their mass scales to the “energy scale” of the observed dark energy that appears in the Universe. Many have suggested plausible mechanisms for such a coupling, but no one has yet solved the hard problem of, “How do we calculate the zero-point energy of space using quantum field theory and the quantum fields that we know exist within our Universe?” We can measure the actual value of dark energy, but as far as understanding the theoretical side of the equation, we can only state, “We do not.”

Another aspect of the story that needs to be included is the fact that, prior to the start of the hot Big Bang, our Universe underwent a separate, earlier period where the Universe was expanding as though we had a positive, finite value to the zero-point energy of space: cosmological inflation. During inflation, however, the energy was much larger than the value it has today, but still not as large as the expected Planck-energy-range values. Instead, the energy scale of inflation is somewhere below ~1025 eV and could have potentially been as low as ~1014 eV: much much larger than today’s value but still much smaller than the value we would’ve naively expected.

axion
This 2018 plot shows the exclusion limits on axion abundances and couplings, under the assumption that axions make up ~100% of the dark matter within the Milky Way. Both the KSVZ and DFSZ axion exclusion limits are shown. Note that if the axion mass is used to calibrate the “energy scale” expected for dark energy, it makes a suggestive candidate.
(Credit: N. Du et al. (ADMX Collaboration) Phys. Rev. Lett., 2018)

Additionally, because there must be some sort of dark matter in the Universe — some particle that isn’t a part of the Standard Model — many have wondered if there couldn’t be some connection between whatever particle is responsible for dark matter with whatever energy scale is responsible for dark energy. One particle that’s a candidate for dark matter, the axion, typically comes in with very low masses that are below ~1 eV but that must be greater than about ~0.00001 eV (a micro-electron-volt), which places it right in the range where it would be very interestingly suggestive for a connection to dark energy.

But the hard problem still remains, and remains unsolved: how do we know, or calculate, what the zero-point energy of empty space actually is, according to our field theories?

That’s something we absolutely must learn how to do. We have to learn how to do this calculation, otherwise we don’t have a good theoretical understanding behind what is or isn’t causing dark energy. And the fact is we don’t know how to do it; we can only “assume it’s all zero” except for some non-zero part. Even when we do that, we have yet to discover why the “mass/energy scale” of dark energy takes on only this low-but-non-zero value any value seems possible. It must make us wonder: are we even looking at the problem correctly?

big crunch
The far distant fates of the Universe offer a number of possibilities, but if dark energy is truly a constant, as the data indicates, it will continue to follow the red curve, leading to the long-term scenario frequently described on Starts With A Bang: of the eventual heat death of the Universe. The Universe was decelerating for the first ~7.8 billion years of cosmic history, but transitioned to accelerating about ~6 billion years ago. If dark energy doesn’t remain constant but rather evolves with time, a Big Rip or a Big Crunch are still admissible, but we don’t have any evidence indicating that this evolution is anything more than idle speculation.
Credit: NASA/CXC/M. Weiss

But there is a great set of reasons to be hopeful: observationally, we’re making tremendous progress. 20 years ago, we thought that dark energy behaves as the zero-point energy of empty space, but our uncertainties on it were something like ~50%. By 15 years ago, the uncertainties were down to about ~25%. Now, they’re down at around ~7%, and with upcoming missions such as ESA’s Euclid, the NSF’s ground-based Vera Rubin Observatory, and NASA’s upcoming Nancy Grace Roman Telescope scheduled to be our next flagship mission now that JWST has launched, we’re poised to constrain the equation-of-state of dark energy to within ~1%.

In addition, we’ll be able to measure whether the dark energy density has changed over cosmic time, or whether it’s been a constant over the past ~8+ billion years. Based on the data we have today, it’s looking like dark energy is very much behaving as a constant: at all times and locations, and that it’s consistent with being the zero-point energy of empty space itself. However, if dark energy behaves differently from this in any way, the next generation of observatories should reveal that as well, with consequences for how we perceive of the fate of our Universe. Even when theory doesn’t pave the way to the next great breakthrough, improved experiments and observations always offer an opportunity to show us the Universe as we’ve never seen it before, and show us what secrets we might be missing!

Sign up for Smart Faster newsletter
The most counterintuitive, surprising, and impactful new stories delivered to your inbox every Thursday.

Related

Up Next