Resolving the Vacuum Catastrophe: A Generalized Holographic Approach ()
1. Introduction
The vacuum energy density predicted by quantum field theory disagrees with cosmological observation by approximately 122 orders of magnitude. It is one of the biggest disagreements between theory, experiment and observation and is known as the vacuum catastrophe [1] . To resolve this discrepancy, we first review the fundamental nature of the vacuum energy density and its relationship to the cosmological constant.
The Einstein field equations of general relativity include a constant
known as the cosmological constant. Originally included to allow for static homogenous solutions to Einstein’s equations, it was subsequently removed when the expansion of the universe was discovered [2] . However, since then the universe was found to be accelerating [3] and many cosmological models have been put forward with a nonzero
e.g. de Sitter, steady state and the Lemaitre models, where
acts as an additional expanding (dark energy) force.
With the inclusion of the cosmological constant, Einstein’s field equations are:
(1)
where
is the Ricci curvature tensor,
is the metric tensor, r is the scalar curvature and
is the stress-energy tensor, which is modeled as a perfect fluid such that:
(2)
The Robertson-Walker solution, which states that the rest frame of the fluid must be the same as the co-moving observer, reduces the Einstein equations to two Friedman equations:
(3)
(4)
where a is the scaling factor, k is the curvature constant and
is the radius of the observable universe (i.e.
, where r is the co-moving radius).
Based on astronomical observations the current cosmological model states that we live in a flat,
dominated, homogeneous and isotropic universe, composed of radiation, baryonic matter and non-baryonic dark matter [3] - [8] .
The Friedman equation for a flat universe (i.e.
) is thus given in the form:
(5)
If we then take the assumption that the universe is pervaded by a form of energy (i.e. dark energy), which is the current consensus in both cosmology and particle physics [9] [10] [11] then the cosmological constant can be interpreted as an energy density [12] [13] and given in terms of the dark energy density,
. Note, this result can also be found by assuming a static universe (i.e.
).
In either case the Friedman equation thus takes the form:
(6)
Friedman’s solutions suggest that there is a critical density at which the universe must be flat, where the ratio of the total mass-energy density to the critical density is known as the density parameter
and is currently measured as
[6] [8] [14] .
The contributions to this density parameter come from: the vacuum density (dark energy),
; the dark matter,
; and the baryonic matter,
, totaling to
[14] .
The Friedman equation thus takes the form of an Einstein-de Sitter model in which the cosmological constant is coupled to the density:
(7)
where
is the density due to baryonic matter;
is the density due to dark matter;
is the density due to dark energy; and
.
Using the current value of
for Hubble’s constant [14] , gives the critical density at the present time as,
and thus
,
and
. The vacuum energy density at the cosmological scale is thus of the order
.
However, quantum field theory determines the vacuum energy density by summing the energies
over all oscillatory modes. See reference [1] for a more detailed overview. As quantum fluctuations predict infinite oscillatory modes [15] [16] this yields an infinite result unless renormalized at the Planck cutoff. In utilizing such a cutoff value, the vacuum energy density is found to be:
(8)
where
is the Planck mass and
is the Planck length. This value is well supported by both theory and experimental results [17] - [23] .
The cosmological vacuum energy density determined from observations,
, is therefore in disagreement with the vacuum energy density at the Planck cutoff, predicted by quantum field theory,
. This discrepancy is a significant 122 orders of magnitude and is thus known as the “vacuum catastrophe”.
Possible attempts to solve this discrepancy, as reviewed by Weinberg [24] , include introducing a scalar field coupled to gravity in such a way that
is automatically cancelled when the scalar field reaches equilibrium [25] . A second approach imagines a deep symmetry that isn’t apparent in the effective field theory but nevertheless constrains parameters of this effective theory so that
is zero or small [26] . Then there is the idea of quintessence which states that the cosmological constant is small because the universe is old and thus imagines a scalar field that rolls down a potential governed by a field equation [27] [28] [29] . When such a slowly varying scalar field is minimally coupled to gravity it can lead to the observed acceleration of the Universe [30] . This idea of quintessence has further been supported by the recent conjecture offered by Obied [31] to explain why string theory has not been able to construct a meta-stable de Sitter vacuum. They found that the resulting “allowed” universe points to an expanding universe in which the vacuum energy decreases at a rate above a specific lower limit i.e. a quintessent universe [31] [32] [33] .
Finally, anthropic considerations apply an anthropic bound on +ve
by setting the requirement that it should not be so large as to prevent the formation of galaxies [34] . Using a simple spherical in-fall model of Peebles [35] the upper bound gives
as being no larger than the cosmic mass density at the time of earliest galaxy formation (z = 5), which is approximately 200 times the present mass density and thus a big improvement from the 122 orders of magnitude. Therefore, as yet, the “vacuum catastrophe” is unresolved.
2. The Generalized Holographic Model
In previous work [36] [37] , a quantized solution to gravity is given in terms of Planck Spherical Units (PSU) in a generalized holographic approach. A brief description of this solution is given below.
Following the holographic principle of ‘t Hooft [38] , based on the Bekenstein-Hawking formulae for the entropy of a black hole [39] [40] , the surface and volume entropy of a spherical system is explored. The holographic bit of information is defined as an oscillating Planck spherical unit (PSU), given as,
(9)
where
.
These PSUs, or Planck “voxels”, tile along the area of a spherical surface horizon, producing a holographic relationship with the interior information mass-energy density (see Figure 1).
In this generalized holographic approach, it is therefore suggested that the information/entropy of a spherical surface horizon should be calculated in spherical bits and thus defines the surface information/entropy in terms of PSUs, such that,
(10)
where the Planck area, taken as one unit of information/entropy, is the equatorial disk of a Planck spherical unit,
and A is the surface area of a spherical system. We note that in this definition, the entropy is slightly greater (~ 5 times) than that set by the Bekenstein bound, and the proportionality constant is taken to be unity (instead of 1/4 as in the Bekenstein-hawking entropy). It has been previously suggested that the quantum entropy of a black hole may not exactly equal A/4 [41] . To differentiate between models, the information/entropy S, encoded on the surface boundary in the generalized holographic model is termed,
.
As first proposed by ‘t Hooft the holographic principle states that the description of a Volume of space can be encoded on its surface boundary, with one discrete degree of freedom per Planck area, which can be described as Boolean variables evolving with time [42] .
Following the definition for surface information
, the information/entropy within a volume of space is similarly defined in terms of PSU as,
(11)
where V is the volume of the spherical entity and r is its radius.
In previous work [36] [37] , it was demonstrated that the holographic relationship between the transfer energy potential of the surface information and the volume information, equates to the gravitational mass of the system. It was thus found that for any black hole of Schwarzschild radius
the mass
can be given as,
(12)
where
is the number of PSU on the spherical surface horizon and R is the number of PSU within the spherical volume. Hence, a holographic gravitational mass equivalence to the Schwarzschild solution is obtained in terms of a discrete granular structure of spacetime at the Planck scale, giving a quantized solution to gravity in terms of Planck spherical units (PSUs). It should be noted that this view of the interior structure of the black hole in terms of PSUs, is supported by the concept of black hole molecules and their relevant number densities as proposed by Miao and Xu [43] and Wei and Lui [44] . As well, the relationship between the interior structure in terms of “voxels” and the connecting horizon pixels is discussed in the work of Nicolini [45] .
Of course, these considerations lead to the exploration of the clustering of the structure of spacetime at the nucleonic scale, where it was found that a precise value for the mass
and charge radius
of a proton can be given as,
(13)
(14)
where
is defined as a fundamental holographic ratio. Significantly, this value is within an
agreement with the latest muonic measurements of the charge radius of the proton [36] [37] , relative to a
variance in the standard approach [46] .
3. Resolving the Vacuum Catastrophe
To resolve the vacuum catastrophe, we must first understand where the value for the vacuum energy density at the Planck scale is coming from. As was previously defined [36] [37] , and summarized above, the physical structure and thus energy density at this scale is more appropriately represented in terms of PSUs, such that the vacuum energy density at the Planck scale
, can be given as,
.
The vacuum energy density at the quantum scale is thus
instead of the value
given in Equation (8).
The generalized holographic model describes how any spherical body can be considered in terms of its PSU packing, or volume entropy, R. The mass-energy
, in terms of PSU, can therefore be given as
and the mass-energy density is given as,
.
In the case of the proton, the mass-energy in terms of Planck mass was calculated as
, which is equivalent to the mass of the observable universe (i.e.
in terms of the Eddington number; and
from density measurements). Since these values for the mass of the observable universe are just approximations, we will take the mass of the observable universe to be the mass-energy of the proton, as calculated above. The mass-energy density of the universe can thus be defined in terms of the mass-energy density of the proton. Thus, at the cosmological scale the mass-energy density, or vacuum energy density, is calculated to be,
(15)
where
and was found by taking
as the Hubble radius
. Thus, when the vacuum energy density of the Universe is considered in terms of the proton density and the protons PSU packing (i.e. its volume entropy, R) we find the density scales by a factor of 10122. As well, it should be noted that this value for the mass-energy density is found to be equivalent to the dark matter density,
.
Similarly, the vacuum energy density can be considered in terms of the PSU surface tiling (i.e. its surface entropy,
), as the radius expands from the Planck scale
to the cosmological scale. The vacuum density at the cosmological scale is thus given as,
(16)
where
is found by assuming a spherical shell Universe of radius
. The resulting change in density, from the vacuum density at the Planck scale to that at the cosmological scale yields an exact equivalent to the currently observed critical density of the universe,
. Thus, when we consider the generalized holographic approach, which describes how any spherical body can be considered in terms of its PSU packing, we show the scale relationship between the PSUs and a spherical shell universe and resolve the 122 orders of magnitude discrepancy between the vacuum energy density at the Planck scale and the vacuum energy density at the cosmological scale.
The solution presented here is in line with the ideas of quintessence in which the mass-energy density is governed by the scale factor
, such that
for
. Following this approach, the Friedman equation can then be written in the form:
(17)
which can also be given in terms of the varying radius, such that
for
and the Friedman equation becomes:
(18)
These findings are in agreement with those of Ali and Das [47] who, in an attempt to resolve the current problems of cosmology, interpret one of the quantum correction terms in the second order Friedman equation as dark energy. From the quantum corrected Raychaudhuri equations they find the first correction term
where
is identified as the current linear dimension of our observable universe, such that
in planck units.
Essentially, they are adding the correction term
whereas we include the scale factor
. However, their solution describes a purely quantum mechanical description of the universe assuming quantum gravity affects are practically absent, whereas the results described here show how, as the density changes with radius we have a scaler field that is coupled to gravity and thus rolls down a potential governed by a generalized holographic quantized solution to gravity [36] .
Similar scale-invariant models have also been proposed by Maeder [48] [49] [50] who much like Milgrom’s modified Newtonian dynamics (MOND) [51] [52] [53] defines a limit where scale invariance is applicable at large scales (i.e. low accelerations in MOND). In his model Maeder utilizes a new co-ordinate system, derived from scale invariant tensor analysis, and much like Milgrom and Verlinde [54] he finds an additional factor
that opposes gravity. Interestingly, and in line with our findings, Maeder notes that with this new co-ordinate system, both the pressure and density are not scale invariant.
It should as well be noted that the equivalence found between the critical density and that found from the surface entropy (Equation (16)) yields a critical mass that obeys the Schwarzschild solution for a universe with a radius of the Hubble radius,
(19)
The idea that the observable universe is the interior of a black hole was originally put forward by Pathria [55] and Good [56] and more recently by Poplowski [57] . If such a solution holds true, then this would give us the prefect opportunity to study the interior of a black hole.
4. Discussion
Previous attempts to resolve the vacuum catastrophe include large quantum corrections (e.g. [47] [58] ). However such theories offer no physical explanation and although solutions such as Zlatev [59] [60] do not depend on any fine tuning of the initial conditions, fine-tuning is still required to set the energy density of the scalar field to equal the energy density of matter and radiation at the present time i.e. at the cross-over from matter dominated to scalar field (or vacuum) dominated. This was the weak point of Hoyle’s [12] steady state universe, as although he was able to show expansion properties with the introduction of the space-time vector C, no physical explanation was proposed.
The solution described in this paper utilizes the generalized holographic approach [36] , offering a physical explanation which is thus inherent within the equations of general relativity such that no correction terms are necessary. Renormalization still occurs, where the cutoff for renormalization is the Planck unit (PSU) which is based on the fundamental constants of nature (within our universe at least).
Similarly, Huang [61] who presents a super-fluid model of the universe attempts to solve the fine-tuning problem by assuming a self-interacting complex scalar field that emerges with the big bang. The potential (defined as the Halpern-Huang potential) then grows from zero as the length scale expands (i.e. it should be asymptotically free) and the cosmological constant, in terms of a high-energy cut-off decreases with the expanding universe.
The nature of the fundamental constants and the large dimensionless numbers resulting from their relationships has been a long-standing puzzle (e.g. [62] - [69] ), and concepts such as a variable G [66] [67] [68] [70] and continuous matter creation have been introduced [66] . The relationship between the number of particles in the universe and Weyls ratio [62] [71] showed that the number of particles in the universe should be increasing proportionally to the square of the age of the universe and therefore matter must be continually created. Steady state cosmology, previously suggested by Hoyle [12] and Einstein [72] , offered such a concept, but with a constant G, as oppose to Dirac and his variable G. In previous work [73] this was resolved by suggesting that it is the mass-energy density that is changing and not G. In this paper we show that the mass-energy density decreases with the increasing size of the universe, so although the number of particles in the universe is increasing, with continuous matter creation the energy/information is conserved i.e. particles passing out of the observable universe are compensated by the creation of new particles where it is only through the creation of matter that an expanding universe can be consistent with the conservation of mass within the observable universe.
Figure 1. Schematic to illustrate the Planck Spherical Units (PSU) packed within a spherical volume.
The standard model of the universe (i.e. concordance ΛCDM) explains the accelerated expansion of the universe in terms of a negative pressure generated by the so-called dark energy. However, although in good agreement with CMB, large scale structure and SNeIa data, it is not yet able to explain the coincidence (fine-tuning) or the cosmological problem. As noted by Corda (2009) [74] extended theories of gravity (e.g. theories of gravity where the Lagrangian is modified by adding high-order terms in the curvature invariants or terms with scalar fields non-minimally coupled to geometry) generate inflationary frameworks which solve many of the problems, including the accelerated expansion. This is in agreement with the theory presented here where the acceleration of the universe can be explained in terms of a pressure gradient due to the information transfer potential at the horizon. The details of this are beyond the scope of this paper and will be addressed in a follow-up paper.
In summary we have shown how the generalized holographic model resolves the 122 orders of magnitude discrepancy between the vacuum energy density at the Planck scale and the vacuum energy density at the cosmological scale. Thus, not only resolving this long-standing problem in physics but also validating this geometrical approach. The details in terms of matter creation and the expansion rate are beyond the scope of this paper and will be addressed in a forth coming paper. The results presented here have profound implications for astrophysics, cosmogenesis, universal evolution and quantum cosmology giving incentive to further exploration and developments.
Acknowledgements
The authors would like to thank Dr. Elizabeth Rauscher, Dr. Michael Hyson, Professor Bernard Carr and Dr. Ines Urdaneta for their helpful notes and discussions, Marshall Lefferts and Andy Day for the use of their diagram (Figure 1) and as well the Royal Society for presenting the research in its preliminary stages at the 2015 satellite discussion meeting.
NOTES
*Presented at the Royal Society satellite discussion meeting―Particle, condensed matter and quantum physics: links via Maxwell’s equations, 18th-19th November 2015, Chicheley Hall, Buckinghamshire, UK.