Expanding or Static Universe: Emergence of a New Paradigm

The no-evolution, concordance expanding universe cosmology and no-evolution, static universe tired light model are compared against observational data on eight cosmology tests. The no-evolution tired light model is found to make a superior fit on all tests. Any attempts to introduce evolutionary corrections to improve the concordance cosmology fit on one test often worsen its fit on other tests. Light curve data of high redshift gamma ray bursts and quasars fail to support claims for cosmological time dilation due to expansion. Also, the SCP supernova light curve test results are considered to be flawed by selection effect biases. The big bang theory also has difficulty accounting for redshift quantization, for the multi-megaparsec periodicity seen in the distribution of galaxy superclusters, and for the discovery of galaxies at redshifts as high as z ~11.9. In overview, it is concluded that a static universe cosmology must be sought to explain the origin of the universe. One possible choice is a cosmology that predicts nonconservative tired-light redshifting in intergalactic space, the continuous creation of neutrons in space, the rate of matter creation scaling with both celestial body mass and temperature, galaxies growing progressively in size, and changing their morphology in the manner suggested by Jeans and Hubble.


Introduction
In the past, there have been two main interpretations of the cosmological red-shift phenomenon, the standard interpretation asserting that it is a recessional cosmological effect due to the expansion of space-time, and the competing view that the universe is not expanding and that the redshift is instead due to a "tired light" energy loss that photons undergo on their journey through space. These two models are most equitably compared to cosmological test data by refraining from introducing ad hoc evolutionary corrections. That is, the discrepancy of each model relative to the data trend becomes most apparent in the absence of evolutionary adjustments. This should not be taken to imply that galaxies do not evolve over time. Indeed, evolution is expected to occur in both the conventional big bang cosmology and to a lesser extent in the static universe, tired light cosmology since the latter does not have the same time limitations for the beginning of creation. Nevertheless, due to the uncertainty in knowing beforehand how much evolution actually would take place in each competing cosmology, it is best to entirely avoid making evolutionary assumptions and to choose the no-evolution model that makes the best fit to the data.
In the next section, the no-evolution, static universe tired light hypothesis and the no-evolution, expanding universe hypothesis are compared against data on several cosmology tests: the angular-size-redshift test, the galaxy number-count-magnitude test, the Tolman surface brightness test, the Hubble diagram test, and its variation the photon-flight-time-redshift test. As will become apparent, the use of multiple tests to compare competing cosmologies is more than just a review of past comparisons. It is an interactive approach to cosmology testing wherein a cosmological model's performance on any given cosmology test is made accountable to its performance on the other tests being examined. Thus, any assumptions introduced to the no-evolution prediction of a given cosmology with the intention of allowing it to better fit one set of test data must be applied as constraints to the interpretation of that cosmology on the other cosmology tests. It is found that in many cases, assumptions applied to make a cosmology fit better on one test, worsen its fit on another test. A final judgment as to the superiority of one cosmology over the other is made by considering the cosmology's performance on all tests, rather than on any one specific test in isolation from the others.
An overview of Section 2 shows that the no-evolution tired light model makes the best fit to the data on all tests without the introduction of ad hoc assumptions. The no-evolution expanding universe cosmology, on the other hand, is able to fit all the data only if numerous ad hoc assumptions are introduced specifying major evolution in galaxy cluster size, galaxy angular size, galaxy radio lobe size, galaxy luminosity, galaxy surface brightness, and galaxy number density. This however raises the question as to why cosmologists should stick with the expanding universe hypothesis with its need of ad hoc corrections chosen in such a way as to allow it to fit the various data trends, when the tired light model already fits the data reasonably well on all tests without making such adjustments. Thus, the tired light cosmology is found to be preferred because of its overall simplicity.
Finally, Section 3 examines the validity of past claims for the occurrence of time dilation in distant supernova since such claims play a critical role in the assumption of whether the universe is static or expanding. Also, in that section we examine the implications of the redshift quantization phenomenon in regard to the choice of the static universe or expanding universe alternatives. In addition, other difficulties for the big bang alternative are examined such as the multi megaparsec supercluster periodicity and the existence of galaxies at high redshift. Section 4 examines the conservative and nonconservative tired light model alternatives and whether it may be necessary to consider that our universe operates at a fundamental level as an open system allowing small departures from perfect energy conservation. Section 5 examines continuous matter creation cosmologies as a possible substitute for the failed big bang paradigm.

The Angular-Size-Redshift Test
The first cosmological test to be considered is the angular-size-redshift test. In this version of the test, distance is judged based on the angular size, θ, which is derived by observing the angular separations between bright galaxies in a cluster, as seen projected on the plane of the sky, and calculating the corrected harmonic mean of these separations. This θ value is then plotted against the cluster's redshift. One suitable θ-z data set is that published by Hickson and Adams [1] for a set of 94 galaxy clusters and which includes clusters at moderately high redshifts reaching up to z = 0.46. Figure 1 is adapted from the paper of LaViolette [2] who had plotted the linear Hubble relation, the no-evolution, static universe Figure 1. Harmonic mean angular separation for the brightest galaxies in a cluster plotted vs. redshift for 94 galaxy clusters. The no-evolution tired light model makes a far better fit to the data than the no-evolution 0 q = 0 Friedmann model assuming universal expansion. (After LaViolette [2] using data from Hickson and Adams [1]).
Euclidean space, predicts that cluster angular size θ should vary with distance r as: where k = d o H 0 /c and d o = 0.75 ± 0.15 Mpc is the intrinsic size determined for a typical cluster. This is plotted in Figure 1 as the downward sloping straight line. The tired light model specifies that a photon loses energy during its journey through Euclidean space as: ( where β = H 0 /c represents the rate of energy attenuation and r the distance traveled. This implies that photon wavelength λ should increase exponentially with distance as: where λ 0 is the wavelength of the photon at the time of emission. This in turn yields: Hence if no cluster evolution is assumed, the static universe, tired light relation predicts that cluster size should vary as: This appears in Figure 1 as the slightly curved solid line, diverging slightly upward from the linear Hubble relation. For small propagation distances, 1 r β , Equation (6) may be approximated by the linear relation, Equation (1). The no-evolution Friedmann model having a 0 q = 0 deceleration parameter and a Λ = 0 cosmological constant, which is here chosen as the expanding universe alternative, predicts that cluster size should vary as: ( ) ( ) proposed that 0 q = 0.09 for a Hubble constant of 65 km/s/Mpc, and in their mass density study Bahcall and Fan [4] have proposed 0 q = 0.1 ± 0.5. So, it is reasonable to choose the 0 q = 0 Friedmann model for comparison to the tired light model since the 0 q = 0.1 model prediction deviates only slightly.
Assumption-laden Friedmann models with larger 0 q values, such as the 0 q = 0.5 model, would plot substantially above the 0 q = 0 cosmology and would depart even further from the data trend. The more commonly cited ΛCDM cosmology with Ω M = 0.3 and Ω Λ = 0.7, which Goldhaber, et al. [5] had used in interpreting their supernova data (discussed in Section 3), when plotted on this test using H 0 = 55 km/s/Mpc, is virtually indistinguishable from the 0 q = 0 model over this redshift range. Hence there is no point to plot it. LaViolette [2] compares the fit of these three models, by assessing the variances between the θ data points and the prediction each model makes. He finds that variances for 1) the linear 1 z θ ∝ relation, 2) the tired light model, and 3) the expanding universe model compare respectively in the ratio 1:1.2:5.0. Repeating the calculation for the 31 most distant clusters (z > 0.1) gives relative variance ratios of 1:1.4:10. Thus the static, Euclidean tired light cosmology is seen to be significantly favored over the 0 q = 0 expanding universe model, requiring no need to introduce ad hoc assumptions about cluster size evolution. Lopez-Corredoira [6] has performed a study of the angular radii of galaxies over the redshift range z = 0.2 to 3.2 and also concludes that the angular sizes for galaxies conform more closely to a static, tired light cosmology with no need to assume size evolution. In his comparison of cosmologies, he includes the no-evolution tired light model and the no-evolution concordance cosmology (H 0 = 70 km/s/Mpc, Ω M = 0.3, Ω Λ = 0.7). The comparison is insensitive to the choice of H 0 . It was necessary for Lopez-Corredoira to use separate graphs for comparing each cosmology to his data set since his distance determinations are galaxy luminosity dependent, which in turn are sensitive to the particular cosmology being assumed.  Figure 2. Comparison of alternative cosmologies to data on the galaxy angular size-redshift test, (a) the no-evolution ΛCDM expanding universe cosmology and (b) the no-evolution tired light static universe cosmology.

P. A. LaViolette
obtained when comparing against galaxy angular size-redshift data: a) the no-evolution concordance cosmology and b) the no-evolution tired light cosmology. As is apparent, the tired light cosmology more closely follows the angular size-redshift data trend (dotted line). So, as in the previous angular-size redshift test, this test also demonstrates that the tired light model makes a superior fit, but using an entirely different angular statistic, namely galaxy angular size rather than the angular separation of galaxies in a cluster. If plotted in Figure  2(a), the 0 q = 0 Friedmann model would track close to the ΛCDM cosmology, but for z > 1.5 would begin to depart somewhat below its trend line.
It is noteworthy that the data trend throughout its redshift range is consistently offset below the tired light prediction by about the same amount over all redshifts, indicating galaxy diameters slightly larger than expected. Lopez Corredoira [6] attributes this to the possibility of dust extinction which affects the luminosity dependent method used in determining galaxy distances for the data set. Figure 3, adapted from Figure 7 of his paper, shows that this gap is closed by assuming a dust extinction of a v = 3.4 × 10 −4 Mpc −1 , which is equivalent to assuming an intergalactic dust density of ρ dust ~ 1.2 × 10 −33 g/cm 3 . It is proposed in Section 2.3 below that dust extinction also plays an important role in the Tolman test.
Yet another type of angular-size-redshift test utilizes, as the angular size, radio lobe separation indouble-lobed radio galaxies and quasars [7] [8]. Ubachukwu and Onuora [9] conducted one such test which compares quasar data extending up to z = 2.1 to various cosmological models; see Figure 4. They conclude that the static-universe, tired light model gives the best fit. The most favorable no-evolution Friedmann model which assumes a minimally curved space with 0 q = 0 predicts angular separations that are high by a factor of two at z = 2. The assumption laden ΛCDM cosmology has also been plotted for comparison and is seen to lie even further from the data than the Friedman model.  . Median values of the angular separation of double radio lobes in quasars plotted against quasar redshift 1 + z. The angular data has been corrected for projection effects arising from the host galaxy's inclination to the line of sight. The tired light model is seen to make a far superior fit to the data in comparison to the various expanding universe predictions (after Ubachukwu and Onuora [9]). For the Friedmann expanding universe model, or ΛCDM cosmology, to attain a data fit similar to the tired light model, strong evolutionary effects would need to be introduced. That is, it would be necessary to invoke the ad hoc assumption that galaxy radio lobes had larger angular separations in earlier epochs and have been gradually decreasing in size over time. However, not only does this further increase the complexity of the expanding universe cosmology vis-a-vis the tired light cosmology, it also requires that one accept that galaxy cluster size and galactic radio lobe separation, which normally would not be expected to be related, both change over time in just the right manner so as to allow the expanding universe model to make a good fit to the data! One might indeed be justified in asking the question about the big bang hypothesis that cosmologist and radio astronomer K. Kellerman [8] posed almost 50 years ago, namely, "Are we drawing too many epicycles?" The law of parsimony would instead point to the tired light model as the candidate model that is capable of explaining the greatest amount of data with the fewest assumptions.
Hoyle [10] has noted that Friedmann models having 0 q > 0 predict a minimum angular size at some finite z, with the expectation that observed angular size should increase with increasing z. For example, a 0 q = 0.5 Friedmann model predicts a minimum radio lobe separation of around 1 arc minute, but no such minimum is observed in the data; see Figure 4. Sandage [11] has noted that this minimum has been sought for in many observational tests but has not been found. He points out that this failure is considered by cosmologists to raise serious doubts as to whether the cosmological redshift is indeed due to a real expansion of space.
Many others have noted that the expanding universe hypothesis makes a poor fit against angular size redshift test data in both the radio, near infrared and visible parts of the spectrum. These include Kapahi [12], Andrews [13], Nabokov, et al. [14], and Lerner [15].

The Galaxy Number Count Magnitude Test
Another kind of cosmological test that has been used to check the predictions of cosmological models compares the differential galaxy number count, dN/dm (the number of galaxies per square degree falling in a given apparent magnitude interval dm), to the average magnitude of that interval, m. Figure 5 displays K-band data obtained up to the 25th magnitude that has been taken from Figure   1 of the paper by Totani, et al. [16]. This is compared to the no-evolution, tired light prediction (solid line), which is seen to make a relatively good fit to the data trend. The number counts for the tired light prediction were reduced in accordance with the selection effect correction which Totani et al. give in Figure 6 of their paper. The dot-dashed line branching above the tired light prediction represents the uncorrected tired light prediction. 1 The expanding universe alternative predicts an additional dimming of galaxy apparent magnitude since it incorporates an additional factor of (1 + z) due to Figure 5. Differential galaxy number counts plotted against uncorrected galaxy K magnitude. Superimposed for comparison are the no-evolution, tired light model corrected for data selection effects (solid line), the tired light model with no correction (dot-dash line), the no-evolution, Ω (0.2, 0.8) accelerating universe model with selection effect correction (dotted line), and the no-evolution, 0 q = 0.5 expanding universe model with selection effect correction (dashed line). 1 The dN/dm vs. m dependence for a static, non-evolving Euclidean universe with no redshift dependent attenuation is represented by the sloping straight line in Figure 5. In such a universe, the integral galaxy number count N would increase with the cube of distance, r as N ∝ r 3 . Galaxy brightness would decrease according to the inverse square of distance or expressed in terms of apparent magnitude m, it would vary as: m ∝ 5log(r/10), or, r ∝ 10 0.2m . Combining these relations gives: N ∝ 10 0.6m , or similarly for differential counts: dN/dm ∝ 10 0.6m . the relativistic time dilation effect (the Hubble "number effect"). The 0 q = 0.5 no evolution, expanding universe prediction, which is plotted as the dashed line, falls substantially below the dN/dm data trend on this test. The accelerating expanding universe cosmology modeled with Ω M = 0.2 and Ω Λ = 0.8, plotted as the dotted line, fairs better, but is also assumption laden. This cosmology plot and the Friedmann cosmology plot are both taken from Figure 9 of Totani's paper and include corrections for data selection effects. The 0 q = 0 Friedmann model, not shown, would plot somewhat below the accelerating universe prediction.
While the accelerating universe model fits the number count data better than the assumption that space was more densely populated with galaxies in earlier epochs and that galaxy number density has been rapidly decreasing over time.
For the 0 q = 0.5 cosmology prediction, galaxies would have had to be 10 times more abundant at m K = 22 (z ≈ 2.3, t ≈10 billion years lookback time) as they are at present. But then this raises the question whether it is justified to assume that the spatial population density of galaxies has been varying to such a great extent and in just the right manner so as to allow the expanding universe model to make a good fit to the data, given that the tired light cosmology already makes a reasonably good fit. In addition, LaViolette [2] has tested the tired light model on the radio galaxy differential number count test and finds that it makes a better fit than the expanding universe prediction. 2

The Tolman Surface Brightness Test
The Tolman surface brightness test, devised in 1930 by Richard Tolman [18], provides another way of distinguishing the predictions of alternative cosmologies. This test uses galaxy surface brightness, S, as a distance indicator for comparison to galaxy redshift. The expanding universe model predicts that surface brightness should vary as the inverse fourth power of redshift, factor of (1 + z) being due to relativistic time dilation, one factor being due to 2 Edward Wright [17] has contested LaViolette's conclusion that the no-evolution, tired-light cosmology makes a good fit on the radio galaxy differential number count test. His objection was that the tired-light cosmology does not fit the small dip in the data trend evident in number counts of the brightest sources, a data trend anomaly that comprises one hundredth of one percent of the total number of sources in that study. Kellermann and others, however, have questioned whether this minor number-count deficiency of bright sources is even real, suggesting that it may be an artifact of the poor sampling statistics in that part of the data set which samples nearby galaxies. By most standards, a model that fits 99.99 percent of a data trend would be considered to be a desirable choice, especially when compared with the non-evolving 0 q = 0 Friedmann cosmology which fits only a small fraction of the data points (<10 −4 ). In 1989 LaViolette wrote to him pointing out that his refutation of the tired-light model is supported by very few data points. He unexpectedly responded that "a valid theory must fit all the data, not just 99.99 percent of it". To the contrary, given that number count data is inherently stochastic, it is meaningless to distinguish 99.99 percent from 100 percent. the cosmological (Doppler) redshift effect, and two factors being due to relativistic geometrical aberration [19]. The tired light model, on the other hand, predicts an inverse surface brightness-redshift relation of ( ) just one factor of (1 + z), the decline in surface brightness being due exclusively to the photon energy loss that produces its nonDoppler redshift. Consequently, the tired light model predicts that galaxy surface brightnesses should appear far brighter than the expanding universe alternative.
In the present discussion, we consider a version of this test which Lubin and Sandage [20]  Lubin and Sandage had considerable flexibility in adjusting the expanding universe model to fit their data. On the one hand, they had a range of expanding universe cosmologies available to them, differing by the value of the deceleration parameter (e.g., 0 q = 0, 0.5, or 1). On the other hand, for a given cosmology, they had a wide range of galaxy luminosity evolution models to choose from to close the gap between theory and observation. These assumed that galaxies were brighter in primordial times to varying extents, ranging from 0.6 up to 1.7 magnitudes brighter in the R-band at a redshift of z = 0.9 [23]. From this wide range, they chose the 0 q = 0.5 cosmology as their best alternative and noted that closing its prediction gap required an assumed luminosity correction of ΔM(R) = 0.99 mag for the R-band data for the z = 0.92 cluster, and ΔM(I) = 0.39 mag and 0.44 mag for the I-band data for clusters at redshifts z = 0.76 and 0.9. But given the flexibility at their disposal in choosing a model to fit their data, is it not inappropriate for them to claim that they were checking the viability of the expanding universe cosmology relative to the tired light cosmology?
If their intention had been to make an equitable comparison, a better approach would have been to plot both cosmologies with no evolutionary corrections and compare each on multiple cosmology tests, a method advocated by LaViolette [2] and utilized as well in the present study. If they had, they would have seen that without evolution their favored 0 q = 0.5 expanding universe cosmology not only makes a poor fit to their data, but also makes a very poor fit on both the angular size redshift tests and galaxy number count test. If one refrains from making ad hoc evolutionary corrections and compares the alternative cosmologies to their R-band surface brightness data, it is found that the no-evolution, tired light cosmology in fact lies closer to the data trend than either of the two no-evolution, expanding universe cosmologies, the tired light model being ≈0.4 magnitudes brighter than the data trend at z = 0.92, while the 0 q = 0.5 and 0 q = 0 cosmologies are 1.2 magnitudes and 1 magnitude dimmer respectively. Here the tired light model predicts surface brightnesses that are slightly higher than the data trend, while the expanding universe cosmology predicts surface brightnesses that lie substantially below the data trend. When the no-evolution cosmologies are compared to their I-band surface brightnesses data, the tired light prediction in this case lies further from the data trend, deviating by 0.89 magnitudes brighter at z = 0.90, as compared with 0.45 and 0.65 magnitudes dimmer respectively for the 0 q = 0.5 and 0 q = 0 expanding universe cosmologies. This comparison changes considerably if corrections are introduced for light extinction due to the presence of galactic and intergalactic dust, something that Lubin and Sandage did not do. Both Aguirre [24] [25] and Goobar et al. [26] argue that light extinction by intergalactic dust may be as high as 0. According to Model B of the paper by Goobar et al. (Figure 9 and Figure 10 in their paper) galaxies at a redshift of z = 0.92 would be dimmed by ≈0.33 magnitudes in the R band and by ≈0.3 magnitudes in the I band. Also, Rowan-Robinson [27] has proposed extinctions originating internal to the galaxy of 0.33 magnitudes for host galaxies in the redshift range z = 0.15 to 0.8. To correct the data for dimming due to both galactic and intergalactic dust extinction, it is here suggested that the R-band surface brightnesses which Lubin and Sandage report for galaxies in the z = 0.92 cluster be increased by 0.4 magnitudes and the I-band surface brightnesses they found for galaxies in the z = 0.76 and z = 0.9 clusters be increased by 0.36 magnitudes, the dust extinction correction in the I-band being assumed to be slightly smaller.
With these extinction corrections, the data trends in the R-band data set would move upward by 0.4 mag, allowing the no-evolution tired light prediction to make a good fit, while increasing the discrepancy of their 0 q = 0.5 and 0 q = 0 expanding universe predictions to 1.6 mag and 1.4 mag respectively. Even against their I-band data, which is about 0.4 to 0.5 magnitudes dimmer than their R-band data, the tired light cosmology with dust extinction makes a better fit than either expanding universe prediction. The tired-light model now would be 0.5 mag brighter than the data set, whereas the 0 q = 0.5 and 0 q = 0 cosmologies would now lie 0.8 mag and 1.0 mag dimmer than the data trend. To be fair, the tired light cosmology should be compared to the 0 q = 0 expanding universe cosmology since the 0 q = 0.5 cosmology requires the introduction of unsupported assumptions about the existence of hidden mass. So, even considering the I band cluster data of Lubin and Sandage, when dust extinction is assumed, the tired-light model makes a far closer fit.
It does not seem unreasonable to introduce the assumption of dust extinction since Lubin and Sandage used comparatively dim clusters in their study. For example, compared to the 102 clusters tabulated in the Hubble diagram study of Kristian, Sandage and Westphal [28] or the 119 clusters tabulated in the paper by Postman and Lauer [29], the three clusters that Lubin and Sandage use lie near the faint limit of the luminosity range in these other studies. In particular, the data points for the first ranked galaxies from each of the three high-z clusters they analyzed range from half a magnitude to a full magnitude dimmer than the data trend for the first-ranked galaxies in the data set of the Hubble diagram of Kristian-Sandage-Westphal (KSW) plotted in Figure 6 (Section 2.4).
To illustrate this the magnitudes of the first ranked galaxies in the three clusters studied by Lubin and Sandage are converted in Table 1 so that they may be properly compared with the KSW data. The magnitudes listed in columns (4) and (6) of Table 1 have been taken from Tables 5-7 of Lubin and Sandage [22] using their values for Petrosian galaxy radii of η = 1.7, except for galaxy No. 9 (z = 0.76) whose magnitude was available only for a Petrosian radius of η = 1.5.
To convert the I-band magnitudes for the first two galaxies listed in the table to R-band magnitudes a Keck R-I color index correction is applied (col. 5). Also, a dust extinction correction similar to that applied to the Tolman test data is added (col. 7) to brighten the magnitudes of all galaxies. The magnitudes are then brightened by an additional -0.25 magnitudes (col. 8) since the photometry for the Kristian-Sandage-Westphal data are based on the Johnson-R system whereas that for the Lubin and Sandage data is based on the Cape-Cousins system, the latter magnitudes being fainter by a zero-point offset of 0.25 magnitudes; see footnote on p. 1072 in Lubin and Sandage [22]. When the corrected magnitudes (col. 9) are compared to the first ranked galaxies plotted in the KSW study, two are seen to be so dim as to lie to the far right of the plot in Figure 6, the third at z = 0.9 lying entirely off the right side of the graph.
In 2002, in correspondence with A. Sandage, the Author pointed out that the clusters they used in their study were relatively faint [30]. To this, Sandage replied [31]: "... as you have noticed, the first ranked galaxy in each of the three clusters that we (Lubin/Sandage) have studied are fainter than the mean of the distribution of either the total sample of Postman and Lauer or of K/S/W. However, each of these brightest galaxies in our (Lubin/S) three clusters is still within the confines of the distribution of absolute magnitudes in either of these two lists (P/L and K/S/W), although, it is true, they are near the limit on the faint side. We have assumed that our three clusters are fainter than average in their brightest member, but not outside the known distribution for 'local' clusters." So, in view of this admission, and the above magnitude comparison, the decision to brighten the Lubin-Sandage data set by the inclusion of a dust extinction correction appears justified and allows the tired-light model to make a superior fit to the Tolman test data. International Journal of Astronomy and Astrophysics Table 1. Magnitude corrections applied to first ranked galaxies in the Lubin/Sandage data. (1) ( In paper number IV of their Tolman test paper set, Lubin and Sandage [23] argue that standard luminosity evolution models require that galaxies should have been brighter in primordial times and that such a luminosity evolution assumption would move the tired light model prediction away from the data trend, instead of closer, to the model's detriment. However, their desire to apply the same luminosity evolution assumptions to the tired light model is poorly founded. For, static-universe tired light cosmologies require nonstandard models of stellar evolution. Indeed, when one adopts the tired light model as being the correct alternative, galaxy distances and look-back times increase in comparison with distances predicted by the expanding universe cosmology, which in turn affects the rate of galaxy evolution. Also as described below, static universe cosmologies lead in the direction of requiring a mechanism of continuous matter creation to explain the origin of matter, something that would drastically alter any assumptions about primordial galaxy evolution.
Lerner [32], Crawford [33], and López-Corredoira [34] have all been critical of the way Lubin and Sandage performed their study, and of their claim that their test data refutes the static universe, tired light model. Lerner [32] has performed a repeat Tolman test whose data set includes galaxies with redshifts of up to z = 6 and concludes that the data is clearly compatible with the static universe hypothesis and clearly incompatible with the expanding hypothesis, even when reasonable brightness evolution is included. Also, Lerner, et al. [35] has performed a repeat of the Tolman test earlier conducted by Pahre, et al. [36] using UV surface brightnesses of galaxies having redshifts as high as z ~ 5. After correcting the errors made by its authors in their cosmology comparison, he demonstrated that the data in fact make a good fit to the static universe Euclidean tired light model and are incompatible with the concordance expanding universe cosmology. Andrews [37] has also compared the expanding universe model and the static universe, tired light model on a Tolman test and has concluded that the data conclusively favor a static universe over an expanding universe. Both of Lerner's test findings and those of Andrews are contrary to the conclusions of Lubin and Sandage.

The Hubble Diagram Test
The Hubble diagram test uses galaxy apparent magnitude, m, as a distance indicator for comparison to galaxy redshift. The m-z curve for the no-evolution tired International Journal of Astronomy and Astrophysics light prediction is given by the following equation: where C = 19.8 [38]. The first term, ln(1 + z), results from the nonlinearity of the tired light redshift distance relation and is derived by substituting the tired light relation r = ln(1 + z)/β into the magnitude relation m = 5logr + constant, with the 1/β term being absorbed into the constant. The second term is the "energy effect", where galaxy dimming is due to the spontaneous diminution of photon energy with travel distance.
In the expanding universe cosmology, galaxy dimming is due both to the cosmological (Doppler) redshift effect and to relativistic time dilation, the latter effect causing the emitted stream of photons to spread out in time. It is also affected by the distances modeled for the galaxies, which differ from those in the tired light cosmology. The m-z curves that plot the standard no-evolution Friedmann model predictions are based on the equations of Mattig given as: where C = 19.8 [39]. At a given redshift, these expanding universe cosmologies predict magnitudes fainter than the tired light model. Figure 6 plots the R-band magnitudes against redshift for the brightest galaxies in each of 103 clusters taken from Figure 4 of the 1978 paper of Kristian, Figure 6. The Hubble diagram charts the R-band magnitude-redshift coordinates for the brightest galaxies in each of 103 clusters; data for z < 0.75 is taken from Kristian, Sandage and Westphal (1978). The data is compared with the no-evolution tired light cosmology (solid line), the 0 q = 0.5, no-evolution expanding universe cosmology (short dashes), the 0 q = 0, no-evolution expanding universe cosmology (long dashes), and the ΛCDM cosmology (dotted line). Sandage and Westphal (KSW) [28]. Although the data is comparatively old and does not rise over z = 0.75, it nevertheless is useful in this multi-test study as a means of constraining the predictions of competing cosmologies with respect to other tests. The data set is here compared to the no-evolution, tired light cosmology (solid line) the 0 q = 0 no-evolution Friedmann prediction (long dashed lines), the 0 q = 0.5 no-evolution Friedmann model (short dashed lines), and the ΛCDM cosmology (dotted line). The latter two cosmologies make less desirable comparisons since they have unsupported assumptions about the existence of hidden mass, dark matter, or dark energy. So again, the tired-light model makes the better fit. Of the three expanding universe predictions, the 0 q = 0.5 Friedman model with the unsupported assumption of hidden mass comes closest to the data trend. However, choice of this cosmology over the 0 q = 0 cosmology worsens the fit of the expanding universe prediction on the angular size redshift tests, moving it further from the data trend on those tests. This shows the advantage of using multiple cosmology tests when one wishes to test competing cosmologies in a fair and consistent manner.

The Photon Flight Time Redshift Test: A Version of the Hubble Diagram
To make a more definitive judgment between expanding and static universe cosmologies using the Hubble diagram test alone, higher redshift data is needed.
To this end, Marosi [40] has examined magnitude-redshift data for 280 supernovae and gamma ray bursts (GRBs) extending from z = 0.01 to z = 8.1 and finds that the tired light model makes a very good fit to the data over the entire data span. To better compare the competing cosmologies, rather than plotting redshift z versus apparent magnitude, m, he plots redshift versus photon flight time, t s . When this is done, the difference between the expanding and static universe alternatives becomes more apparent. The t s -z diagram shown in Figure 7 line, which is virtually indistinguishable from the t s -z supernova/GRB data trend. The no-evolution 0 q = 0 Friedmann cosmology, the lower straight line, has been added to Marosi's plots for comparison. While this Friedmann model is currently not as popular as the ΛCDM cosmology, it is preferred in the present multi-test comparison since it makes no ad hoc assumptions about the presence of dark matter and dark energy.
The ΛCDM and Friedmann cosmologies both assume that the universe expands linearly according to Marosi found that when H 0 = 2.024 × 10 −18 /s, this static universe cosmology yields a best fit to the data trend with a very high confidence level. Rewriting (11) as: and substituting t s = r/c in the left term of Equation (12) yields essentially the r-z relation of Equation (5). As Marosi notes, the tired light prediction essentially closely overlays the supernovae z-t s data trend out to z = 8.1, while the ΛCDM expanding universe cosmology with H 0 = 72.6 departs significantly from the data trend for redshifts > 3. At z = 8.1, this ΛCDM cosmology predicts a value of z ~ 25, over 3 times greater than the data trend! Compared to the ΛCDM cosmology, 0 q = 0 Friedmann cosmology avoids making ad hoc assumptions about dark matter and dark energy. But it makes a poor fit for z > 2, consistently predicting higher t s values than the data trend.
Marosi [41] has also performed a photon-flight-time vs. redshift test for 84 gamma ray bursts ranging from z = 0.033 to z = 8.1. The t s -z diagram shown in   [44]. It also falls substantially below the value H 0 = 69.8 ± 0.8 km/s/Mpc obtained by Freedman, et al. [45] by calibrating TRGB in galaxies at distances of 7 to 20 Mpc using eclipsing binary stars in the Large Magellenic Cloud and tying them to Type 1a supernovae. It also lies far below the ΛCDM cosmology value of H 0 = 74.0 ± 1.4 km/s/Mpc which Riess, et al. [46] have derived based on analysis of Cepheids in the Large Magellenic Cloud.
The discrepancy between these various H 0 determinations is found to be up to ten-fold greater than the accuracy of estimating the individual H 0 values. This conflict among determinations, termed the Hubble tension, has been problematic for the concordance cosmology. Some have suggested that the introduction of new physics may be needed to resolve it. Others such as Lombriser [47] and Ding, et al. [48] suggest that use of local calibrators may bias H 0 toward higher values since our Galaxy resides in an under dense region of the cosmos.
Sandage, et al. [42] attribute these high H 0 values to a faulty choice of LMC P-L relations resulting in moduli that predict distances too short compared to their own. However, considering that the cosmology tests considered above collectively rule out the expansion hypothesis as the best choice cosmology, it seems more prudent to choose the H 0 = 64.0 ± 1.6 value of Tammann and Reindl which approximates the 65.3 ± 2.9 km/s/Mpc value that emerged from Marosi's tired light fit to supernovae and gamma ray burst data.

No Evidence for Time Dilation
The expanding universe cosmology hypothesizes that high redshift galaxies are receding from us at close to the speed of light and that, due to the relativistic time dilation effect, clocks in those galaxies should be ticking slower or alternatively that supernova explosions should be taking longer to occur. One indication that we live in a cosmologically stationary, non-time-dilated universe comes from studies of gamma ray bursts. These are believed likely to be produced by supernova explosions [49]. So, if the light curves of distant supernova were in fact being time dilated, one would expect to see a similar effect in the duration of gamma ray bursts, the more distant, more highly redshifted gamma ray bursts being expected to last longer on the average. But such is not seen to be the case. It has now been established that X-ray bursts are essentially the same phenomenon as gamma ray bursts, except that they originate from much greater distances, the gamma rays in the original burst being cosmologically redshifted down to the X-ray energy band. However, the X-ray bursts are found to last about as long as gamma ray bursts.
Another study examined the durations of 195 Swift detected gamma ray bursts ranging out to a redshift of 8.1 yet found no evidence of time-dilation broadening in the light curves [50]. The data from this study, which is presented here in Figure 9, shows no evidence of any correlation between redshift and duration. Crawford [51] has also examined gamma ray burst data and finds no evidence for any time dilation of their light curves. Quasar light curves also show no evidence of time-dilation broadening. One such study performed a Fourier analysis of the light curves of 800 quasars which were monitored on time scales from 50 days to 28 years to see if more distant quasars exhibited longer duration outbursts [52]. Low redshift quasars (z < 1) were compared to high redshift quasars (z > 1), but no evidence for time dilation was seen. The results of these various studies not only support the static universe, tired light cosmology, but also call into question the validity of the time dilation conclusions of the Supernova Cosmology Project (SCP).
Let us take a moment to analyze that study. The SCP study analyzed the durations of 60 supernovae out to a redshift of 0.83 and the authors claimed that their results indicated that supernova explosions in distant galaxies are time-dilated in accordance with the predictions of the expanding universe cosmology. They claimed their data was best fit by a ΛCDM cosmology having cosmological parameters Ω M = 0.28 and Ω Λ = 0.72, and that their analysis refutes the static universe tired light model, which predicts that supernova duration should be independent of redshift. But the dependence of light-curve width on supernova redshift which their data showed was likely an artifact of data selection bias. In particular, this study is handicapped by the Malmquist bias, the tendency to preferentially detect intrinsically bright objects. As Phillips [53] has pointed out, searches for distant type Ia supernovae (z > 0.2) will clearly favor the discovery of superluminous events and this could introduce a significant Malmquist bias into the survey. At higher redshifts our telescopes necessarily sample a much greater volume of the universe and hence a much greater number of supernova-producing galaxies. In the redshift range 0.4 to 0.5, one would be surveying a volume of space that was 34 times larger than that of the local neighborhood which spans the redshift range 0 to 0.1. Going out to a redshift range of 0.7 to 0.8, this observational volume increases to 66 times greater than the local volume. Hence because the high redshift domain subtends a very large volume of space, compared with the local environment, there is a much greater probability of observing extremely luminous high-energy supernovae, which normally occur very rarely. Since the light curve of such supernovae persist much longer than those of less luminous supernovae [54], high-z supernova searches will be skewed to discovering high-luminosity, long-duration supernovae, giving exactly the result that they found even with time dilation left out of the picture.
Moreover, shorter duration supernovae, being less luminous at maximum light, would not be as easily seen, especially at high redshifts where both the greater distances and greater dimming due to photon energy loss conspire to create conditions adverse to their detection. Hence there would be a tendency for supernovae with lower light curve width values to pass undetected resulting in a flux limited selection. Such flux limitation could explain why the SCP data set contains progressively fewer supernovae at progressively higher redshifts where instead progressively more supernovae should be observed due to the progressive increase in the volume of surveyable space. For example, the SCP data set contains 24 supernovae in the redshift range of 0.3 to 0.5; 13 supernovae in the redshift range of 0.5 to 0.66 (defining a volume of space 30% larger); and only 3 supernovae in the redshift range of 0.66 to 0.85 (a 210% larger space volume). It is not a question of whether this data set might be flux limited. It is clear that it is flux limited since there is no other explanation that could account for this kind of number drop off. Standard theories of galaxy evolution cannot explain this drop off since such theories predict that high-z galaxies were bluer and had greater star formation rates with more frequent supernovae.
The occurrence in their data of a supernova at z = 0.46 with a width factor of 2.26 provides evidence that some supernovae can have very long light curve durations, more than three times greater than what the time dilation assumption would predict. The width of this supernova even surpasses the widths of two supernovae that have almost twice the redshift. So, this extremely long duration event may simply be a supernova that is at the upper end of the spectrum in terms of luminosity and duration. Moreover, a decade after this SCP study was published, Quimby, et al. [55] reported discovering supernovae that are an order of magnitude brighter than type I supernovae and that take much longer to fade away. If we were to wait long enough, such long duration supernovae would be seen also in the local low-z neighborhood.
Although there is a tendency for brighter supernovae to last longer and to decline slower, supernovae having the same peak absolute magnitude can nevertheless have widely varying light curve decline rates, as represented by the parameter Δm 15 (B), the amount that the supernova's B-band apparent magnitude decreases from supernova maximum by the 15th day. In Figure 7 of his paper, Rowan-Robinson [27] has plotted Δm 15 (B) versus peak absolute magnitude for local supernovae that were first observed before their maximum and finds a large amount of data point scatter. For a given absolute magnitude, Δm 15 (B) duration can vary by up to ±30%. If we were to study a sample of supernovae collected over a much longer time period, say over two centuries, we would most likely find an even greater amount of data scatter perhaps as large as ±50% since the more rarely occurring, brighter and longer lasting supernovae, would be more likely to be seen. So, to avoid a selection effect bias, a larger data scatter value of about ±50% would be more appropriate when interpreting data observed at high redshifts. Given this amount of data scatter and a flux limitation inherent in observing at high redshifts, one is left to conclude that the high-z data published by Goldhaber et al. were significantly biased toward the high width factor end of the data scatter range and could have yielded width factors as much as 50% too high compared to the norm even with no time dilation effect.
A related problem with the SCP study is that the local neighborhood is under sampled. To get a fair sampling of supernova light curve widths in the local environment that would compare in a fair manner with what is seen at high redshifts, we would need to observe for a total of 260 years, or 65 times longer than the four-year period over which Goldhaber et al. collected their data. If the SCP group were to extend their study of the local environment for a few more centu-ries, quite likely they would discover supernovae that had durations just as long as the one's they observed at high redshift. Adding in these potentially observable long duration supernovae would shift the light curve width data upward toward higher width values.
In summary, due to the comparatively short time span for their search for local supernovae, the Supernova Cosmology Project's data set is biased toward charting lower width values at low redshifts. At the same time, due to flux limited sample selection effects their data set is biased toward charting higher width values at high redshifts. Together, these two effects conspire to produce the observed upward sloping trend line. With proper sampling, it is expected that such data would yield a flat trend line with no evidence for time dilation.
In Section 4 of their paper, Goldhaber et al. briefly acknowledge that their supernova data might suffer from such a selection effect bias. David Crawford [56] has also noted that the SCP supernova data set may be biased by selection effects that could affect the outcome of the study's time dilation conclusions. John Masreliez [57], another critic of the study's conclusions, also makes a convincing case that the SCP supernova sample is flux limited and that selection effects could entirely account for the 1 + z rise in supernova light curve width factor. He also notes that the positive slope in the redshift-width relation arising from selection effects should have added on to the slope predicted by the 1 + z time dilation effect, if in fact it were present, to produce a slope much steeper than 1. Since such a steep upward slope is not seen, we are left to suppose that there is no 1 + z correlation and that the slope is due to other factors fortuitously mimicking the slope predicted by the expanding universe hypothesis.
As stated earlier, a particular cosmology cannot be proven or disproved on the basis of a single cosmology test. We must take a more holistic approach and view a cosmology's performance on several tests simultaneously. On the one hand, we have the supernova-light-curve-width test of Goldhaber et al. whose results are questionable due to selection effect biases and which favors a specific accelerating universe ΛCDM cosmology that does not perform well on other cosmology tests. On the other hand, we have the angular-size-redshift test, Tolman test, number count magnitude test, Hubble diagram test, and the photon-flight-time-redshift test, all of which favor the tired light cosmology which consistently makes a better fit to the data. To be considered remotely plausible the expanding universe hypothesis must be modified to include specific assumptions regarding the evolution of galaxy cluster size, galaxy radio lobe size, galaxy luminosity, and galaxy number density, etc. But the required assumptions are numerous, and some even produce opposing results, worsening the fit of the expansion cosmology on certain tests. The tired light model, on the other hand, being free from the need for ad hoc evolutionary corrections, is preferred on the basis of its simplicity.

The Redshift Quantization Effect
Tifft [ Cocke [65] find that cosmological redshifts are "quantized," in one sixth submultiples of cΔz = 72.45 km/s, i.e., 12 km/s, the 24 km/s and 36 km/s harmonics being most prevalent. After further study, Tifft [66] concluded that these other redshift quantizations were higher multiples of either 8.05 km/s or 2.68 km/s which he then regarded as the most basic quantizations (1/9 and 1/27 of the originally discovered unit). These findings have more recently been confirmed by Guthrie and Napier [67] and Napier and Guthrie [68] who find 37.5 and 71.5 km/s periodicities in the redshifts of the local supercluster. The existence of the 72 km/s periodicity is now well established with a probability of only 10 −6 that it is due to chance.
Cocke and Tifft [64] suggest that the redshifts may be due to cosmological expansion and that the observed quantization indicates that the expansion of the universe is quantized, or that the universe is stationary and that the photon emission properties of atoms are quasi-stationary, for example, with the Rydberg constant monotonically changing its value over time in discrete steps. Here they may be alluding to the variable mass theory of Narlikar [69] which suggests that matter is being continuously created in the universe and that the inertial mass of matter gradually increases from the time of its creation, thereby creating a cosmological redshift-distance effect.
LaViolette [2] has proposed an alternative interpretation of this quantization effect suggesting that the observed incremental change of redshift represents discrete steps in the decay of photon energy as photons journey through space. Thus, rather than losing energy continuously, as Equation (1) describes, photon quanta would change their energy (and wavelength) in incremental fashion. Taking Δr as the distance over which an average photon travels before undergoing a redshift transition of amount Δz, then over n increments it would accumulate a redshift of nΔz. Thus, given that Δr = cΔz/H 0 , this implies Δr = 0.188 Mpc, given that cΔz ~ 12 km/s and H 0 = 64 km/s/Mpc, if we adopt the value of Tammann & Reindl [43]. So, a photon would travel a distance of about 610,000 light years before undergoing an incremental decline in energy and corresponding increase in redshift. Over extended distances tired light energy loss could be expressed as: which is an updated version of Equation (4) above. On the other hand, if the expanding universe hypothesis were adhered to, it would be necessary to assume that space-time globally expands in quantized fashion, which appears to border on the incredulous. Another interesting quantization result comes from the observations of Arp [70] which show that galaxy redshifts relative to the main galaxies in the Local Group, M31 and M81, and in the Sculptor Group are quantized in steps of approximately 72.4 km/s, matching the quantization interval that Tifft and Cocke find for more distant galaxies. Arp reports that since the redshifts are known with a precision of about ±8 km/s, and for seven of these galaxies even more precisely, about ±4 km/s, one is led to conclude that the galaxies in these groups are unusually "quiet" (relatively motionless). Observing that galaxy redshifts in the M31 and M81 groups are distributed around the 72.4 km/s periodicity with a standard deviation of 17 km/s, Arp [71] has quite reasonably interpreted this as evidence that the peculiar motions of galaxies in these clusters can have velocities no larger than this. This presents a strong argument that the differential redshifts of galaxies in these local groups are not due to relative motion of the galaxies, but to some intrinsic photon energy change phenomenon.

Multi-Megaparsec Structures
Another problem with the expanding universe hypothesis is its inability to adequately account for the regular spacing of galaxy superclusters [72]. That is, as we look further and further out into space in the direction of the north and south galactic poles, the number of galaxies per unit volume is found to alternately increase and decrease in cyclic fashion. Galaxies group into wall-like structures that are seen to be spaced from one another by about 180 Mpc along our line of site, their wave-like pattern being seen to stretch out 1500 Mpc (5 billion l.y.) in either direction. This poses a problem for the expanding universe theory which predicts a space-time dimension doubling during that period. If the universe was expanding, and a supercluster wave pattern did for some reason emerge, its wavelength would be expected to vary with time, being twice as long now as it was 5 billion years ago. But this is not the case. The wavelength stays constant with look-back time. This is more logically explained if the universe is static and Euclidean and that galaxies in some manner became preferentially created in certain regions of space so as to form a wave pattern of cosmic proportions. Moreover, assuming that galaxy clusters had been initially uniformly distributed in space and typically had gravitationally induced peculiar velocities of 1000 km/s, the time taken to traverse 90 Mpc to form just one of these supercluster aggregations calculates to be 100 billion years, which far exceeds the age of a big bang universe.

The Age of the Universe
Another difficulty with the concordance expanding universe theory is that it predicts an age for the universe that is too short in comparison to the ages found for the highest-z galaxies. For example, galaxies have recently been discovered having redshifts as high as 11.1 [73] and 11.9 [74]. According to the ΛCDM H 0 = 70 km/s/Mpc concordance cosmology, a galaxy at z = 11.9 would be seen in existence 300 million years after the time of the big bang. This would place it well before the reionization epoch which is theorized to have occurred around z = 7.7 [44]. This raises the question of how this galaxy would have had time to develop given that recent models of galaxy formation indicate that it should take a galaxy at least 750 million years to form. How does cosmology explain that stars would have started to form and develop into a galaxy prior to the beginning of the big bang! For other problems with the big bang/expanding universe hypothesis, see International Journal of Astronomy and Astrophysics López-Corredoira [34].

Conservative and Nonconservative Tired Light Models
It is useful to review here some history of the tired light model. A number of cosmologists have proposed energy conserving tired light mechanisms in which the energy lost from the original photon remains in the universe as low-grade heat. For example, seven months after Hubble published his redshift-distance relation, Zwicky [75], publishing in the same journal, proposed an energy-conserving tired light mechanism as a nonvelocity explanation of Hubble's findings. His theory assumed that photons have a nonzero rest mass and lose energy as a result of a gravitational drag resulting from their interaction with ambient matter. Alternatively, Pecker and Vigier [76] have suggested that cosmological photons have a nonzero rest mass and lose energy through their interaction with a bath of φ-particles that have masses much smaller than that of an electron. Also, Marmet [77] [78] [79] has proposed a conservative energy loss mechanism involving photon scattering from intergalactic hydrogen nuclei that has some basis in laboratory experiments. His experiments demonstrated that photon interactions with intergalactic gas should produce no angular deflection and hence no image blurring. Also, Zheng [80] has proposed that "soft photon" scattering from intergalactic electrons can cause redshifting. However, the scattering theories of both Marmet and Zheng have difficulty explaining why the cosmological redshifts are the same at radio frequencies since an intergalactic medium of hydrogen gas or electron ions would not scatter radio waves in the same manner as optical wavelength photons.
Image blurring and spectral flatness criticisms, however, do not apply to most non-conservative tired light mechanisms. James Clerk Maxwell may have been one of the first to propose nonconservative photon energy loss. His original electromagnetic wave equation contained the energy damping term σ 0 μ 0 (∂φ/∂t), where σ 0 and μ 0 represented the electrical conductivity and magnetic permeability of background space [81] [82]. Nernst [83] [84] put forth a nonconservative tired light idea in which he proposed that Olber's paradox might be resolved if photons were assumed to undergo nonconservative energy damping during their journey through intergalactic space [85]. As in Maxwell's damped EM wave, in Nernst's version the lost energy was proposed to physically disappear from the universe. Vigier [86] has proposed a non-energy-conserving tired light model in which photons lose energy through energy dissipating interactions with stochastic vacuum fluctuations.
LaViolette [2] [87] has also proposed a non-conservative tired light effect. But the version he proposed was not devised specifically to explain redshift-distance observations, as was the case for the other tired light theories described above.
Rather, his tired-light relation emerged as a prediction of the subquantum kinetics (SQK) physics methodology [87] [88] [89]. His main intention was to test the validity of this photon energy loss prediction against astronomical data. La-International Journal of Astronomy and Astrophysics Violette's tired-light relation emerges from a methodology that adopts a very different approach to conceiving the nature of physical phenomena, nevertheless, is one that has been extensively published in the literature. It is advisable to summarize a bit about it here. According to SQK, all space is pervaded by a subquantum reaction-diffusion medium whose constituents are able to self-organize into concentration inhomogeneities thereby forming "bunched" field potentials at the quantum level that constitute observable subatomic particles and photons.
The underlying reaction and diffusion processes of this medium are mathematically described by an open, nonlinear reaction system, termed Model G [87]- [93]. One characteristic of such systems is that the entropy or energy magnitude of a photon (reaction-diffusion wave) does not necessarily remain constant over time as is assumed in standard physics. Instead, its initial energy magnitude may progressively increase or decrease over time depending on the state of criticality of the underlying reactions.
Wave amplitude (energy) in such nonlinear open reaction systems is generally described by the following wave equation which is applicable to reaction-diffusion waves consisting of small amplitude excursions [φ] from the ambient potential [94]: where 0 A is the initial magnitude of the wave (photon) and where κ R and κ i are the real and imaginary parts of its wave number κ. The frequency and wavelength of the wave are given respectively as f = ω/2π and λ = 2π/κ and the wave velocity is given as 0 c fλ ω κ = = . The oscillatory real term in Equation (14), the first exponent on the right, is consistent with energy wave behavior in standard physics. The imaginary term, the second exponent on the right, though, is new to quantum electrodynamics/physics. It dictates nonconservative wave damping when κ i > 0 (when subcritical conditions prevail in the reaction system) and nonconservative wave amplification when κ i < 0 (when supercritical conditions prevail in the reaction system).
Equation (14) may be restated as follows to portray the manner in which photon energy changes as a function of photon travel distance in SQK: E(r) signifies the wave's electric potential amplitude, or energy, at distance r, and is equivalent to [φ] in Equation (14). Term E 0 represents the wave's initial electric potential amplitude, or energy, at r = 0 and is equivalent to the wave amplitude term [ 0 A ] in Equation (14). The exponent ( ) e g c r αϕ − is essentially the same as the second exponent in Equation (14) where κ i = αφ/c. Here, α is a constant of proportionality, c is the velocity of light, and φ g signifies the ambient gravity potential in the wave's vicinity. This serves as the bifurcation parameter determining the system's mode of behavior. Perfect energy conservation holds when the photon is traveling through regions of space sufficiently close to galaxies where the gravity potential is at its critical threshold, zero value, φ g (r) = 0. In the im-mediate vicinity of galaxies and galaxy clusters, where φ g (r) is negative, the subquantum reaction-diffusion processes become supercritical and dictate photon energy amplification. In intergalactic space, where φ g (r) becomes positive, the underlying reaction-diffusion processes become subcritical and dictate photon energy damping. These various modes of photon behavior are illustrated in Figure 10. This ability for gravity potential to affect the nature of photon energy conservation according to the gravity theory of SQK, of course, is not predicted by general relativity. Nevertheless, the gravity theory of SQK does predict effects consistent with all other aspects of general relativity [91] [95].
Since a photon from a distant galaxy would spend far more time traveling through subcritical intergalactic void regions than through supercritical regions surrounding galaxy clusters, its energy on average would progressively decline, in accordance with Equation (15). The average rate of energy attenuation that a photon would experience during its flight may then be expressed as: where β, the average attenuation coefficient takes the place of term αφ g /c in Equation (15). Expressed in terms of photon wavelength, λ, this would be rewritten as: This is essentially the same as the "tired light" relation which historically has been devised to explain cosmological observations. Or, if photon energy loss and redshift occur in quantum increments, as portrayed earlier by Equation (13) This SQK cosmological redshift relation was not observationally motivated to explain cosmological data, as had been the case for other tired-light models, but rather emerged as a corollary of the Model G reaction-diffusion system. The value of β given above is chosen to be β = H 0 /c, where H 0 is the observed value for the Hubble constant. Here we may choose the value H 0 = 64.0 ± 1.6 km/s/Mpc, Figure 10. According to SQK, a photon's energy behavior depends on the ambient value of the gravity potential in the photon's vicinity judged relative to the critical threshold value, φg = 0. Photons are predicted to progressively increase their energy within the supercritical gravity wells that surround galaxies and galaxy clusters (where φg(r) < 0), and to progressively decrease their energy in intergalactic space (where φg(r) > 0). as proposed by Tammann and Reindl [43], which falls close to the H 0 value indicated by the fit of the tired light relation on the t s -z cosmology tests of Marosi discussed above. This yields a photon energy decline rate of β = 6.54 ± 0.16% per billion light years (bly). Expressing β in time units, by multiplying by c = 3.17 ×10 −17 bly/s, yields an energy loss rate µ = −βc = −2.07 × 10 −18 /s. This is about 10 orders of magnitude smaller than the smallest change observable in the laboratory. Hence such a nonconservative energy loss certainly has no observable influence on laboratory measurements. It becomes important only at the astronomical scale, and as has been seen above, it makes a good fit to astronomical data.
The cosmology test results discussed in Section 2, which vindicate the no-evolution, static universe tired light cosmology, are generally valid for all tired light models discussed above. However, the SQK tired light relation has a number of advantages over other tired light theories that have been proposed. First, since it predicts that the energy of the photon decays over time without emitting secondary photons from recoils with intergalactic particles, the photon should suffer no angular deflection that might contribute to image blurring. 3 Second, by assuming that Model G operates sufficiently close to its threshold of marginal stability, its tired light energy loss effect will show no detectable frequency dependence. Hence radio frequency photons should redshift at the same rate as visible photons [91].
Third, the SQK paradigm, predicts that as stellar gravity fields extend outward away from their parent galaxy, beyond ~3 kpc from their source masses, their potentials decline and ultimately approach the steady state zero-gradient potential value present in intergalactic space. This is illustrated in Figure 10 by the ambient value attained in the subcritical space between the two galaxies. Hence over great distances the force of gravity departs from a Newtonian radial inverse square decline, as is the case in Milgrom's theory of modified Newtonian dynamics (MOND). As a result, in SQK, there is no gravitational potential summation problem leading to universal gravitational collapse, a difficulty that plagues many static universe cosmologies. Like the tired light effect, this limited range for gravity emerges as a prediction of SQK Model G and is not introduced as an ad hoc assumption [91].
A fourth advantage of the SQK tired light model is that its energy attenuation coefficient can vary according to the value of the ambient gravity potential. Thus, photons crossing subcritical void regions where the ambient gravity potential is substantially positive would exhibit a Hubble constant value much greater than that characterizing their passage through more typical regions of intergalactic space where their trajectory would occasionally traverse supercritical gravity well regions associated with clusters and superclusters. Such a gravity 3 In their Tolman test cosmology paper, Lubin and Sandage [23] incorrectly suggested that the tired light mechanism that LaViolette [2] described in his 1986 paper accomplished its photon energy loss through photon scattering and commented that such scattering would inappropriately cause the images of distant galaxies to blur. This was an apparent misreading of his paper since LaViolette instead proposed a spontaneous nonconservative energy loss mechanism for the cosmological redshift. potential dependency could explain why voids in redshift space are found to appear elliptical with their direction of elongation oriented along the line of sight to observer, this being the so-called Fingers-of-God effect. That is, if photons undergo a greater than normal redshift rate as they travel through a void, galaxies on the void's far side would appear to have a greater than normal redshift and therefore be displaced away from the observer in redshift space, creating an apparent elongation [91] [95] [96]. Alternatively, SQK predicts that blueshifting would occur when photons pass through galactic clusters or superclusters, an effect that is able to account for the Kaiser effect and for the Fingers-of-God effect seen there as well [91] [95] [96].
A fifth advantage of the SQK tired light model is that the Model G reaction system from which it is derived also predicts that matter should be continuously created in supercritical regions of space. Consequently, with the emergence of the static universe tired light model in cosmology, SQK has a matter creation theory ready to explain the origin of the universe in lieu of the big bang explanation; see next section.

A Theory of Continuous Creation
The cosmological test data results summarized above lead inevitably to the conclusion that the universe must be globally static, and that the cosmological redshift is due to a process other than recessional velocity. We find then that the big bang theory is no longer a viable theory to account for the origin of things and that we must look elsewhere for alternatives to explain the nucleation of matter and energy quanta, a cosmology that does not require creation to take place all at once in a singular primordial space-time explosive event. One is then left to consider the possibility that matter might be continuously created throughout a cosmologically static universe. In some ways this scenario is more plausible than a big bang. For instead of being asked to believe that all the matter-energy of the universe was born into existence in a brief instant, we may now consider individual subatomic particles springing into being in a leisurely manner throughout the vastness of space at a rate so slow as to be virtually undetectable in the laboratory.
The idea that matter might be continuously created in regions of high matter density dates back to Sir James Jeans [97] whose observations of galaxies led him to speculate that there is a progressive evolution of galactic morphology from elliptical to spiral which involves a centrifugal ejection of material from their centers. Some years later he proposed that matter was being continuously created throughout the universe, at a time when the universe was believed to be cosmologically static. He held that the centers of galaxies may be similar to singular points at which matter pours into our universe from some other external dimension [98]. In his scheme, matter creation occurs in a nonconservative manner, as he essentially proposes that our universe may behave as an open system, at least in the cores of galaxies which he believed to be the most active creation centers.
Jean's idea of galaxy evolution was supported by Hubble [99] who found that for galaxies of a fixed total magnitude, the major axis diameter of a galaxy progressively increases as one proceeds through the morphology sequence from E0 to Sc. He embodied this sequence in his well-known "tuning fork-like" galaxy classification scheme. This implied to him that the size and mass of a galaxy progressively increases as one proceeds from early to late type galaxies. He wrote that the entire series can be represented by various configurations of an originally globular stellar mass expanding equatorially. William McCrea [100] was also led to the idea of continuous matter creation on the basis of astronomical observation. He proposed that matter is continuously created throughout space, with creation being assumed to proceed most rapidly in regions of negative gravitational field potential, e.g., within stars and condensed masses. He proposed that all matter may potentially promote the creation of new matter. Furthermore, he proposed that since all matter normally resides in galaxies, the creation of fresh matter promotes the growth of galaxies. He notes that when on occasion a clump of matter becomes detached from its galaxy, it may serve as the embryo for a new galaxy.
Victor Ambartsumian [101] [102] further developed Jeans' idea proposing that a galaxy's evolution was largely shaped by energetic events taking place in its nucleus. Based on his observations of a variety of galaxies, he concluded that supermassive cores, most evident within the more massive galaxies, not only emit tremendous amounts of energy during their active phase, but large quantities of matter as well. He proposed that matter is thrown off in the form of an intense flux of relativistic particles as well as in the form of nonrelativistic ejections. He suggested that the latter would include gas moving at speeds of thousands of km/s, eruptive ejections of entire gas clouds, and even ejections of entire "galaxy embryos", i.e., supermassive cores. J. L. Sérsic [103], who echoed the ideas of Ambartsumian, suggested that giant elliptical galaxies with active cores may on occasion explosively fragment to produce progeny galaxies which could develop into various morphologies. Halton Arp [104] [105] also espoused similar ideas. Based on his observations of active galaxies, he concluded that active galactic nuclei can fission and explosively eject highly energetic clumps of matter, usually at steep angles to the galactic plane.
The novel physics methodology of SQK predicts a nonconservative matter creation cosmology that is consistent with the proposal of McCrea, as well as those of Jeans, Hubble, Ambartsumian, Sérsic, and Arp. Besides predicting tired light energy loss for photons traveling through intergalactic space, the Model G open reaction-diffusion system also predicts that matter may be spontaneously created in supercritical regions of space, this occurring wherever the ambient gravity potential lay slightly below its critical threshold zero value [89] [91] [106]. Such matter-spawning supercritical regions would prevail over large stretches of space scattered throughout the cosmos. In such regions, a zero-point energy fluctuation of sufficient magnitude is able to nucleate the formation of a neutron. Just as the energy of a photon would continuously blueshift in a supercritical environment, so too a zero-point energy fluctuation of critical size, sto-chastically arising in a matter-free region of space, would be able to grow in size. Rather than growing indefinitely, its magnitude instead converges to a new stable state, locally changing the former homogeneous steady-state of the reaction-diffusion system, the vacuum state, to an inhomogeneous steady-state as it forms a stable subatomic particle, i.e., a dissipative soliton wave pattern of well-defined wavelength and field potential magnitude. In SQK, this matter creation process is termed parthenogenesis, meaning virgin birth.
This parthenogenic matter creation process is illustrated in computer simulations of the partial differential equation system that constitutes Model G. A simulation showing the successive growth of an initial electric potential seed fluctuation into a dissipative soliton field pattern representative of a neutron is shown in Figure 4 of Pulver and LaViolette [93]. Also, a video simulation showing this particle materialization process may be viewed at: https://tinyurl.com/ybfphshf. As LaViolette [90] has shown, the emergent neutron has a form closely resembling that observed for the neutron in particle scattering experiments conducted by Kelly [107]. That is, the electric charge density distribution of the neutron's core is found to have a haystack-like shape and be surrounded by a radial periodicity. In fact, the electric potential field pattern for the Model G neutron yields a closer resemblance to Kelly's results than any previous nuclear field model [90].
To nucleate a neutron, the seed fluctuation must be of positive charge since in Model G positive fluctuations generate supercritical gravity wells which allow the fluctuation to grow and ultimately spawn a fully grown particle. Once formed, the particle's φ g (r) well continues to stabilize the particle, allowing the particle to persist in spite of potentially destabilizing zero-point energy fluctuations arising in its an environment. Seed fluctuations of negative potential, which might otherwise spawn the antiparticle state, e.g., the antineutron, fail to grow spontaneously since they generate a subcritical gravity potential hill in their vicinity and hence are self-extinguishing. Because of this matter-antimatter bias, the SQK matter creation process leads to a universe filled mainly with matter, rather than antimatter. This is an advantage since to date there has been no detection of antimatter galaxies. The apparent lack of an equal amount of antimatter in the universe has been a major setback for the big bang theory.
Besides creating a supercritical gravity well in its core, the SQK neutron also produces a peripheral gravity well shell at one Compton wavelength intervals from its center. The inner most two gravity well shells serve as supercritical regions where a randomly emerging zero-point energy fluctuation can self-amplify and nucleate a daughter neutron. Computer simulations of Model G have borne out this mother-daughter matter creation process; see Pulver and LaViolette [93]. Also, a video simulation showing this mother-daughter materialization process may be viewed at: https://tinryurl.com/yde4bvk3.
Once a neutron self-nucleates there is a very high probability that it will decay into a proton, beta particle, and neutrino before it has a chance to spawn a progeny particle. So, protons being stable particles would tend to accumulate in space as the most common matter nucleation sites. According to SQK, the pro-ton has a periodic Turing wave pattern very similar to that of a neutron, except that its core electric field would be positively biased relative to that of a neutron. Such biasing of the proton's wave pattern was also borne out by the particle scattering results of Kelly [107] which reveal that the proton's charge density profile is in fact positively biased relative to that of the neutron.
The probability of matter creation occurring in the immediate vicinity of an existing nucleon greatly outstrips that of a particle nucleating on its own in empty space. Consequently, with Model G this mother-daughter creation process becomes the dominant means of matter creation, with matter producing more matter at an exponentially increasing rate. Although initially devised for the purpose of modeling subatomic particles, Model G leads to a continuous creation scenario that is generally compatible with that of McCrea [100] wherein matter is continuously created throughout space with the creation process proceeding most rapidly in regions of negative gravitational field potential.
Nascent particles would first emerge as neutrons, whether they did so autonomously or assisted by the supercritical region of an existing nucleon. Neutrons emerging in the vicinity of a proton could either detach to continue as isolated neutrons that would later decay or could form a nuclear bond with its proton parent transforming it into a deuteron. Alternatively, a deuteron could serve as a nucleation site for a nascent neutron which could either detach from its parent or form a nuclear bond, transforming the deuteron into a tritium particle. This nuclear transmutation sequence could conceivably repeat causing a tritium nucleus to transform into 3 He, and finally into 4 He. Further transformation would be unable to occur through parthenogenic creation. Admittedly, the synthesis of D, 3 He, and 4 He would be rare events. The main production of these elements, including Li and the light elements would occur later through thermonuclear fusion once stars had formed and grown sufficiently massive. Burbidge and Hoyle [108] have shown that stellar fusion could account for the observed abundances of these elements and that big bang nucleosynthesis is unnecessary.
Beta decay protons and electrons would be the most common particles existing in primordial times hence producing a diffuse ionized hydrogen gas heated to an X-ray emitting plasma state by collisions with the 0.78 Mev beta particles. This could account for the observed diffuse X-ray emission that is observed coming from all directions of space. This X-ray emitting intergalactic gas has been referred to as the Warm Hot Intergalactic Medium, or WHIM. Its presence is also indicated by the so-called Lyman alpha forest, diffuse Lyman alpha emission radiated by the ionized portion of this gas. The big bang theory cannot account for its temperature because it predicts that the gas of its initially hot fireball should have long ago cooled down. Crawford [33] [109] has shown that the electrons in such a heated X-ray emitting plasma have a temperature and density sufficient to generate the observed 2.73˚K cosmic microwave background radiation (CMBR). The source of ionizing radiation for the WHIM has puzzled astronomers since no stars are visible in these clouds and radiation from active ga-International Journal of Astronomy and Astrophysics lactic cores falls short of the energy requirements. LaViolette [91], however, has shown that beta decay of parthenogenic neutrons would supply more than enough energy to power this emission. Moreover Arp, et al. [110] have argued that iron whiskers present in intergalactic space could thermalize the 3˚K radiation field while allowing transparency at other wavelengths. So, with the theory that the WHIM may be the source of the CMBR we still retain the idea that the microwave background is of cosmic origin, energized not from a big bang, but from beta particle radiation arising from the decay of continuously created neutrons. Although, Burbidge and Hoyle [108] have also suggested that hydrogen burning in stars may be an energy source contributing to the 3˚K radiation.
There is no 13.8 to15 billion-year time restriction in which to generate the cosmos in the SQK continuous matter creation cosmology. Once some regions in the WHIM had cooled sufficiently to allow hydrogen to condense into its liquid state, hydrogen could eventually coalesce to form comet sized bodies. Such planetesimals would grow both through accretion and through internal matter creation. As time passed, the planetesimal would grow in size, becoming a brown dwarf, then a red dwarf, and then a sun-like star. In the early stages of matter creation, isolated particles would have been the predominant sites of matter creation. Later, once matter had condensed to form material bodies and stars, the matter creation rates per nucleon would have risen considerably since the gravity potential well for such bodies would be far deeper than that of an isolate baryon. Such bodies would be the seat of both matter creation as well as energy creation through photon blueshifting.
The rate at which neutrons would self-create in the vicinity of nucleons located inside a star would depend not only on the ambient gravity potential within the star, which in turn determines the prevailing degree of supercriticality, but also on the star's internal temperature. Temperature would be a factor since thermal collisions would induce field potential fluctuations in the star's nucleons which would help to excite the materialization of neutrons. In general, the rate of matter creation per nucleon within a stellar body would far outstrip that occurring in space. Equation (19) gives a relation, admittedly tentative, of how a stellar body's rate of matter creation (g/s) may be modeled: k g = 10 −40 s/cm 2 /K being a constant of proportionality, g ϕ (cm 2 /s 2 ) being the body's average internal gravity potential, M (grams) being the body's mass, and T (K) being its average internal temperature [91] [106].
As the star's core grew past a critical temperature and density, nucleosynthesis would commence. The gas expelled in the star's stellar wind would generate a surrounding nebula in which orbiting gaseous planets would form. Over time these daughter planets would themselves grow into stars. Meanwhile the primordial mother star would continue to grow, proceeding up the stellar main sequence until it became a blue supergiant. This would either explode as a super-nova or discharge its atmosphere leaving behind a white dwarf, bare stellar core. This core would not be a dead star energized only by matter accretion, as standard physics proposes, but would be a stellar body that would continue to grow by internal matter creation. Hydrogen continuously generated in its interior through parthenogenesis would continue to fuel its fusion reactions. But in addition, there would be supplementation by energy spontaneously created in its interior through nonconservative photon blueshifting. As a result, a stellar core would not cool off, but would continue to radiate energy mainly in the form of x-rays and cosmic rays.
This photon blueshifting prediction of Model G may be understood as follows. In supercritical regions of space, that is where φ g (r) < φ gc , photon energy will progressively increase in nonconservative fashion where Equation (16) now becomes expressed as: (20) For φ g < 0, β here takes on the role of an amplification coefficient, in effect acting as a "negative Hubble constant" that dictates exponential energy increase rather than tired light energy decrease. Evidence that photons exhibit a negative Hubble constant when passing through the gravity well of a galaxy cluster or supercluster is discussed by LaViolette [91] [96]. This relation is similar to that described earlier for the SQK cosmological redshift prediction, but with an amplification coefficient of opposite sign. Since observation shows that the cosmological redshift does not occur continuously, but takes place in discrete quantized jumps as mentioned earlier, a discrete energy increase of form similar to that described in Equation (13) may be inferred for this cosmological blueshifting phenomenon, in which case Equation (20) would be written as: ( ) 0 e n r E r E β∆ = . (21) The excess energy evolved from photon energy amplification is termed genic energy. The genic energy of a body of mass M would be given as: where H represents the body's total heat capacity, given by the product of its average gravity potential g ϕ , mass M, average specific heat C , and average internal temperature T . Among other things, this photon blueshifting prediction has been found to successfully predict the planetary-stellar M-L relation [111]. The reader is here referred to LaViolette [91] for further information on the astronomical significance of this blueshifting effect.
According to the SQK paradigm, a stellar core is unable to collapse into a black hole singularity due to its continuous creation of genic energy; see LaViolette [91], Section 9.9. For example, since gen L , the rate of genic energy production depends on both gravity potential and temperature, as depicted in Equation (21), L gen of a collapsing stellar core would scale according to 1/R 2 , where R is stellar radius. That is, assuming that . Furthermore, since the surface area of the collapsing core varies as P. A. LaViolette 1/R 2 , the genic energy radiation pressure per unit surface area opposing collapse would vary according to 1/R 4 . The inward pull of gravity opposing this, however, would increase only as 1/R 2 . Consequently, with radiation pressure increasing faster than the inward force of gravity, a point would be reached where a core would ultimately cease to collapse. Consequently, a singularity would be unable to form.
Another reason why a black hole would not be able to form in SQK is because according to this physics the gravitational field potential within a subatomic particle should taper to a zero gradient at the particle's center. Hence as particles within a collapsing stellar core were pressed increasingly close together, the gravitational force attracting them to one another would approach zero. It is also worth noting that stellar cores may be supported from collapse entirely by genic energy radiation pressure, rather than by electron degeneracy. Consequently, in the cosmology of SQK, a supermassive galactic core would likely be a nondegenerate stellar core, rather than a black hole.
As stars continue to proliferate, they would collectively form a star cluster, and at their center would lie the supermassive stellar core that had given birth to their lineage. By this time the mother star core would have grown to hundreds to thousands of solar masses and would be expelling a wind of ionized gas and cosmic ray particles. Based on Equations (19) and (20), it is evident that the rate of matter and energy creation, and hence the nonconformance with energy conservation, would be most extreme in such supermassive cores, which is compatible with the ideas of Jeans, Ambartsumian, Sérsic, and Arp as mentioned earlier.
SQK refers to such stellar cores as mother stars, the name highlighting the characteristic that a mother star would serve as a galaxy's primary matter and energy birthing site, being the most supercritical region in a galaxy.
As the star cluster continues to proliferate and grow in size, it eventually turns into a dwarf elliptical galaxy with its less massive stars orbiting about the Mother star along a preferred orbital plane. As the Mother star grows further in mass and creates an increasingly supercritical internal environment, its occasional outbursts would become increasingly energetic. Upon reaching a mass of a hundred thousand solar masses or more, its outbursts would begin to resemble those seen to come from the cores of Seyfert galaxies. These would propel stars and gas outward, causing the dwarf elliptical galaxy to evolve into a compact spiral galaxy and over time into a mature spiral galaxy. This progressive evolution from dwarf elliptical, to S0 spiral to Sc spiral would match the galaxy evolution progression suggested by Jeans and Hubble. Some of the more violent outbursts would cause the Mother star to fission and spew out a part of itself as a star cluster or even as a core embryo that could grow into a dwarf daughter galaxy. Halton Arp [104] [105] has catalogued many examples of what appear to be such core ejections; also see Sérsic [103]. These daughter bodies would orbit the spiral galaxy, forming a star cluster halo around the galaxy as well as spawning satellite galaxy progeny, as is considered to be happening in our own Milky Way. Eventually, as a result of continued core ejections, the spiral galaxy develops a spher-International Journal of Astronomy and Astrophysics ical shape and evolves into a giant elliptical galaxy [91].

Conclusions
Arp [104] has stated that the observational evidence against the big bang theory has become overwhelming and that in reality the theory has been toppled. This is also the conclusion of the present overview. As shown above, the no-evolution, tired light model makes a better fit than the expanding universe hypothesis when compared to the observational data of seven cosmology tests. Including the results of the radio galaxy differential number count test [2], the total number of tests favoring the no-evolution, static universe tired light model comes to nine.
Also, it is concluded that the supernova light curve test of Goldhaber, et al. is flawed by selection effect biases. Studies of X-ray bursts, gamma ray bursts, and quasar variability show no evidence of time dilation with increasing redshift. So based on the current evidence, one may conclude that the universe must be cosmologically stationary. Furthermore, the finding that the cosmological redshift values are quantized introduces a serious challenge to the Doppler redshift interpretation. The tired light theory fares much better since discrete quantum energy transitions are commonly known to microphysics. Hence tired light photons may be assumed to lose energy in discrete quantum steps.
Of the tired light theories that have been proposed, most account only for the cosmological redshift phenomenon. They do not simultaneously provide a matter creation cosmology that may replace the big bang theory. The subquantum kinetics physics paradigm, on the other hand, predicts tired light photon energy loss in intergalactic space and also provides a mechanism for the continuous creation of matter. Furthermore, SQK has been shown to spontaneously produce matter rather than antimatter, something that the big bang theory fails to do.
Also, this physics predicts that a galaxy's gravity potential field should begin to depart from a Newtonian decline at distances greater than about 3 kpc and ultimately plateau to a finite local extragalactic gravity potential value. This not only provides an answer as to why the static universe does not spontaneously undergo gravitational collapse, but also, in accordance with the conclusions of MOND, makes it unnecessary to assume the presence of dark matter in galaxies. Finally, the SQK tired light relation, whose energy attenuation coefficient is gravity potential dependent, may provide a resolution to the Fingers-of-God effect. If the present tight grip on the First Law can be ever so slightly relinquished, a new era should await the future of physics.

Conflicts of Interest
The author declares no conflicts of interest regarding the publication of this paper.