Can Information Theory Help to Formulate an Optimal Model of a Physical Phenomenon?

This article deals with the problem of calculating the comparative uncertainty of the main variable in the model of the studied physical phenomenon, which depends on a qualitative and quantitative set of variables. The choice of variables is determined by preliminary information available to the observer and dependent on his knowledge, experience and intuition. The finite value of the amount of information available to the researcher leads to the inevita-ble aberration of the observed object. This causes the existence of an unre-movable and intractable processing by any statistical methods, a comparative (respectively, relative) uncertainty of the model. The goal is to present a theoretical justification for the existence of this uncertainty and proposes a procedure for its calculation. The practical application of the informational method for choosing the preferred model for the Einstein formula and for calculating the speed of sound is demonstrated.


Introduction
Intuition, life experience and knowledge (the human paradigm ref. [1]) have creative power because they create those prisms through which we look at the world. Formulating a model of the phenomenon under study, the researcher creates a picture that is unique to him and reflects his philosophical view of the order of things that exist regardless of his desire. How accurately the observed process is recreated can only be clarified by carrying out accurate measurements.
Any experiment is preceded by the process of building a model of the phenomenon being studied by the observer. The great Newton used three basic va-riables to formulate laws: length (m), mass (kg) and time (sec). Only in the 18th-19th centuries scientists proposed and began to use other basic variables: I-unit of electric current-ampere (A), θ-unit of thermodynamic temperature-kelvin (K), F-amount of substance (mol), J-unit of luminous intensity-candela (cd). On the basis of various variables, systems of units were built: SI, British, Planck system, CGS. However, scientists approached the idea of the need to use the system of units only in the 20th century. Despite the fact that systems of units were created by human genius and did not exist in nature, they are an indisputable tool for understanding the world around us. When formulating a model and choosing certain variable, the researcher is "doomed" to use any system of units.
Depending on the subject area under study, the problem being solved, on the mathematical background of the researcher and the requirements of the customer, mathematical models can have various forms and ways of representation.
The model plays a system-forming and sense-forming role in scientific cognition; it allows us to understand the phenomenon and the structure of the object under study. Without building a model, it is unlikely to be able to understand the logic of the system and conduct a high-precision experiment. This means that the model, from the position of a particular researcher, allows you to decompose the system into elements, connections, mechanisms, explains the operation of the system and determines the cause-and-effect relationships and the nature of the interaction of the variables taken into account. In the simplest case, modeling technology involves at least three steps: 1) formalization, 2) actual modeling, 3) interpretation. A huge number of studies are devoted to the second and third stages ref. [2]. This article will consider only the process of formalization (identifying the essential features of the object under study and describing the relationships between them using the language of mathematics ref. [3]), which, in the opinion of the author, has not received sufficient attention. This is also due to the fact that it is at this stage that the minimum total costs can be achieved, including the damage from its application (inaccuracy) and significant resources spent (time, cost, reasonable complexity) for constructing the model.
It is equally important to note that the formalization stage determines the choice of the required measuring complex and the technology for implementing the experiment.

Initial Background
In the process of model formalization, there are two entities: the system of units like the International System of Units (SI), which "provides" the model with information about the phenomenon under study using dimensional and dimensionless variables, and the researcher, who takes into account all available information and selects certain variables from the SI using intuition, knowledge and experience to analyze the uncertainty of the model. In relation to the formalization process, this may mean that a certain likelihood function is created in the researcher's Given the above and taking into account the current interest in the applications of information theory in various types of human activity ref. [4] [5] [6], some researchers recognize that it is possible to consider the model as an information channel between the observed object and the researcher ref. [7].
In this case, you can represent the input data in the form X{x 1 , ..., x j }, as the total number of variables (hereinafter referred to as variables with finite information, FIQ ref. [8]) inherent in the observed physical object. The number of variables in X is equal to the number of FIQs contained in the system of units used by the researcher; for example, SI, μ SI = 38.265 ref. [9]. Y{y 1 , ..., y p }-output data, the number of FIQs included in the model at the will of the observer. Y is a "noisy" version of X, in which the actually observed phenomenon "shrinks" in size, i.e., the number of FIQs is significantly reduced, but without energy consumption. This is explained by the fact that at the stage of formalization, the observer only considers the model and does not introduce any interference into the real process. Given that μ SI is a constant, each FIQ has a limited amount of information ref.
where -Δ is the a priori model absolute uncertainty (systematic effect ref. [12]) caused by choice of the CoP and the number of recorded FIQs, S is the interval of the observation of the main researched FIQ chosen by the observer; -z' is the number of FIQs in the selected CoP, β' is the number of base quantities in the selected CoP, z" is the number of FIQs recorded in a model, and β" is the number of independent quantities recorded in a model; μ is the number of dimensionless FIQs that can be constructed based on the seven base SI quantities, μ = 38,265 ref. [9]. The concept of "relative uncertainty" r is clear to all scientists and is widely used in science and technology. However, r can only serve to reflect a subjective judgment ref. [13]. In addition, r does not imply the need to indicate the results of measurements and at the same time, consider the measure of confidence in them in the form of an interval within which most of the distribution of the values of the measured variable lies. Therefore, as a universal indicator for a quantitative assessment of the proximity of the model to the object under study, the comparative uncertainty of the model ε = Δ/S is proposed. Until now, researchers have not considered the value of ε, although it is of vital importance in information theory ref. [14]. Equation (1)-the "ε-equation"-applies to models that use both dimensional and dimensionless FIQs ref. [15] [16]. It should be noted that ε cannot be statistically tested using tools such as consistency, asymptotic normality, weighted estimates, or coefficients.
Following the proposed FIQ-based approach, the results of scientific research are recommended to analyze from the perspective of comparing the comparative uncertainty achieved in the model ε mod with the theoretically justified ε opt ( Table  1). The ratio ε mod /ε opt is an objective criterion for establishing the acceptability of a particular model, measurement method and assessing the accuracy achieved when comparing different models for one specific physical phenomenon or technological process. The similarity between these two uncertainties (1 > ε mod / ε opt ≈ 1) proves the applicability of the proposed model in describing the process being studied. Conversely, a significant difference between these uncertainties indicates that the proposed model is unreliable. Importantly, there is no guarantee that this limit (ε opt ) will ever be reached (ε mod is always less than ε opt ), regardless of the advances made by scientists and engineers. The following analysis of research results will highlight the obstacles that must be bypassed or overcome before the various objectives can be achieved.
This approach indicates the possibility of using the concept of information transfer, accumulation, and transformation directly in information theory, as well as in theoretical studies in other areas of knowledge, including applied problems.
Using data of ref. [4], the optimal values of ε opt can be identified for different CoPs and the recommended number of FIQs corresponding to each CoP, as shown in Table 1.
Analyzing the data from Table 1, it should be noted, firstly, that with an increase in the number of basic variables in CoP, the optimal number of variables used in the model and necessary to achieve the optimal value of comparative uncertainty sharply increases. Secondly, the use of a model with a CoP containing a small number of basic variables, for example, LMT or LMTF, makes it impossible to achieve the optimal value of comparative uncertainty. This indicates the inexpediency of using a model with a low number of basic variables to describe the physical object under study.

Light Speed
Suppose that within the framework of the presented informational method, the researcher sets out to "improve or modify" Einstein's formula.
where E is energy, m is the mass of the object, c is the speed of light, c = 299.792.458 m·s -1 ref. [17]. The formula corresponds to CoP SI ≡ LMT. It must be said that, at present, measurements using this formula are carried out with a relative error of 4 × 10 −7 ref. [18]. Despite these results, this does not mean that Einstein's ideas will always be true. Future physicists will no doubt test it even more precisely because more precise tests imply that our theory of the world is, in fact, becoming more and more perfect.
Let's imagine that from the standpoint of a researcher who proceeds from his own philosophical views, the temperature of the environment surrounding the object should be taken into account. It should be noted that the expediency of B. Menin Journal of Applied Mathematics and Physics taking into account temperature in the relativistic theory was noted in ref. [19].
Taking into account the "Landauer limit" ref. [20] and the results of ref. [21], in which it was proved that the amount of information of any physical system must be finite if the space of the object and its energy are finite, can be proven ref. [22] [23] that ( ) ( ) where ϒ E is the energy of the object contained in the sphere of radius R in terms of ordinary energy (m 2 ·kg·s -2 ) and includes the total mass-energy of the observed sphere E and the energy due to the information ϒ I contained in the object, ħ is the reduced Planck constant, k b is the Boltzmann constant, θ is the ambient temperature. The formula corresponds to CoP SI ≡ LMTθ.
Let us analyze (2) and (3) from the point of view of comparing the comparative uncertainty ε mod achieved in the model with the theoretically substantiated ε opt . Considering the dimensions of FIQ in (2), the problem belongs to CoP SI ≡ LMT, and we can assume that z'' − β'' = 1 (according to the π-theorem ref. [24]. For (2) where z' − β' = 91 is the number of dimensionless complexes for CoP ≡ LMT and ε LMT = 0.0048 (Table 1). ε 1 /ε LMT = 2.79.

Taking into account the dimensions of FIQ in (3), the problem refers to CoP SI
≡ LMTθ, and we can assume that z'' − β'' = 3 (according to the π-theorem ref. where z' − β' = 846 is the number of dimensionless complexes for CoP SI ≡ LMTθ and ε LMTθ = 0.0442 (Table 1). ε 2 /ε LMTθ = 0.58. Analyzing ε 1 /ε LMT =2.79 and ε 2 /ε LMTθ = 0.58, we can draw the following conclusions. Although the value of 0.58 is much less than 1, from the point of view of the informational method, the inclusion of temperature in Einstein's Equation (3) is perhaps an advance in understanding the nature of the surrounding world. Undoubtedly, Einstein's formula is considered the greatest achievement of the 20th century, embodies the idea of simplicity and depth of scientific thought, is consistent with a set of known experimental results, and also allows you to make predictions on obtaining new scientific data. However, given that ε 1 /ε LMT = 2.79 > 1, it can be assumed that the theory of relativity can be based on the subjective philosophical view of the researcher at the most fundamental level, which raises deep epistemological questions about the fundamental nature of reality. At the same time, the informational approach is designed to reveal the smallest deviations from the generally accepted principles of modeling physical phenomena, which can give the first signs of new physics. This equation is formulated in terms of the Helmholtz energy and can be used to calculate all thermodynamic properties through combinations of the function itself and derivatives with respect to its natural variables. The speed of sound was considered as a function of temperature, pressure, path length difference and time. Thus, the model for measuring the speed of sound refers to CoP SI ≡ LMTθ. There was mentioned that because the speed of sound measurements for hydrogen chloride in the liquid and dense vapor phases are the first data for this property in the literature, they are important for modeling caloric data. The accuracy of the equation was analyzed by extensive comparisons to experimental data. Furthermore, the physical and extrapolation behavior of the equation of state was carefully monitored, which is an important aspect for the application to mixture models.

Sound Speed
The speed of sound was considered as a function of temperature (θ), pressure (p), path length difference (l) and time (T). Thus, the sound velocity measurement model refers to CoP SI ≡ LMTθ. We can assume that z'' − β'' = 1 (according to the π-theorem ref. [24]. For this case, theoretically achievable comparative uncertainty equals: where z' − β' = 846 is the number of dimensionless complexes for CoP SI ≡ LMTθ and ε LMTθ = 0.0442 (Table 1). In this way, In contrast to the previous example, suppose that a group of researchers set out to reveal the deep physical nature of the speed of sound. So, scientists proved ref. [26] that the upper limit of the speed of sound in condensed phases depends on the combination of two important dimensionless fundamental constants, the fine structure constant α and the ratio of the masses of an electron to the mass of a proton:

B. Menin Journal of Applied Mathematics and Physics
The research was carried out in the framework of CoP SI ≡ LMTIF. It means that the dimensions used as the dimensions of the bases L, M, T, I, and F, in different degrees ref. [11]. A total of 22 (z") variables were used to calculate vu. For the case involving the selection of four independent variables (β" = 4), in accordance with the π-theorem ref. [24], the number of dimensionless criteria in a model, γmod, equals γ mod = z" − β" = 18.
Unfortunately, in that study ref. [26], the authors did not indicate the ranges of variation and the measurement of uncertainty for each considered variable. Additionally, the total absolute uncertainty of the key parameter (the speed of sound in condensed phases) was not calculated. This information allows for the possibility of comparing the theoretical comparative uncertainty, ε opt , calculated using (1), with the experimental comparative uncertainty, ε 4 .
Despite the lack of knowledge about the proposed information method, the study authors presented a very plausible model and result comparable to those obtained in ref. [27], in which researchers reached ε mod /ε opt ≈ 0.9 with 130 variables selected in the model.
In another study ref. [28] the performance of three established reference thermodynamic models based on the Helmholtz energy for binary mixtures (N 2 + H 2 ) is evaluated by comparing experimental sound speed data and acoustic virial coefficients with the results predicted by these reference thermodynamic models. In addition, this study provides new and more accurate experimental data, which form the basis for improving these models.
The achieved relative uncertainty is 220 × 10 −6 (0.022%). The number of va- Taking into account (7), (9), (10), it follows from the above reasoning and calculations that the idea that simplicity (a small number of variables in the constructed model that reflects the observed object from the point of view of the researcher) is the path to truth turns out to be far from reality. It is precisely this characterization of simplicity in the hitherto discovered laws of nature that it would be erroneous to generalize since it is obvious that simplicity was one of the reasons for their discovery and, therefore, cannot serve as a basis for supposing that other undiscovered laws are just as simple.

B. Menin Journal of Applied Mathematics and Physics
Thus, the achieved maximum accuracy of the representation of the observed physical phenomenon actually depends on the qualitative and quantitative set of variables in the model.

Discussion
Since 1927, Heisenberg's uncertainty principle has put an end to dreams of a fully knowable world and the hope that only human ingenuity limits the accuracy of measuring things. Since 2017 ref. [9] it became apparent that the progress toward more accurate measurements is also limited not only by quantum uncertainty but also by human consciousness. This is explained by the fact that when building a model that precedes any experiment, a conscious observer uses (forced, obliged) some system of units, including a finite qualitative-quantitative set of variables. This, in turn, determines the non-infinitely small amount of information contained in the model. Thus, initially, before conducting any experiment designed to confirm or refute the proposed model of nature, the researcher is doomed to get a distorted (blurred) picture of the world. This limit is greater than the uncertainty dictated by the Heisenberg inequality.
In the theoretical modeling of physical phenomena, the decisive step is to determine those variables that allow you to get an approximate (through human thinking) but an informative idea about it. In some cases, this choice can be made based on the intuition and experience of scientists and engineers for many complex systems. We present, using thermodynamic information theory, a methodology for determining the appropriate accuracy of a model of a physical phenomenon based on the calculation of its information content. We use the informational approach to identify the class of any physical or technological phenomenon under study and the optimal number of variables to achieve its most plausible representation that satisfies the researcher's philosophical vision.
Thus, by means of the informational method, a bridge is built between the consciousness of the observer, the object of study and such a still not fully known structure as information.
The importance and diversity of the processes and phenomena under study require a much greater role from testing models than is customary in the physical and technical sciences. The informational approach offers a flexible approach to testing. We emphasize the need to use comparative uncertainty, which does not require the consistency of numerical data.

Conclusions
The results obtained can help reveal the features of the subjective experience of

Conflicts of Interest
The author declares no conflicts of interest regarding the publication of this paper.