From Kalman to Einstein and Maxwell: The Structural Controllability Revisited

Abstract

In the Special Relativity paper of Einstein (1905), only a footnote provides a reference to the conformal group of space-time for the Minkowski metric ω . We prove that General Relativity (1915) will depend on the following cornerstone result of differential homological algebra (1990). Let K be a differential field and D=K[ d 1 ,, d n ] be the ring of differential operators with coefficients in K . If M is the differential module over D defined by the Killing operator D:T S 2 T * :ξΩ=( ξ )ω and N is the differential module over D defined by the Cauchy=ad( Killing ) adjoint operator with torsion submodule t( N ) , then t( N )ex t D 1 ( M )=0 and the Cauchy operator can be thus parametrized by stress functions (Airy for n=2 , Beltrami, Maxwell, Morera for n=3 , Einstein for n=4 ) having strictly nothing to do with Ω . This result is largely superseding the Kalman controllability test in classical OD control theory and is showing that controllability is a structural “built-in” property of an OD/PD control system not only depending on the choice of inputs and outputs, contrary to the engineering tradition. Indeed, as illustrated by many examples, using any control system as a way to define the above differential operators and modules, the above result amounts to prove that the system is controllable if the adjoint operator is injective. In actual practice, we invite the reader to pick up in textbooks any example depending on some parameters, treat it by the Kalman test and make any exchange between inputs and outputs to check that the controllability conditions on the parameters are still unchanged! It also points out the terrible confusion done by Einstein (1915) while following Beltrami (1892), both of them using the Einstein operator but ignoring that it was self-adjoint in the framework of differential double duality (1995). Following Weyl, we finally prove that the structure of electromagnetism and gravitation only depends on the nonlinear elations of the conformal group of space-time, showing thus that nothing is left from the mathematical foundations of both general relativity and gauge theory. These results also question the origin and existence of gravitational waves and black holes, not because of a problem of detections but because of a problem of equations.

Share and Cite:

Pommaret, J.-F. (2025) From Kalman to Einstein and Maxwell: The Structural Controllability Revisited. Advances in Pure Mathematics, 15, 570-628. doi: 10.4236/apm.2025.159031.

1. Introduction

Let D:ξη be a linear differential operator of order q . A direct problem is to look for a differential operator D 1 :ηζ such that D 1 η=0 is generating the compatibility conditions (CC). For example, starting with q=2 and the operator d 22 ξ= η 2 , d 12 ξ= η 1 we obtain at once the single first order CC d 1 η 2 d 2 η 1 =0 . Now, multiplying on the left by a test function λ and integrating by parts, the adjoint operator becomes d 1 λ= μ 2 , d 2 λ= μ 1 with the single first order CC d 1 μ 1 + d 2 μ 2 =0 . However, multiplying now ( η 1 , η 2 ) by two test functions ( μ 1 , μ 2 ) , adding and integrating by parts, we obtain the second order adjoint operator d 22 μ 2 + d 12 μ 1 which is NOT generating the CC of ad( D 1 ) even if it is well known that ad( D )ad( D 1 )=ad( D 1 D )=0 . This result is just showing that, when D 1 generates the CC of D , then ad( D ) does not generate in general ALL the CC of ad( D 1 ) . If M is the differential module that can be defined by D , the above “ defect “ is measured by a module called ex t 1 ( M ) that only depends on M and not of the way to define it as we shall see. We shall denote by N the differential module which is defined by ad( D ) and introduce the torsion submodule t( N ) made by all the elements satisfying at least one OD or PD equation FOR ITSELF. In the present example, N is defined by d 12 η 1 + d 22 η 2 = d 2 ( d 1 η 1 + d 2 η 2 )=0 and t( N ) is generated by ν= d 1 η 1 + d 2 η 2 . The following formula:

t( N )ex t 1 ( M )t( M )ex t 1 ( N )

is a cornerstone of homological algebra and the equivalence exists because ad( ad( D ) )=D . All the second section will be presenting in a rather self-contained way all the definitions and homological results needed for understanding these new concepts and refer the reader to the Zentralblatt review Zbl 1079.93001 for understanding how difficult it is to collect in the literature all the references needed. However, sometimes it may work, that is ad( D ) may generate the CC of ad( D 1 ) or, equivalently, ex t 1 ( M )=0 as we shall see in the following motivating examples that we set up together in this Introduction. For the moment, we ask the reader to stop for a few minutes in order to imagine any link that could exist between this formula and Kalman test in control theory on one side or between this formula and Einstein or Maxwell equations on the other side.

Being a former student of A. Lichnerowicz, specialist of systems of PDE in group theory and control theory, it has been a challenge for me to apply the new methods of Differential Homological Algebra introduced around 1990 by U. Oberst [1] and E. Zerz [2] for systems with constant coefficients or by M. Kashiwara [3] for systems with variable coefficients in order to study gravitational waves. The three last papers published in 2024 [4]-[6] with a short summary in [7] could be roughly summarized by the single formula:

t( N )ex t 1 ( M )=0

where M is the differential module defined by the Killing operator and N is the differential module defined by the Cauchy=ad( Killing ) adjoint operator with torsion submodule t( N ) , as extension modules are torsion modules that do not depend on the resolution of M that MUST be used, namely the differential sequence in which the order of an operator is under its arrow, which is exact in the sense that any operator generates the CC of the previous one:

0Θn Killing 1 n( n+1 ) 2 Riemann 2 n 2 ( n 2 1 ) 12 Bianchi 1 n 2 ( n 2 1 )( n2 ) 24

Such a differential sequence sequence can be found, from a purely computational way, in any textbook of general relativity. However, its adjoint differential sequence made by the respective adjoint operators going therefore “backwards” (that is from right to left) and totally unknown, is also surprisingly exact, the stress ( σ ij ) n T * S 2 T being a tensor density and not only a tensor [4]:

0n Cauchy 1 n( n+1 ) 2 Beltrami 2 n 2 ( n 2 1 ) 12 Lanczos 1 n 2 ( n 2 1 )( n2 ) 24

This result points out the confusion (the word is weak !) done by Einstein (1915) for space-time while following the work of Beltrami (1892) for space only, both using the Einstein operator but ignoring that it was was self-adjoint in the framework of differential double duality. The Cauchy operator can be thus parametrized (backwards !) by n( n+1 )/2 stress functions having strictly nothing to do with the metric, exactly like in the case of the single Airy stress function for plane elasticity, because the Airy parametrization is only the adjoint of the Riemann operator when n=2 . Though unpleasant it is, this result questions the origin and existence of gravitational waves, not because of a problem of detection but because of a problem of equations as we shall prove that Einstein did confuse the Cauchy=ad( Killing ) operator with the so-called “divergence” operator induced from the Bianchi operator. In a rough way, it is like claiming that the single Airy (for plane) or the six Beltrami (for space) stress functions in elasticity should have something to do with earthquakes.

One striking byproduct of our claim is provided by the next example [8] but we advise the reader to read all the following examples that are presented totally independently.

Example 1.1: (Kalman system): The previous formula is the Kalman test in classical OD control theory. As any operator is the adjoint of its own adjoint because ad( ad( D ) )=D as we already said and one can thus exchange M and N in the formula, that is t( N )ex t 1 ( M )t( M )ex t 1 ( N ) . Hence, if M is the differential module defined by the formally surjective Kalman operator ( y k , u r )( d y k + A l k y l + B r k u r ) with time derivative d , inputs u and outputs y while N is the differential module defined by its adjoint operator with torsion submodule t( N ) , then we prove that the Kalman controllability test amounts to say that the given control system is controllable if and only if N=0 . Indeed, introducing Lagrange multipliers λ=( λ k ) , the kernel of the adjoint operator is defined by the OD equations ( y k d λ k + λ l A k l =0 , u r λ k B r k =0 ) with all their differential consequences, namely:

dλ+λA=0,λB=0d( λB )=( dλ )B=( λA )B=0λAB=0,

and so on, as a way to recover the well known controllability matrix ( B,AB, A 2 B, ) . It follows that t( N )=N is already a torsion module and that the Kalman system is controllable if and only if N=t( N )=0 as claimed. Moreover, it is well known that a control system is controllable if and only if it is parametrizable, that is M can be embedded into a free differential module. In fact, when n=1 , D=K[ d ] is a principal ideal domain, that is any ideal can be generated by a single element, and it is well known that any torsion-free module over D is indeed free. Accordingly, the kernel of the projection of Dy+Du D 2 onto M is free too and there is no loss of generality by supposing that the control system is made by differentially independent equations. The controllability of an OD control system is thus the purely structural property t( M )=0 independently of the presentation, a fact amounting to the impossibility to find any torsion element, that is any linear combination of the the control variables that could be a solution of an autonomous OD equation for itself.

Example 1.2: With m=3,n=1 and a parameter a=a( x )K , let us consider the formally surjective first order operator with d= d x :

D 1 :( η 1 , η 2 , η 3 )( d η 1 a  η 2 d η 3 = ζ 1 , η 1 d η 2 +d η 3 = ζ 2 )

Multiplying on the left by two test functions ( λ 1 , λ 2 ) and integrating by parts , we obtain:

ad( D 1 ):( λ 1 , λ 2 )( d λ 1 + λ 2 = μ 1 ,a λ 1 +d λ 2 = μ 2 ,d λ 1 d λ 2 = μ 3 )

In order to look for the CC of this operator, we obtain:

λ 2 a λ 1 = μ 1 + μ 2 + μ 3 ( a+ a 2 a ) λ 1 =d( μ 1 + μ 2 + μ 3 )+( a1 ) μ 2 +a μ 3

but we have also ( da ) λ 1 = μ 2 + μ 3 and may easily eliminate λ 1 .

Introducing the notation j q ( μ ) for all the derivatives of μ up to order q , we obtain therefore:

( a+ a 2 a )λ j 1 ( μ )

When the structural controllability condition is satisfied, that is when a is not a solution of the Riccati equation in the bracket, we may obtain a second order CC operator of the form:

ad( D ):( μ 1 , μ 2 , μ 3 ) d 2 μ 1 + d 2 μ 2 + d 2 μ 3 +=ν

Multiplying on the left by a test function ξ and integrating by parts, we obtain the second order injective parametrization, provided that a+ a 2 a0 :

D:ξ( d 2 ξ+= η 1 , d 2 ξ+= η 2 , d 2 ξ= η 3 )

We have the long exact (splitting) sequence and its adjoint (splitting) sequence which is also exact:

0 ξ D 2 η D 1 1 ζ 0 0 ν ad( D ) 2 μ ad( D 1 ) 1 λ 0

At no moment one has to decide about the choice of inputs and outputs and we advise the reader to effect any choice for applying the Kalman test when a is a constant parameter. Of course a=csta( a1 )0 in a coherent way with a classical approach but we have thus been able to extend the controllability test even for variable coefficients, a result still not known because it essentially depends on a systematic use of the adjoint operators.

Example 1.3: (Double pendulum): Many examples can be found in classical ordinary differential control theory because it is known that a linear control system is controllable if and only if it is parametrizable [6]. In our opinion, the best and simplest one is the so-called double pendulum in which a rigid bar is able to move horizontally with reference position x and we attach two pendulums with respective length l 1 and l 2 making the (small) angles θ 1 and θ 2 with the vertical, the corresponding control system does not depend on the mass of each pendulum and the two equations easily follow by projection from the Newton laws:

D 1 η=0 d 2 x+ l 1 d 2 θ 1 +g θ 1 =0, d 2 x+ l 2 d 2 θ 2 +g θ 2 =0

where g is the gravity. A first result, still not acknowledged by the control community, is to prove that this control system is controllable if and only if l 1 l 2 without using a tedious computation through the standard Kalman test but, equivalently, to prove that the corresponding second order operator ad( D 1 ) is injective. Though this is not evident, such a result comes from the fact D is a principal ideal ring when n=1 and thus, if the differential module M 1 is torsion-free, then M 1 is also free and has a basis allowing to split the short exact resolution 0 D 2 D 1 D 3 M 1 0 with M 1 D in this case. When learning control theory, it has also been a surprise to be unable to find examples in which the controllability was explicitly shown not to depend on the choice of inputs and outputs among the system variables, like in such an example as we shall see.

Hence, multiplying on the left the first OD equation by λ 1 , the second by λ 2 , then adding and integrating by parts, we get:

ad( D 1 )λ=μ{ x d 2 λ 1 + d 2 λ 2 = μ 1 θ 1 l 1 d 2 λ 1 +g λ 1 = μ 2 θ 2 l 2 d 2 λ 2 +g λ 2 = μ 3

The main problem is that the operator ad( D 1 ) is not formally integrable because we have:

l 2 λ 1 + l 1 λ 2 = 1 g ( l 2 μ 2 + l 1 μ 3 l 1 l 2 μ 1 )

and is thus injective if and only if l 1 l 2 because, differentiating twice this equation, we also get:

( l 2 / l 1 ) λ 1 +( l 1 / l 2 ) λ 2 j 2 ( μ )

Hence, as the determinant of this 2×2 matrix is exactly l 1 l 2 , if l 1 l 2 , we finally obtain λ j 2 ( μ ) and, after tricky substitutions, a single fourth order CC for μ , namely:

l 1 l 2 d 4 μ 1 g( l 1 + l 2 ) d 2 μ 1 g 2 μ 1 + l 2 d 4 μ 2 +g d 2 μ 2 + l 1 d 4 μ 3 +g d 2 μ 3 =ν

showing that ad( D ) and thus D is indeed a fourth order operator a result not evident at first sight. It follows that we have thus been able to work out the parametrizing operator D of order 4, namely:

Dϕ=η{ l 1 l 2 d 4 ϕg( l 1 + l 2 ) d 2 ϕ g 2 ϕ = x l 2 d 4 ϕ+g d 2 ϕ = θ 1 l 1 d 4 ϕ+g d 2 ϕ = θ 2

This parametrization is injective iff l 1 l 2 because we have successively with g0 :

l 2 d 2 ϕ+gϕ=0 l 1 d 2 ϕ+gϕ=0g( l 1 l 2 )ϕ=0ϕ=0

We have the short exact splitting sequence and its adjoint splitting sequence which is also exact:

0 1 D 4 3 D 1 2 2 0 0 1 ad( D ) 4 3 ad( D 1 ) 2 2 0

We now study the way to split these sequences. As any operator is the adjoint of its own adjoint, we define the lift of the lower sequence as follows:

g 2 ( l 1 l 2 ) λ 1 =g( l 1 l 2 ) μ 2 g ( l 1 ) 2 μ 1 + l 1 l 2 d 2 μ 2 + ( l 1 ) 2 d 2 μ 3 ( l 1 ) 2 l 2 d 2 μ 1

g 2 ( l 1 l 2 ) λ 2 =g( l 1 l 2 ) μ 3 +g ( l 2 ) 2 μ 1 ( l 2 ) 2 d 2 μ 2 l 1 l 2 d 2 μ 3 + l 1 ( l 2 ) 2 d 2 μ 1

obtain the lift of the upper sequence, up to a factor g 2 ( l 1 l 2 ) , namely:

g ( l 1 ) 2 ζ 1 +g ( l 2 ) 2 ζ 2 ( l 1 ) 2 l 2 d 2 ζ 1 + l 1 ( l 2 ) 2 ζ 2 =x

g( l 1 l 2 ) ζ 1 + l 1 l 2 d 2 ζ 1 ( l 2 ) 2 d 2 ζ 2 = θ 1

g( l 1 l 2 ) ζ 2 + ( l 1 ) 2 d 2 ζ 1 l 1 l 2 d 2 ζ 2 = θ 2

We finally consider the case l 1 = l 2 =l . Subtracting the two OD equations, we discover that z= θ 1 θ 2 is an observable quantity that satisfies the autonomous system l d 2 z+gz=0 existing for a single pendulum. It follows that z is a torsion element and the system cannot be controllable. When z=0 θ 1 = θ 2 =θ we let the reader prove that the remaining OD equation d 2 x+l d 2 θ+gθ=0 can be parametrized by l d 2 ξ+gξ=x , d 2 ξ=θ .

At this stage of the reading, we invite the reader to realize this experiment with a few dollars, check how the controllability depends on the lengths and wonder how this example may have anything to do with the Cosserat, Einstein or Maxwell equations !.

Example 1.4 (RLC electrical circuit) As we shall prove below, we do believe that the standard control theory of electrical circuits does not allow at all to study the structure of the various underlying differential modules defined by the corresponding systems (torsion submodules, extension modules, resolutions, ...), in particular if some of the RLC components depend on time.

Let us consider a RLC electrical circuit made up by a battery with voltage u delivering a current y to a parallel subsystem with a branch containing a capacity C with voltage x 1 between its two plates and a resistance R 1 while the other branch, crossed by a current x 2 , is containing a coil L and a resistance R 2 . The corresponding OD equations are easily seen to be:

R 1 Cd x 1 + x 1 =u,Ld x 2 + R 2 x 2 =u, x 1 R 1 x 2 =u R 1 y

Such a system can be set up at once in the standard matrix form x ˙ =Ax+Bu , y=Cx+Eu with input u , state ( x 1 , x 2 ) and output y , but we shall avoid the corresponding Kalman criterion that could not be used if R 1 , R 2 ,L or C should depend on time. The two first order OD equations for ( u, x 1 , x 2 ) are defining a differential module N over the differential field K=( R 1 , R 2 ,L,C ) while the elimination of ( x 1 , x 2 ) is providing the input submodule Du=LN and the output submodule Dy=MN with ( L,M )N . Nothing can be said as long as the prolongation/projection (PP) procedure has not been achieved like in the previous example but it has never been used in control theory, in particular for electrical circuit. The idea is to forget about the state by eliminating it but also to forget about the distinction between the input u and the output y because we already know that controllability is a structural property of the control system for ( u,y ) that we have now to work out.

Differentiating the third zero order OD equation, we get

d x 1 R 1 d x 2 =du R 1 dy . Using the two first order OD equations, we obtain:

x 1 + ( R 1 ) 2 R 2 C L x 2 = ( R 1 ) 2 Cdy R 1 Cdu+( ( R 1 ) 2 C L 1 )u and thus

( L R 1 R 2 C ) x 2 = R 1 CLdy+LyCLdu R 1 Cu with a similar result for x 1 . Hence, in the present situation, we have to distinguish carefully between two cases:

If R 1 R 2 CL , we have ( x 1 , x 2 ) j 1 ( u,y ) and obtain a single second order CC for ( u,y ) :

R 1 CL d 2 y+( L+ R 1 R 2 C )dy+ R 2 yCL d 2 u( R 1 + R 2 )Cduu=0

The system is observable, that is we have indeed the strict equality ( L,M )=N . Surprisingly, multiplying the first OD equation by λ 1 the second by λ 2 , adding and integrating by parts, the kernel of the adjoint system is:

x 1 R 1 Cd λ 1 λ 1 =0, x 2 Ld λ 2 R 2 λ 2 =0,u λ 1 + λ 2 =0 Ld λ 1 R 2 λ 1 =0

and thus λ 1 = λ 2 =0 because the determinant of the system for the couple ( d λ 1 , λ 1 ) is just L R 1 R 2 C0 , a result showing that N is a torsion-free module, that is t( N )=0 . The reader may compare this approach to the Kalman procedure. For example, if R 1 =C=L=1 , R 2 =2 , we get the system d 2 y+3dy+2y d 2 u3duu=0 which is easily seen to be controllable. It is quite difficult to find such examples.

If R 1 R 2 C=L , we have only a single first order CC equation for ( u,y ) , namely:

Ldy+ R 2 y R 2 Cduu=0

Multiplying by a test function λ and integrating by parts, we have to solve the two equations Ldλ+ R 2 λ=0 and R 2 Cdλλ=0 . Hence the system is controllable if and only if the only possible solution is λ=0 , that is when L ( R 2 ) 2 C or R 1 R 2 when C0 and we have the strict inclusion

( L,M )N . Indeed, setting z=y 1 R 1 u , we get R 1 Cdz+z=0 and z is a

torsion element. For example, if R 1 = R 2 =L=C=1 , we get the system dy+yduu=0 which is not controllable because z=yu is a torsion element with dz+z=0 .

Remark 1.5: In the general situation, one can use the differential submodules t( N ) , L and M of N both with the new differential modules L =L+t( N ) and M'=M+t( N ) in order to study all the problems concerning poles and zeros of control systems. As we are only interested by controllability, we have just to study the differential submodules of the torsion-free differential module N/ t( N ) . If we suppose that LM=0 , we have the following commutative diagram of inclusions, in which the upper commutative square is the so-called minimum controllable realization used as a logo for our ERCIM courses (1990 to 1996):

N L M t( N ) L M 0

Studying the differential correspondence between ( x 1 , x 2 ) and ( u,y ) , we have to eliminate ( x 1 , x 2 ) in order to find the resolvent system for ( u,y ) . These results could be extended to time dependent electrical components and open a large domain for future research on electrical circuits. We notice that at no moment we have used the fact that u is called input and y is called output !

This paper is also a kind of Summary Note sketching in a rather self-contained but condensed way the results presented through a series of lectures at the Albert Einstein Institute (AEI, Berlin/Potsdam), October, 23-27, 2017 [9]. The initial motivation for studying the methods used in this paper has been a 1000$ challenge proposed in 1970 by J. Wheeler in the physics department of Princeton University while the author of this paper was a visiting student of D.C. Spencer in the close-by mathematics department, namely:

Is it possible to express the generic solutions of Einstein equations in vacuum by means of the derivatives of a certain number of arbitrary functions, like the potentials for Maxwell equations ?.

After recalling the negative answer we already provided in 1995 [10] [11], the main purpose of this paper is to use again these new techniques of differential double duality in order to revisit the mathematical foundations of the concepts and equations involved in general relativity and gauge theory that are leading to gravitational waves. At the same time, we point out the fact that the above parametrization problem is equivalent to the controllability property of a control system, such a result showing for the first time that it is a structural property, that is a property that does not depend on the choice of inputs and outputs or even on the presentation of the system, that is on a change of all the independent variables used to describe the system, contrary to the commonly accepted point of view of the control community. Many explicit examples are illustrating the paper, ranging from ordinary differential (OD) or partial differential (PD) control theory to mathematical physics, explaining in particular why the mathematical foundations of both gravitation and electromagnetism only depend on the structure of the conformal group of space-time. Accordingly, the foundations of control theory, engineering and mathematical physics must be revisited within this new framework, though striking it may sometimes look like. Of course, it is rather easy to study systems involving OD equations as we saw and we shall need new tools for studying systems of PD equations, though these new methods can also be used for OD equations. An additional difficulty will be met when dealing with operators having variable coefficients.

Example 1.6: While using Kalman test in control theory, it is often useful to transform a second order system d 2 y=0 into a first order system d z 1 z 2 =0 , d z 2 =0 by setting y= z 1 , dy= z 2 , transforming one OD equation for one unknown to two OD equations for two unknown. However, the mathematical community is not aware that, more generally, this has been exactly the procedure followed by Spencer from transforming ANY system of PD equations of order q with n independent variables and m unknowns to a new system of PD equations of order one. One of the best elementary examples to be met in the literature has been provided by F. S. Macaulay in 1916 [12] [13]. With n=2,m=1,q=2 and using the jet notation d 23 y= y 23 , let us consider the second order homogeneous system Py y 22 =0 , Qy y 12 y 11 =0 . Differentiating once, we notice that all the derivatives of order 3 vanish. We obtain therefore four arbitrary parametric jets:

{ z 1 =y, z 2 = y 1 , z 3 = y 2 , z 4 = y 11 }

satisfying the non-homogeneous first order “equivalent system”, called “First Spencer operator”:

d 2 z 1 z 3 =0, d 2 z 2 z 4 =0, d 2 z 3 =0, d 2 z 4 =0, d 1 z 1 z 2 =0, d 1 z 2 z 4 =0, d 1 z 3 z 4 =0, d 1 z 4 =0

In the present situation, we can integrate the system explicitly. We have indeed at once a basis of four solutions, namely { θ τ ( x )|1τ4 }={ 1, x 1 , x 2 , x 1 x 2 + 1 2 ( x 1 ) 2 } and the space of solutions is a vector space V of dimension 4 over the field of constants of K .

Changing slightly the notations along with the other examples, we may consider the second order system R 2 J 2 ( E ) written d 22 ξ= η 1 , d 12 ξ d 11 ξ= η 2 or Dξ=η . Differentiating it once, we obtain the trivially involutive third order system R 3 J 3 ( E ) with corresponding Janet tabular:

{ d 222 ξ = d 2 η 1 d 122 ξ = d 1 η 1 d 112 ξ = d 1 η 1 d 2 η 2 d 111 ξ = d 1 η 1 d 2 η 2 d 1 η 2 d 22 ξ = η 1 d 12 ξ d 11 ξ = η 2 1 2 1 1 1

with the CC d 22 η 2 d 12 η 1 + d 11 η 1 =( PQQP )ξ=0 or D 1 η=0 in the exact differential sequence:

0Θ1 D 2 2 D 1 2 10

which is nevertheless far from being a Janet sequence because D is not involutive. In the present situation, we let the reader check that ad( D ) indeed generates the CC of ad( D 1 ) .

Example 1.7: With two independent variables ( x 1 , x 2 ) and one unknown y , let us consider the following second order system with constant coefficients:

{ Py d 22 y =u Qy d 12 yy =v

where now P and Q are PD operators with coefficients in the subfield k= of constants of the differential field K=k( x 1 , x 2 ) . We obtain at once from a first use of crossed derivatives:

d 2 y= d 1 u d 2 v

and from a second use:

y= d 11 u d 12 vv

and could hope to obtain the 4 th -order generating compatibility conditions (CC) by substitution, that is to say:

{ A d 1122 u d 1222 v d 22 vu =0 B d 1112 u d 11 u d 1122 v =0

with the only generating CC : w d 11 A d 12 BB=0 .

However, in this particular case, there is an unexpected unique second order generating CC:

C d 12 uu d 22 v=0

as we now have indeed PQQP=0 both with A d 12 C+C and B d 11 C , a result leading to C d 22 B d 12 A+A . Accordingly, the systems A=0,B=0 on one side and C=0 on the other side are completely different though they have the same solutions in u,v which can be parametrized injectively by y .

Finally, setting u=0,v=0 , we notice that the preceding homogeneous system can be written in the form Dy=0 and admits the only solution y=0 . More precisely, if a linear system R q J q ( E ) of order q on E is given we may find two integers ( r,s ) such that, prolonging r+s times to obtain R q+r+s and keeping only the equations of order q+r , we obtain a system R q+r ( s ) providing all the informations on the solutions up to any order (prolongation/projection (PP) procedure) [14] [15]. In the present case, we get successively:

0= R 2 ( 4 ) R 2 ( 3 ) R 2 ( 2 ) R 2 ( 1 ) R 2 J 2 ( E )

with strict inclusions and respective dimensions: 0<1<2<3<4<6 .

Example 1.8: Denoting by y i k = d i y k for i=1,2 and k=1,2,3 the formal derivatives of the three differential indeterminates y 1 , y 2 , y 3 , we consider the system of three PD equations for 3 unknowns and 2 independent variables ( x 1 , x 2 ) which is defining a differential module M over the non-commutative ring D=( a )( x 1 , x 2 )[ d 1 , d 2 ] of differential operators with coefficients in ( a )( x 1 , x 2 ) when a is a constant parameter:

{ Φ 1 y 2 3 y 2 2 =0 Φ 2 y 2 2 + y 1 1 a x 2 y 1 =0 Φ 3 y 1 3 y 1 2 =0

No one among ( y 1 , y 2 ) can be given arbitrarily and that there is a unique generating CC, namely:

Ψ d 2 Φ 3 d 1 Φ 1 =0

Also, setting z= y 3 y 2 , we get both z 1 =0, z 2 =0 and z is an autonomous element. Then one can easily prove that any other autonomous element can be expressible by means of a differential operator acting on z which is therefore a generator of the torsion module t( M )M . Accordingly, in the present situation, any autonomous element is a constant multiple of z .

Finally, setting z=0 and thus y 3 = y 2 , we obtain for ( y 1 , y 2 ) , after substitution:

Φ y 2 2 + y 1 1 a x 2 y 1 =0

which is defining an operator D :( y 1 , y 2 ) Φ and a torsion-free module M M/ t( M ) in the short exact sequences:

0D D 2 M 0,0t( M )M M 0

Multiplying the previous operator by a test function λ and integrating by parts, the kernel of the adjoint operator ad( D ):λ( μ 1 , μ 2 ) is defined by:

y 2 d 2 λ= μ 2 , y 1 d 1 λa x 2 λ= μ 1 d 1 μ 2 d 2 μ 1 a x 2 μ 2 =aλ

We have thus two quite different situations:

  a=0 : The adjoint operator is not injective and we are in the situation of the div operator when n=2 that can be parametrized by the curl operator in such a way that M is neither free nor projective but M D with a strict inclusion.

  a0 , say a=1 : The adjoint operator is injective and, using the fact that any operator can be written as the adjoint of an operator, we have obtained a lifting operator such that . We shall prove later on that M is not free but projective, thus torsion-free, because this lift provides an isomorphism D 2 M D and an isomorphism Mt( M ) M which may not exist in general.

Example 1.9: (Elasticity) A first striking result that does not seem to have been even noticed by mechanicians up till now, let us consider the situation of classical elasticity theory where D is the Killing operator for the euclidean metric, namely ΩDξ=( ξ )ω S 2 T * and D 1 the corresponding CC, namely the linearized Riemann curvature with n 2 ( n 2 1 )/ 12 components that can be found in any textbooks of elasticity theory or general relativity. In that case, as it is well known that the Poincaré sequence for the exterior derivative is self-adjoint up to sign (for n=3 the adjoints of grad,curl,div are respectively div,curl,grad ) then the first extension module does not depend on the differential sequence used and therefore vanishes. Accordingly, ad( D ) generates the CC of ad( D 1 ) . Hence, in order to parametrize the Cauchy stress equations, that is ad( D ) , namely:

σ 12 = σ 21 , d 1 σ 11 + d 2 σ 21 =0, d 1 σ 12 + d 2 σ 22 =0

one just needs to compute ad( D 1 ) . For n=2 , we get:

Ω 11 =2 d 1 ξ 1 , Ω 12 = d 1 ξ 2 + d 2 ξ 1 , Ω 22 =2 d 2 ξ 2 d 11 Ω 22 + d 22 Ω 11 2 d 12 Ω 12 =0

ϕ( d 11 Ω 22 + d 22 Ω 11 2 d 12 Ω 12 )=( d 22 ϕ ) Ω 11 2( d 12 ϕ ) Ω 12 +( d 11 ϕ ) Ω 22 + d i ( ) i

and recover the parametrization by means of the Airy function in a rather unexpected way:

σ 11 = d 22 ϕ, σ 12 = σ 21 = d 12 ϕ, σ 22 = d 11 ϕ

Exhibiting a parametrization for n3 thus becomes an exercise of computer algebra, the number of (pseudo)-potentials being the number n 2 ( n 2 1 )/ 12 of components of the Riemann tensor.

We now treat the case of Cosserat equations with zero second members, namely [16] [17]:

d 1 σ 1,1 + d 2 σ 2,1 =0, d 1 σ 1,2 + d 2 σ 2,2 =0, d i μ i,12 + σ 1,2 σ 2,1 =0

For this, instead of using the Janet sequence as before, we now use the Spencer sequence which is isomorphic to the gauge sequence, namely the tensor product of the Poincaré sequence by a lie algebra. However, according to the general theorems of homological algebra, the existence of a parametrization does not depend on the differential sequence used and therefore follows again, like in the previous example, from the fact that the Poincaré sequence is self-adjoint up to the sign. In the present situation, we have with dim( G )= n( n+1 )/2 . We have shown that the Cosserat equations were just ad( D 1 ) , their first order parametrization is thus described by ad( D 2 ) and needs dim( C 2 )= n 2 ( n 2 1 )/4 (pseudo)-potentials. We provide the details when n=2 but we know at once that we must use 3 (pseudo)-potentials only. The case n=3 could be treated similarly and is left as an exercise.

In fact, for constructing the adjoint of D 1 , we have just to integrate by parts the duality summation σ j i ( i ξ j ξ i j )+ μ k ij ( i ξ j k ξ ij k ) while taking into account that ξ ij k =0 . Lowering the indices by means of the constant Euclidean metric, the Spencer operator D 1 is described by the equations:

1 ξ 1 = A 11 , 1 ξ 2 ξ 1,2 = A 12 , 2 ξ 1 ξ 2,1 = A 21 , 2 ξ 2 = A 22 , 1 ξ 1,2 = B 1 , 2 ξ 1,2 = B 2

because R 1 is defined by the equations ξ 1,1 =0, ξ 1,2 + ξ 2,1 =0, ξ 2,2 =0 .

Accordingly the 3 CC describing the Spencer operator D 2 are:

2 A 11 1 A 21 + B 1 =0, 2 A 12 1 A 22 + B 2 =0, 2 B 1 1 B 2 =0

Multiplying these equations respectively by ϕ 1 , ϕ 2 , ϕ 3 , summing and integrating by part, we get ad( D 2 ) and the first order parametrization in the form [16] [17]:

σ 1,1 = 2 ϕ 1 , σ 1,2 = 2 ϕ 2 , σ 2,1 = 1 ϕ 1 , σ 2,2 = 1 ϕ 2 , μ 1,12 = 2 ϕ 3 + ϕ 1 , μ 2,12 = 1 ϕ 3 + ϕ 2

as announced previously. As we are dealing with PD equations with constant coefficients, it is important to notice that such a parametrization could also have been obtained by localization later on. When the stress is symmetric, that is σ 1,2 = σ 2,1 , the Airy parametrization can be recovered if we cancel the couple-stress with ϕ 1 = 2 ϕ 3 , ϕ 2 = 1 ϕ 3 and set ϕ 3 =ϕ .

Changing the presentation will be studied later on as we shall need a lot of additional tools.

Example 1.10: (Electromagnetism) A similar comment can be done for electromagnetism through the exterior derivative as the first set of Maxwell equations can be parametrized by the EM potential 1-form while the second set of Maxwell equations, adjoint of this parametrization, can be parametrized by the EM pseudo-potential. With more details, the beginning of the classical Poincaré sequence for the exterior derivative is:

0 T * d 1 T * d 2 T * d 3 T *

Using standard notations, we denote by A 1 T * = T * the EM potential, by F=( F ij ) 2 T * the EM field and the first set of Maxwel equations, namely dF=0 , is parametrized by dA=F . Denoting by =( ij ) 4 T * 2 T the EM induction, a tensorial density, the second set of Maxwell equations is usually written as i ij = J j and thus ad( d )=J 4 T * T , with the only CC j J j =0 describing the so-called conservation of current. The problem that we faced while teaching EM during twenty years, is that only tensors are used in most textbooks and the above formulas, if they are used by physicists, are not correct at all from a mathematical point of view. When E is any vector bundle over a manifold X of dimension n , the idea is to introduce the adjoint vector bundle ad( E )= n T * E * with E * defined by patching the inverse transition matrices, exactly like T * is obtained from T . Such a formal approach, totally lacking in the literature, allows to describe both the second set of Maxwell equations and the conservation of current in the following dual sequence existing when n=4 :

0ad( 0 T * ) ad( d ) ad( 1 T * ) ad( d ) ad( 2 T * ) ad( d ) ad( 3 T * )

in which ad( d ) is going “backwards”, that is from right to left. For the reader knowing more mathematics, such a procedure may be simplified by using Hodge duality with the volume form dx=d x 1 d x n as a natural way to obtain the dual sequence when n=4 in the form:

0 4 T * d 3 T * d 2 T * d 1 T *

Such a confusing procedure has in fact to do with the so-called side changing functor in differential homological algebra but is far out of the purpose of this paper. Of course, in the actual practice of computer algebra and electromagnetism, the two dual sequences can be written, up to sign, as:

1 d 4 d 6 MI 4 0 1 d 4 MII 6 d 4 MaxwellI MaxwellII

Let us finally simply say that it is a way to transform a left differential module into a right differential module and vice-versa, one of the most difficult concepts that must be used when studying differential extension modules and the reason for which an adjoint operator must always be written “backward” as we saw (See [18] [19] for more details and examples).

We end this Introduction with one of the best academic examples we know in order to understand that working out differential sequences is not an easy task, even on elementary examples.

Example 1.11: (Macaulay) Let us revisit Example 1.6 using more advanced methods. With m=1,n=2,q=2,D=[ d 1 , d 2 ] , let us consider the linear second order system R 2 J 2 ( E ) with dim( E )=1 defined by the two PD equations Pξ d 22 ξ=0 , Qξ d 12 ξ d 11 ξ=0 . We let the reader check easily that g 3 =0 with d ijk ξ=0 and thus dim( R 2 )=4 with parametric jets ( z 1 =ξ, z 2 = ξ 1 , z 3 = ξ 2 , z 4 = ξ 11 ) , a result leading to dim( R 2+r )=4 if we differentiate r times. We recall the dimensions:

q 0 1 2 3 4 5 S q T * 1 2 3 4 5 6 J q ( E ) 1 3 6 10 15 21

both with the commutative and exact diagram allowing to construct inductively the Spencer bundles C r C r ( E ) and the Janet bundles F r for r=0,1,,n with F 0 = J q ( E )/ R q and C 0 = R q J q ( E )= C 1 ( E ) while replacing the system R q J q ( E ) of order q on E by the system R q+1 J 1 ( R q ) of order 1 on R q when q is large enough, that is q=3 in the present example because g 3 =0 . The following diagram allows to start the induction:

0 0 0 R q+1 J 1 ( R q ) C 1 0 0 J q+1 ( E ) J 1 ( J q ( E ) ) C 1 ( E ) 0 0 R q+1 J q+1 ( E ) J 1 ( F 0 ) F 1 0 0 0

In the present situations with q=3 we have the dimensions:

0 0 0 4 12 8 0 0 15 30 15 0 0 4 15 18 7 0 0 0

Inductively, using the Spencer δ -map, we have indeed:

C r = r T * R q /δ ( r1 T * g q+1 )

C r ( E )= r T * J q ( E )/δ ( r1 T * S q+1 T * E )

F r = r T * J q ( E )/ ( r T * R q +δ( r1 T * S q+1 T E ) )

When R q J q ( E ) is involutive, that is formally integrable (FI) with an involutive symbol g q , then these three differential sequences are formally exact on the jet level and, in the Spencer sequence:

0Θ j q q C 0 D 1 1 C 1 D 2 1 D n 1 C n 0

the first order involutive operators D 1 , D 2 ,, D n are induced by the standard Spencer operator d: R q+1 T * R q that can be extended to d: r T * R q+1 r+1 R q [11] [15].

A similar condition is also valid for the Janet sequence:

0ΘE D q F 0 D 1 1 F 1 D 2 1 D n 1 F n 0

which can be thus constructed “as a whole” from the previous extension of the Spencer operator (See [15], p 183 + 185 + 391 for the main diagrams, [6], [11] for other explicit computations on the Macaulay example and application to group theory). However, such a result is still neither known and nor even used today in mathematical physics, particularly in general relativity which is never using the Spencer δ -cohomology in order to define the Riemann or Bianchi operators. In the present example, as the coefficients are constant, the only second order CC is PQQP=0 and the simplest formally exact resolution, quite far from being a

Janet sequence. With a basis of solutions ( θ τ )=( 1, x 1 , x 2 , 1 2 ( x 1 ) 2 + x 1 x 2 ) , we may

introduce the general section of R 3 , namely ξ μ = λ τ ( x ) μ θ τ ( x ) and obtain for the Spencer operator:

( d ξ 3 ) μ,i ( x )= i ξ μ ( x ) ξ μ+ 1 i ( x )=( i λ τ ( x ) ) μ θ τ ( x )

The upper Spencer sequence in the following Fundamental Diagram I is isomorphic to the tensor product of the Poincaré sequence for the exterior derivative by a vector space V of dimension 4 over . Using now the involutive system R 3 instead of R 2 , we get:

0 0 0 0 Θ j 3 4 D 1 1 8 D 2 1 4 0 0 1 j 3 3 10 D 1 1 15 D 2 1 6 0 0 Θ 1 D 3 6 D 1 1 7 D 2 1 2 0 0 0 0

In each sequence, the Euler-Poncaré alternate sum of dimensions is indeed vanishing. Taking the adjoint of each operator and inverting the arrows, we obtain the commutative diagram:

0 0 0 0 4 ad( D 1 ) 1 8 ad( D 2 ) 1 4 0 1 ad( j 3 ) 3 10 ad( D 1 ) 1 15 ad( D 2 ) 1 6 0 0 1 ad( D ) 3 6 ad( D 1 ) 1 7 ad( D 2 ) 1 2 0 0 0 0 0

which is not formally exact because a delicate chase allows to prove that the cohomology H=Z/B at 7 is isomorphic to the kernel of ad( D 2 ) and is thus ex t 2 ( M )0 because n=2 though ex t 1 ( M )=0 . Cutting the last diagram vertically after 7 , we notice that Z is the kernel of the north west arrow. Indeed, starting with a6 killed by the upper north west arrow, we get b15 coming from a unique c7 killed by the lower north west arrow and thus killed by ad( D 1 ) , that is cZ . Such a result is allowing to obtain the following commutative and exact diagram:

0 8 ad( D 2 ) 4 ker( ad( D 2 ) ) 0 0 6 = 6 0 0 H Z B 0 0 0

A snake chase finally provides the desired isomorphism. We also notice that the two central exact sequences of these diagrams both split. Such a situation is one of the rare ones encountered in the study of exact canonical Spencer/Janet sequences. The similar but more delicate study of another example, also provided by Macaulay, can be found for the dimension n=3 [20].

The content of the paper will follow this Introduction. In the long section 2 we shall recall, in the most self-contained and elementary way as possible, the concepts and main results of homological algebra before extending them to the differential framework (see Zentralblatt review Zbl 1079.93001). In section 3 we shall apply them in order to revisit the mathematical foundations of general relativity. In section 4 we shall prove that the structure of the conformal group must also be carefully revisited because, contrary to Riemannian geometry, the corresponding differential sequence will drastically depend on the dimension of the ground manifold. As a byproduct, we also revisit the mathematical foundations of both electromagnetism and gravitation by chasing in the fundamental diagram II, before concluding in section 5.

We now sketch the main result that will be proved and illustrated through this paper, pointing out that just learning about the many tools involved should take more that a full year. Its application to Einstein general relativity and Maxwell electromagnetism will prove that the mathematical foundations of these two apparently well established theories will have to be entirely revisited but for quite different reasons. Roughly, let K be a differential field of characteristic zero, that is K , with derivations ( 1 ,, n ) and D=K[ d 1 ,, d n ] is the eventually non-commutative ring of differential operators with coefficients in K . If the differential module M over D is defined by a linear differential operator D with coefficients in K and we denote by N the differential module defined by the (formal) adjoint operator ad( D ) , we shall prove in a rather self-contained way that cannot be found easily elsewhere. Of course, the specific situation of a principal ideal domain D=K[ d ] with n=1 met in classical control theory has specific properties not held when n2 .

Theorem 1.12: The differential module ex t 1 ( M )=ex t D 1 ( M,D )=t( N ) is a torsion module that does not depend on the finite presentation of M .

We invite the reader to keep constantly in mind the motivating examples presented in the Introduction as these new methods, found by pure mathematicians, have never been applied to OD/PD control theory with variable coefficients or mathematical physics (general relativity and gauge theory), a fact explaining why we have not been able to find other references.

2. Differential Homological Algebra

It becomes clear from the previous motivating examples that there is a need for classifying the properties of systems of OD (classical control theory) or PD (mathematical physics) equations in a way that does not depend on their presentations and this is the purpose of differential homological algebra. The crucial idea will be indeed to obtain such a classification from the families of modules they allow to define over integral domains in the following way (see [18] or Zbl 1079.93001) but a much more advanced “ purity” classification in which torsion-free amounts to 0-pure [21]:

FREE    PROJECTIVE    REFLEXIVE    TORSIONFREE

pointing out the fact that such a classification just disappears when n=1 .

2.1. Module Theory

Before entering the heart of this section, we need a few definitions and results from commutative algebra, in particular for localization. The reader may look at the textbook [22] for most of the proofs as we are using quite standard notations, having in mind the previous examples.

Definition 2.1.1: A ring A is a non-empty set with two associative binary operations respectively called addition and multiplication, respectively sending a,bA to a+bA and abA in such a waythat A becomes an abelian group for the multiplication, so that A has a zero element denoted by 0, every aA has an additive inversedenoted by a and the multiplication is distributive over the addition, that is to say a( b+c )=ab+ac,( a+b )c=ac+bc,a,b,cA .

A ring A is said to be unitary if it has a (unique) element ab=ba,a,bA .

A non-zero element aA is called a zero-divisor if one can find a non-zero bA such that ab=0 and a ring is called an integral domain if it has no zero-divisor.

Definition 2.1.2: A ring K is called a field if every non-zero element aK is a unit, that is one can find an element bK such that ab=1K .

Definition 2.1.3: A module M over a ring A or simply an A - module is a set of elements x,y,z, which is an abelian group for an addition ( x,y )x+y with an action A×MM:( a,x )ax satisfying:

a( x+y )=ax+ay,aA,x,yM

a( bx )=( ab )x,a,bA,xM

( a+b )x=ax+bx,a,bA,xM

1x=x,xM

The set of modules over a ring A will be denoted by mod( A ) . A module over a field is called a vector space.

Definition 2.1.4: A map f:MN between two A -modules is called a homomorphism over A if f( x+y )=f( x )+f( y ),x,yM and f( ax )=af( x ),aA,xM . We successively define:

ker( f )={ xM|f( x )=0 }

im( f )={ yN|xM,f( x )=y }

coker( f )=N/ im( f )

Definition 2.1.5: We say that a chain of modules and homomorphisms is a sequence if the composition of two successive such homomorphisms is zero. A sequence is said to be exact if the kernel of each map is equal to the image of the map preceding it. An injective homomorphism is called a monomorphism, a surjective homomorphism is called an epimorphism and a bijective homomorphism is called an isomorphism. A short exact sequence is an exact sequence made by a monomorphism followed by an epimorphism.

Proposition 2.1.6: If one has a short exact sequence:

0 M f M g M 0

then the following conditions are equivalent:

There exists a monomorphism v: M M such that gv=i d M .

There exists an epimorphism u:M M such that uf=i d M .

There exist isomorphisms φ=( u,g ):M M M and ψ=f+v: M M M that are inverse to each other and provide an isomorphism M M M .

Definition 2.1.7: In the above situation, we say that the short exact sequence splits and u( v ) is called a lift for f( g ) . In particular we have the relation: fu+vg=i d M .

Definition 2.1.8: A left (right) ideal a in a ring A is a submodule of A considered as a left (right) module over itself. When the inclusion aA is strict, we say that a is a proper ideal of A .

Lemma 2.1.9: If a is an ideal in a ring A , the set of elements rad( a )={ aA|n, a n a } is an ideal of A containing a and called the radical of a . An ideal is called perfect or radical if it is equal to its radical.

Definition 2.1.10: For any subset SA , the smallest ideal containing S is called the ideal generated by S . An ideal generated by a single element is called a principal ideal and a ring is called a principal ideal ring if any ideal is principal. The simplest example is that of polynomial rings in one indeterminate over a field. When a and b are two ideals of A , we shall denote by a+b ( ab ) the ideal generated by all the sums a+b (products ab ) with aa,bb .

Definition 2.1.11: An ideal p of a ring A is called a prime ideal if, whenever abp ( aAbp in the non-commutative case) then either ap or bp . The set of proper prime ideals of A is denoted by spec( A ) and called the spectrum of A .

Definition 2.1.12: The annihilator of a module M in A is the ideal an n A ( M ) of A made by all the elements aA such that ax=0,xM .

From now on, all rings considered will be unitary integral domains, that is rings containing 1 and having no zero-divisor. For the sake of clarity, as a few results will also be valid for modules over non-commutative rings, we shall denote by   A M B a module M which is a left module for A with operation ( a,x )ax and a right module for B with operation ( x,b )xb . In the commutative case, lower indices are not needed. If M= M A and N= N A are two left A -modules, the set of A -linear maps f:MN will be denoted by ho m A ( M,N ) or simply hom( M,N ) when there will be no confusion and there is a canonical isomorphism hom( A,M )M:ff( 1 ) with inverse x( aax ) . When A is commutative, hom( M,N ) is again an A -module for the law ( bf )( x )=f( bx ) as we have indeed:

( bf )( ax )=f( bax )=f( abx )=af( bx )=a( bf )( x ).

In the non-commutative case, things are much more complicate and we have:

Lemma 2.1.13: Given   A M B and   A N , then ho m A ( M,N ) becomes a left module over B for the law ( bf )( x )=f( xb ) .

Proof: We just need to check the two relations:

( bf )( ax )=f( axb )=af( xb )=a( bf )( x ),

( b ( b f ) )( x )=( b f )( x b )=f( x b b )=( ( b b )f )( x ).

A similar result can be obtained (exercise) with   A M and   A N B , where ho m A ( M,N ) now becomes a right B -module for the law ( fb )( x )=f( x )b .

Now we recall that a sequence of modules and maps is exact if the kernel of any map is equal to the image of the map preceding it and we have:

Theorem 2.1.14: If M, M , M are A -modules, the sequence:

M f M g M 0

is exact if and only if the sequence:

0hom( M ,N )hom( M,N )hom( M ,N )

is exact for any A -module N .

Proof: Let us consider homomorphisms h:MN , h : M N , h : M N such that h g=h , hf= h . If h=0 , then h g=0 implies h ( x )=0, x M because g is surjective and we can find xM such that x =g( x ) . Then h ( x )= h ( g( x ) )= h g( x )=0 . Now, if h =0 , we have hf=0 and h factors through g because the initial sequence is exact. Hence there exists h : M N such that h= h g and the second sequence is exact.

We let the reader prove the converse as an exercise.

Corollary 2.1.15: The short exact sequence:

0 M M M 0

splits if and only if the short exact sequence:

0hom( M ,N )hom( M,N )hom( M ,N )0

is exact for any module N .

Definition 2.1.16: If M is a module over a ring A , a system of generators of M over A is a family { x i } iI of elements of M such that any element of M can be written x= iI a i x i with only a finite number of nonzero a i .

Definition 2.1.17: An A -module is called noetherian if every submodule of M (and thus M itself) is finitely generated.

One has the following technical lemma:

Lemma 2.1.18: In a short exact sequence of modules, the central module is noetherian if and only if the two other modules are noetherian.

We obtain in particular:

Proposition 2.1.19: If A is a noetherian ring and M is a finitely generated module over A , then M is noetherian.

Proof: Applying the lemma to the short exact sequence 0 A r1 A r A0 where the epimorphism on the right is the projection onto the first factor, we deduce by induction that A r is noetherian. Now, if M is generated by { x 1 ,, x r } , there is an epimorphism A r M:( 1,0,,0 ) x 1 ,,{ 0,,0,1 } x r and M is noetherian because of the lemma.

In the preceding situation, the kernel of the epimorphism A r M is also finitely generated, say by { y 1 ,, y s } and we therefore obtain the exact sequence A s A r M0 that can be extended inductively to the left.

Definition 2.1.20: In this case, we say that M is finitely presented.

We now present the basic elements of the technique of localization in the non-commutative case as it will be needed later on in a few proofs. We start with a basic definition:

Definition 2.1.21: A subset S of a ring A is said to be multiplicatively closed if s,tSstS and 1S . For simplicit, we shal suppose from now that A is an integral domain and consider S=A{ 0 } .

In a general way, whenever A is a non-commutative ring, that is abba when a,bA , we shall set the following definition:

Definition 2.1.22: By a left ring of fractions or left localization of a noncommutative ring A with respect to a multiplicatively closed subset S of A , we mean a ring denoted by S 1 A and a homomorphism θ= θ S :A S 1 A such that:

1) θ( s ) is invertible in S 1 A,sS .

2) Each element of S 1 A or fraction has the form θ ( s ) 1 θ( a ) for some sS,aA .

3) ker( θ )={ aA|sS,sa=0 } .

A right ring of fractions or right localization can be similarly defined.

In actual practice, a fraction will be simply written s 1 a and we have to distinguish carefully s 1 a from a s 1 . We shall meet four problems:

How to compare s 1 a with a s 1 ?

How to decide when we shall say that s 1 a= t 1 b ?

How to multiply s 1 a by t 1 b ?

How to find a common denominator for s 1 a+ t 1 b ?

The following proposition is essential and will be completed by two technical lemmas that will be used for constructing localizations.

The following proposition is essential for constructing localizations:

Proposition 2.1.23: If there exists a left localization of A with respect to S , then we must have:

1) SaAs0,aA,sS .

2) If sS and aA are such that as=0 , then there exists tS such that ta=0 .

Proof: The element θ( a )θ ( s ) 1 in S 1 A must be of the form θ ( t ) 1 θ( b ) for some tS,bA . Accordingly, θ( a )θ ( s ) 1 =θ ( t ) 1 θ( b )θ( t )θ( a )=θ( b )θ( s ) and thus θ( tabs )=0uS,uta=ubs with utS,ubA . Finally, as=0θ( a )θ( s )=0θ( a )=0 because θ( s ) is invertible in S 1 A . Hence tS such that ta=0 .

Definition 2.1.24: A set S satisfying the condition 1) is called a left Ore set.

Lemma 2.1.25: If S is a left Ore set in a noetherian ring, then S also satisfies the condition 2) of the preceding lemma.

Lemma 2.1.26: If S is a left Ore set in a ring A , then AsAtS0,s,tS and two fractions can be brought to the same denominator.

Let K be a differential field with n commuting derivations ( 1 ,, n ) and consider the ring D=K[ d 1 ,, d n ]=K[ d ] of differential operators with coefficients in K with n commuting formal derivatives satisfying d i a=a d i + i a in the operator sense. If P= a μ d μ D=K[ d ] , the highest value of | μ | with a μ 0 is called the order of the operator P and the ring D with multiplication ( P,Q )PQ=PQ is filtred by the order q of the operators. We have the filtration 0K= D 0 D 1 D q D =D . As an algebra, D is generated by K= D 0 and T= D 1 / D 0 with D 1 =KT if we identify an element ξ= ξ i d i T with the vector field ξ= ξ i ( x ) i of differential geometry, but with ξ i K now. It follows that D = D D D is a bimodule over itself, being at the same time a left D -module by the composition PQP and a right D -module by the composition PPQ . We define the adjoint functor ad:D D op :P= a μ d μ ad( P )= ( 1 ) | μ | d μ a μ and we have ad( ad( P ) )=P both with ad( PQ )=ad( Q )ad( P ),P,QD . Such a definition can be extended to any matrix of operators by using the transposed matrix of adjoint operators (see [18] for more details and applications to control theory or mathematical physics).

Proposition 2.1.27: D is an Ore domain and S=D{ 0 } S 1 D=D S 1 .

Proof: For this, if P,QD , let us consider the inhomogeneous system ad( P )y=u , ad( Q )y=v . As the number of derivative of ( u,v ) . is quite larger than the number of derivatives of the single y, there is at least one compatibility condition (CC) for ( u,v ) of the form Uu=Vv leading to the identity UP=VQ and D is an Ore domain. Conversely, if U,VD , we may repeat the same procedure with ad( U ),ad( V ) in order to get ad( P ),ad( Q ) such that ad( P )ad( U )=ad( Q )ad( V ) and thus to get P,QD such that UP=VQ and thus U 1 V=P Q 1 , a result showing the importance of the adjoint (Compare to [1], p. 27).

Accordingly, if y=( y 1 ,, y m ) are differential indeterminates, then D acts on y k by setting d i y k = y i k d μ y k = y μ k with d i y μ k = y μ+ 1 i k and y 0 k = y k . We may therefore use the jet coordinates in a formal way as in the previous section. Therefore, if a system of OD/PD equations is written in the form Φ τ a k τμ y μ k =0 with coefficients aK , we may introduce the free differential module Dy=D y 1 ++D y m D m and consider the differential module of equations I=DΦDy , both with the residual differential module M= Dy/ DΦ or D -module and we may set M= M D if we want to specify the ring of differential operators. We may introduce the formal prolongation with respect to d i by setting d i Φ τ a k τμ y μ+ 1 i k +( i a k τμ ) y μ k in order to induce maps d i :MM: y ¯ μ k y ¯ μ+ 1 i k by residue with respect to I if we use to denote the residue DyM: y k y ¯ k by a bar like in algebraic geometry. However, for simplicity, we shall not write down the bar when the background will indicate clearly if we are in Dy or in M . As a byproduct, the differential modules we shall consider will always be finitely generated ( k=1,,m< ) and finitely presented ( τ=1,,p< ). Equivalently, introducing the matrix of operators D=( a k τμ d μ ) with m columns and p rows, we may introduce the morphism D p D D m :( P τ )( P τ Φ τ ) over D by acting with D on the left of these row vectors while acting with D on the right of these row vectors by composition of operators with im( D )=I . The presentation of M is defined by the exact cokernel sequence D p D D m M0 . We notice that the presentation only depends on K,D and Φ or D , that is to say never refers to the concept of (explicit local or formal) solutions. It follows from its definition that M can be endowed with a quotient filtration obtained from that of D m which is defined by the order of the jet coordinates y q in D q y . We have therefore the inductive limit 0 M 0 M 1 M q M =M with d i M q M q+1 and M=D M q for q0 with prolongations D r M q M q+r ,q,r0 . It is important to notice that it may be sometimes quite difficult to work out I q or M q from a given presentation which is not involutive [15].

We are now in position to construct the ring of fractions S 1 A whenever S is a left Ore set. For this, usingthe preceding lemmas, let us define an equivalence relation on S×A by saying that ( s,a )~( t,b ) if one can find u,vS such that us=vtS and ua=vb . Such a relation is clearly reflexive and symmetric, thus we only need to prove that it is transitive. So let ( s 1 , a 1 )~( s 2 , a 2 ) and ( s 2 , a 2 )~( s 3 , a 3 ) . Then we can find u 1 , u 2 A such that u 1 s 1 = u 2 s 2 S and u 1 a 1 = u 2 a 2 . Also we can find v 2 , v 3 A such that v 2 s 2 = v 3 s 3 S and v 2 a 2 = v 3 a 3 . Now, from the Ore condition, one can find w 1 , w 3 A such that w 1 u 1 s 1 = w 3 v 3 s 3 S and thus w 1 u 2 s 2 = w 3 v 2 s 2 S , that is to say ( w 1 u 2 w 3 v 2 ) s 2 =0 . Hence, unless A is an integral domain, using the second condition of the last proposition, we can find tS such that t( w 1 u 2 w 3 v 2 )=0t w 1 u 2 =t w 3 v 2 . Changing w 1 and w 3 if necessary, we may assume that w 1 u 2 = w 3 v 2 w 1 u 1 a 1 = w 1 u 2 a 2 = w 3 v 2 a 2 = w 3 v 3 a 3 as wished. We finally define S 1 A to be the quotient of S×A by the above equivalence relation with θ:A S 1 A:a 1 1 a .

The sum ( s,a )+( t,b ) will be defined to be ( us=vt,ua+vb ) and the product ( s,a )×( t,b ) will be defined to be ( st,ab ) .

A similar approach can be used in order to define and construct modules of fractions whenever S satifies the two conditions of the last proposition. For this we need a preliminary lemma:

Lemma 2.1.28: If S is a left Ore set in a ring A and M is a left module over A , the set:

t S ( M )={ xM|sS,sx=0 }

is a submodule of M called the S - torsion submodule of M .

Proof: If x,y t S ( M ) , we may find s,tS such that sx=0,ty=0 . Now, we can find u,vA such that us=vtS and we successively get us( x+y )=usx+vty=0x+y t S ( M ) . Also, aA , using the Ore condition for S , we can find bA,tS such that ta=bs and we get tax=bsx=0ax t S ( M ) .

Definition 2.1.29: By a left module of fractions or left localization of M with respect to S , we mean a left module S 1 M over S 1 A both with a homomorphism θ= θ S :M S 1 M:x 1 1 x such that:

1) Each element of S 1 M has the form s 1 θ( x ) for sS,xM .

2) ker( θ S )= t S ( M ) .

In order to construct S 1 M , we shall define an equivalence relation on S×M by saying that ( s,x )~( t,y ) if there exists u,vA such that us=vtS and ux=vy . Checking that this relation is reflexive, symmetric and transitive can be done as before (exercise) and we define S 1 M to be the quotient of S×M by this equivalence relation.

The main property of localization is expressed by the following theorem:

Theorem 2.1.30: If one has an exact sequence:

M f M g M

then one also has the exact sequence:

S 1 M S 1 f S 1 M S 1 g S 1 M

where S 1 f( s 1 x )= s 1 f( x ) .

We now turn to the definition and brief study of tensor products of modules over rings that will not be necessarily commutative unless stated explicitly.

Let M= M A be a right A -module and N= N A be a left A -module. We may introduce the free -module made by finite formal linear combinations of elements of M×N with coefficients in .

Definition 2.1.31: The tensor product of M and N over A is the -module M N A obtained by quotienting the above -module by the submodule generated by the elements of the form:

( x+ x ,y )( x,y )( x ,y ),( x,y+ y )( x,y )( x, y ),( xa,y )( x,ay )

and the image of ( x,y ) will be denoted by xy .

It follows from the definition that we have the relations:

( x+ x )y=xy+ x y,x( y+ y )=xy+x y ,xay=xay

and there is a canonical isomorphism M A A M,A N A N . When A is commutative, we may use left modules only and M A N becomes a left A -module.

Example 2.1.32: If A=,M=/ 2 and N=/ 3 , we have ( / 2 ) ( / 3 )=0 because xy=3( xy )2( xy )=x3y2xy=00=0 .

As a link with localization, we let the reader prove that the multiplication map S 1 A×M S 1 M given by ( s 1 a,x ) s 1 ax induces an isomorphism S 1 A A M S 1 M of modules over S 1 A when S 1 A is considered as a right module over A and M as a left module over A .

When A is a commutative integral domain and S=A{ 0 } , the field K=Q( A )= S 1 A is called the field of fractions of A and we have the short exact sequence:

0AKK/A 0

If now M is a left A -module, we may tensor this sequence by M on the right with AM=M but we do not get in general an exact sequence. The defect of exactness on the left is nothing else but the torsion submodule t( M )={ mM|0aA,am=0 }M and we have the long exact sequence:

0t( M )MK A MK/A M A 0

as we may describe the central map as follows:

m1m= a a m= 1 a am,0aA

Such a result, based on the localization technique, allows to understand why controllability has to do with the so-called “simplification” of the transfer matrix. In particular, a module M is said to be a torsion module if t( M )=M and a torsion-free module if t( M )=0 .

Definition 2.1.33: A module in mod( A ) is called a free module if it has a basis, that is a system of generators linearly independent over A . When a module F is free, the number of generators in a basis, and thus in any basis (exercise), is called the rank of F over A and is denoted by ran k A ( F ) . In particular, if F is free of finite rank r , then F A r .

More generally, if M is any module over a ring A and F is a maximum free submodule of M , then M/F =T is a torsion module. Indeed, if xM,xF , then one can find aA such that axF because, otherwise, F{ F,x } should be free submodules of M with a strict inclusion. In that case, the rank of M is by definition the rank of F over A and one has equivalently :

Lemma 2.1.34: r k A ( M )=di m K ( K A M ) .

Proof: Taking the tensor product by K over A of the short exact sequence 0FMT0 , we get an isomorphism K A FK A M because K A T=0 (exercise) and the lemma follows from the definition of the rank.

We now provide two proofs of the additivity property of the rank, the second one being also valid for non-commutative rings.

Proposition 2.1.35: If 0 M f M g M 0 is a short exact sequence of modules over a ring A , then we have r k A ( M )=r k A ( M )+r k A ( M ) .

Proof 1: Using localization with respect to the multiplicatively closed subset S=A{ 0 } , this proposition is just a straight consequence of the definition of rank and the fact that localization preserves exactness.

Proof 2: Let us consider the following diagram with exact left/right columns and central row:

0 0 0 0 F F F F 0 i i i 0 M f M g M 0 p p p 0 T T T 0 0 0 0

where F ( F ) is a maximum free submodule of M ( M ) and T = M / F ( T = M / F ) is a torsion module. Pulling back by g the image under i of a basis of F , we may obtain by linearity a map σ: F M and we define i=f i π +σ π where π : F F F and π : F F F are the canonical projections on each factor of the direct sum. We have i| F =f i and gi=gσ π = i π . Hence, the diagram is commutative and thus exact with r k A ( F F )=r k A ( F )+r k A ( F ) trivially. Finally, if T and T are torsion modules, it is easy to check that T is a torsion module too and F F is thus a maximum free submodule of M .

Definition 2.1.36: If f:MN is any morphism, the rank of f will be defined to be r k A ( f )=r k A ( im( f ) ) .

We provide a few additional properties of the rank that will be used in the sequel. For this we shall set M * =ho m A ( M,A ) and, for any moprphism f:MN we shall denote by f * : N * M * the corresponding morphism which is such that f * ( h )=hf , hho m A ( N,A ) .

Proposition 2.1.37: When A is a commutative integral domain and M is a finitely presented module over A , then r k A ( M )=r k A ( M * ) .

Proof: Applying ho m A ( ,A ) to the short exact sequence in the proof of the preceding lemma while taking into account T * =0 , we get a monomorphism 0 M * F * and obtain therefore r k A ( M * )r k A ( F * ) . However, as F A r with r< because M is finitely generated, we get F * A r too because A * A . It follows that r k A ( M * )r k A ( F * )=r k A ( F )=r k A ( M ) and thus r k A ( M * )r k A ( M ) .

Now, if F 1 d F 0 M0 is a finite presentation of M , applying ho m A ( ,A ) to this presentation, we get the ker/coker exact sequence:

0N F 1 * d * F 0 * M * 0

Applying ho m A ( ,A ) to this sequence while taking into account the two useful isomorphisms F 0 ** F 0 , F 1 ** F 1 , we get the ker/coker exact sequence:

0 N * F 1 d F 0 M0

Counting the ranks, we obtain:

r k A ( N )r k A ( M * )=r k A ( F 1 * )r k A ( F 0 * ) =r k A ( F 1 )r k A ( F 0 ) =r k A ( N * )r k A ( M )

and thus:

( r k A ( M )r k A ( M * ) )+( r k A ( N )r k A ( N * ) )=0

As both two numbers in this sum are non-negative, they must be zero and we finally get the very important formulas r k A ( M )=r k A ( M * ),r k A ( N )=r k A ( N * ) .

Corollary 2.1.38: Under the condition of the proposition, we have r k A ( f )=r k A ( f * ) .

Proof: Introducing the ker/coker exact sequence:

0KM f NQ0

we have: r k A ( f )+r k A ( Q )=r k A ( N ) . Applying ho m A ( ,A ) and taking into account Theorem 2.A.14, we have the exact sequence:

0 Q * N * f * M *

and thus : r k A ( f * )+r k A ( Q * )=r k A ( N * ) . Using the preceding proposition, we get r k A ( Q )=r k A ( Q * ) and r k A ( N )=r k A ( N * ) , that is to say r k A ( f )=r k A ( f * ) .

2.2. Homological Algebra

Having in mind the previous section, we now need a few definittions and results from homological algebra [18] [22]. In all that follows, A,B,C, are modules over a ring A or vector spaces over a field k and the linear maps are making the diagrams commutative.

We start recalling the well known Cramer’s rule for linear systems through the exactness of the ker/coker sequence for modules. We introduce the notations rk=rank , nb=number , dim=dimension , ker=kernel , im=image , coker=cokernel . When Φ:AB is a linear map (homomorphism), we introduce the so-called ker/coker exact sequence:

0ker( Φ )A Φ Bcoker( Φ )0

where coker( Φ )=B/ im( Φ ) .

In the case of vector spaces over a field k , we successively have rk( Φ )=dim( im( Φ ) ) , dim( ker( Φ ) )=dim( A )rk( Φ ) , dim( coker( Φ ) )=dim( B )rk( Φ )=nb of compatibility conditions, and obtain by substraction:

dim( ker( Φ ) )dim( A )+dim( B )dim( coker( Φ ) )=0

In the case of modules, using localization, we may replace the dimension by the rank and obtain the same relations because of the additive property of the rank. The following theorem is essential:

Snake Theorem 2.2.1: When one has the following commutative diagram resulting from the the two central vertical short exact sequences by exhibiting the three corresponding horizontal ker/coker exact sequences:

0 0 0 0 K A A Q 0 Φ Φ 0 L B B R 0 Ψ Ψ 0 M C C S 0 0 0 0

then there exists a connecting map MQ both with a long exact sequence:

0KLMQRS0

We may now introduce cohomology theory through the following definition:

Definition 2.2.2: If one has a sequence A Φ B Ψ C , then one may introduce

with coboundary=im( Φ )ker( Ψ )=cocycleB and define the cohomology at B to be the quotient cocycle/coboundary.

Theorem 2.2.3: The following commutative diagram where the two central vertical sequences are long exact sequences and the horizontal lines are ker/coker exact sequences:

0 0 0 0 K A A Q 0 Φ Φ 0 L B B R 0 Ψ Ψ 0 M C C S 0 Ω Ω 0 N D D T 0 0 0 0

induces an isomorphism between the cohomology at M in the left vertical column and the kernel of the morphism QR in the right vertical column.

We now introduce the extension functor in an elementary manner, using the standard notation ho m A ( M,A )= M * . First of all, by a free resolution of an A -module M , we understand a long exact sequence:

d 2 F 1 d 1 F 0 M0

where F 0 , F 1 , are free modules, that is to say modules isomorphic to powers of A and M=coker( d 1 )= F 0 / im( d 1 ) . We may take out M and obtain the deleted sequence:

d 2 F 1 d 1 F 0 0

which is of course no longer exact. If N is any other A -module, we may apply the functor ho m A ( ,N ) and obtain the sequence:

d 2 * ho m A ( F 1 ,N ) d 1 * ho m A ( F 0 ,N )0

Definition 2.2.4: One may define:

ex t A 0 ( M,N )=ker( d 1 * )=ho m A ( M,N ),ex t A i ( M,N )= ker( d i+1 * )/ im( d i * ) ,i1

One can prove that the extension modules do not depend on the resolution of M chosen and have the following two main properties, the first of which only is classical [18] [22].

Proposition 2.2.5: If 0 M M M 0 is a short exact sequence of A -modules, then we have the following connecting long exact sequence:

0ho m A ( M ,N )ho m A ( M,N )ho m A ( M ,N )ex t A 1 ( M ,N )

of extension modules.

We provide two different proofs of the following proposition:

Proposition 2.2.6: ex t i ( M )=ex t A i ( M,A ) is a torsion module, i1 .

Proof 1: Let F be a maximal free submodule of M . From the short exact sequence:

0FMM/F 0

where M/F is a torsion module, we obtain the long exact sequence:

ex t i1 ( F )ex t i ( M/F )ex t i ( M )ex t i ( F )

As ex t i ( F )=0,i1 from the definitions, we get ex t i ( M )ex t i ( M/F ),i2 . Now it is known that the tensor by the field K of any exact sequence is again an exact sequence. Accordingly, we have from the definition:

K A ex t i ( M/F ,A )ex t A i ( M/F ,K )ex t K i ( K A M/F ,K )=0,i1

We finally obtain from the above sequence K A ex t i ( M )=0ex t i ( M ) torsion, i1 .

Proof 2: Having in mind that B i =im( d i * ) and Z i =ker( d i+1 * ) , we obtain rk( B i )=rk( d i * )=rk( d i ) and rk( Z i )=rk( F i * )rk( d i+1 * )=rk( F i )rk( d i+1 ) . However, we started from a resolution, that is an exact sequence in which rk( d i )+rk( d i+1 )=rk( F i ) . It follows that rk( B i )=rk( Z i ) and thus rk( H i )=rk( Z i )rk( B i )=0 , that is to say ex t i ( M ) is a torsion module for i1 , Mmod( A ) .

As we have seen in the Motivating Examples of the Introduction, the same module may have many very different presentations. In particular, we have the Schanuel lemma [18] [22]:

Lemma 2.2.7: If F 1 d 1 F 0 M0 and F 1 d 1 F 0 M0 are two presentations of M , there exists a presentation F 1 d 1 F 0 M0 of M projecting onto the preceding ones.

Definition 2.2.8: An A -module P is projective if there exists a free module F and another (thus projective) module Q such that PQF . Any free module is projective.

Proposition 2.2.9: The short exact sequence:

0 M f M g M 0

splits whenever M is projective.

Proposition 2.2.10: When P is a projective module and N is any module, we have:

ex t A i ( P,N )=0,i1

Proposition 2.2.11: When P is a projective module, applying ho m A ( P, ) to any short exact sequence gives a short exact sequence.

2.3. Differential Duality

The main but highly not evident trick will be to introduce the adjoint operator D ˜ =ad( D ) by the formula of integration by part:

λ,Dξ = D ˜ λ,ξ +div( )

where λ is a test row vector and denotes the usual contraction. The adjoint can also be defined formally, as in computer algebra packages, by setting:

ad( a )=a,aK,ad( d i )= d i ,ad( PQ )=ad( Q )ad( P ),P,QD

Another way is to define the adjoint of an operator directly on D by setting:

P= 0| μ |p a μ d μ ad( P )= 0| μ |p ( 1 ) | μ | d μ a μ

for any PD with ord( P )=p and to extend such a definition by linearity.

We shall denote by N the differential module defined from ad( D ) exactly like M was defined from D and we have the following fundamental theorem which is not easily accessible to intuition [3] [8]:

Theorem 2.3.1: There is a long exact sequence:

0ex t 1 ( N )M ϵ M ** ex t 2 ( N )0

and the two following statements are equivalent:

The corresponding operator is simply (doubly) parametrizable.

The corresponding module is torsion-free (reflexive).

Proof: Let us start with a free presentation of M :

F 1 d 1 F 0 M0

By definition, we have M=coker( d 1 )N=coker( d 1 * ) and we may exhibit the following free resolution of N where M * =ker( d 1 * )=im( d 0 * )coker( d 1 * ) :

0N F 1 * d 1 * F 0 * d 0 * F 1 * d 1 * F 2 * M * = M * 0 0

The deleted sequence is:

0 F 1 * d 1 * F 0 * d 0 * F 1 * d 1 * F 2 *

Applying ho m A ( ,A ) and using the canonical isomorphism F ** F for any free module F , we get the sequence:

0 F 1 d 1 F 0 d 0 F 1 d 1 F 2 M ϵ M ** 0 0

in which ϵ:M M ** is defined by ( ϵ( m ) )( f )=f( m ),fho m A ( M,A ) .

Denoting as usual a coboundary space by B , a cocycle space by Z and the cohomology by H=Z/B , we get the commutative and exact diagram:

0 B 0 F 0 M 0 ϵ 0 Z 0 F 0 M **

An easy snake chase provides at once H 0 = Z 0 / B 0 =ex t 1 ( N )ker( ϵ ) and it follows that ker( ϵ )M is a torsion module, that is ker( ϵ )t( M ) .

Now, if mt( M ) , then we can find 0aA such that am=0 . Hence, fho m A ( M,A ) , we have a( f( m ) )=f( am )=f( 0 )=0 and thus f( m )=0 because A is an integral domain. We obtain therefore t( M )ker( ϵ )M and thus t( M )=ker( ϵ ) .

Finally, as B 1 =im( ϵ ) and Z 1 M ** , we finally obtain:

H 1 = Z 1 / B 1 =ex t 2 ( N )coker( ϵ )

Accordingly, a torsion-free (reflexive) module is described by an operator that admits a single (double) step parametrization.

As ad( ad( D ) )=D , it is important to notice that one can exchange M and N in any case.

The same proof also provides an effective test for applications by using D and ad instead of A and in the differential framework. In particular, a control system is controllable if it does not admit any “ autonomous element”, that is to say any finite linear combination of the control variables and their derivatives that satisfies, for itself, at least one OD or PD equation. More precisely, starting with the control system described by an operator D 1 , one MUST construct D ˜ 1 and then D such that D ˜ generates all the compatibility conditions of D ˜ 1 . Finally, M is torsion-free if and only if D 1 generates all the compatibility conditions of D . Though striking it could be, this is the true generalization of the standard Kalman test as we already claimed in the Introduction.

Corollary 2.3.2: The constructive test in order to know if an operator D 1 can be parametrized by an operator D has five successive steps along with the following diagram in operator language:

ζ 5 D 1 4 ξ D η D 1 ζ 1 3 ν ad( D ) μ ad( D 1 ) λ 2

D 1 parametrizedbyD D 1 = D 1 ex t 1 ( N )=0ϵinjectivet( M )=0

Any new CC brought by D 1 is a torsion element of the differential module defined by D 1 .

Proof: We have used the fact that ad( ad( D ) )=D and the parametrization is existing if and only if we may have D 1 = D 1 whenever D 1 generates the CC of D as ad( D )ad( D 1 )=0 D 1 D=0 , that is D 1 is surely among the CC of D but other CC may also exist. In addition, denoting by M 1 the differentia module determined by D 1 and using the fact that r k D ( D 1 )=r k D ( D 1 )=pr k D ( D ) because r k D ( ad( D ) )=pr k D ( ad( D 1 ) ) , then any new CC provides an element of t( M 1 ) .

Corollary 2.3.3: The constructive test in order to know if an operator D 1 can be parametrized by an operator D which can be itself parametrized by an operator D 1 has 5 steps which are drawn in the following diagram where ad( D ) generates the CC of ad( D 1 ) and D 1 generates the CC of D=ad( ad( D ) ) while ad( D 1 ) generates the CC of ad( D ) and D generates the CC of D 1 :

η ζ 5 D D 1 4 ϕ D 1 ξ D η D 1 ζ 1 3 θ ad( D 1 ) ν ad( D ) μ ad( D 1 ) λ 2

D 1 parametrizedbyD D 1 = D 1 ex t 1 ( N )=0ϵinjectivet( M )=0

Dparametrizedby D 1 D= D ex t 2 ( N )=0ϵsurjective

Corollary 2.3.4: In the differential module framework, if F 1 D 1 F 0 p M0 is a finite free presentation of M=coker( D 1 ) with t( M )=0 , then we may obtain an exact sequence F 1 D 1 F 0 D E of free differential modules where D is the parametrizing operator. However, there may exist other parametrizations F 1 D 1 F 0 D E called minimal parametrizations such that coker( D ) is a torsion module and we have thus r k D ( M )=r k D ( E ) .

Example 2.3.5 : When n3 , the existence of the Poincaré differential sequence:

0Θ 0 T * d 1 T * d d n1 T * d n T * 0

for the exterior derivative “ d ”, proves that the differential module defined by the last operator is surely reflexive. However, when n=3 , the operators involved, namely ( grad,curl,div ) , are such that the div may be parametrized by an operator defining a torsion module as follows by considering the involutive system:

{ d 3 y 3 = z 1 d 3 y 2 = z 2 d 2 y 1 d 1 y 2 = z 3 1 2 3 1 2 3 1 2 d 1 z 1 + d 2 z 2 + d 3 z 3 =0

Now, in order to have a full picture of the correspondence existing between differential modules and differential operators, it just remains to explain why and how we can pass from left to right modules and conversely. By this way, we shall be able to take into account the behaviour of the adjoint of an operator under changes of coordinates. We start with a technical lemma [4] [18]:

Lemma 2.3.6: If faut( X ) is a local diffeomorphism of X , we may set y=f( x )x= f 1 ( y )=g( y ) and introduce the jacobian Δ( x )=det( i f k ( x ) )0 . Then, we have the identity:

y k ( 1 Δ( g( y ) ) i f k ( g( y ) ) )0.

Accordingly, we notice that, if D:EF is an operator, the way to obtain the adjoint through an integration by parts proves that the test function is indeed a section of the adjoint bundle F ˜ = F * Λ n T * and that we get an operator ad( D ): F ˜ E ˜ . This is in particular the reason why, in elasticity, the deformation is a covariant tensor but the stress is a contravariant tensor density and, in electromagnetism, the EM field is a covariant tensor (in fact a 2-form) but the induction is a contravariant tensor density.

Also, if we define the adjoint formally, we get, in the operator sense:

ad( 1 Δ i f k y k )= y k ( 1 Δ i f k )= 1 Δ i f k y k = 1 Δ x i

and obtain therefore:

x i = i f k ( x ) y k ad( x i )= x i =Δad( 1 Δ i f k ( x ) y k )

a result showing that the adjoint of the gradient operator d: Λ 0 T * Λ 1 T * is minus the exterior derivative d: Λ n1 T * Λ n T * .

If A is a differential ring and D=A[ d ] as usual, we may introduce the ideal I={ PD|P( 1 )=0 } and obtain AD/I both with the direct sum decomposition DAI . In fact, denoting by D q the submodule over A of operators of order q , A can be identified with the subring D 0 D of zero order operators and we may consider any differential module over D as a module over A , just “forgetting” about its differential structure. Caring about the notation, we shall set T= D 1 / D 0 ={ ξ= a i d i | a i A } with ξ( a )= ξ i i a,aA , so that D can be generated by A and T .

The module counterpart is more tricky and is based on the following theorem [23]:

Theorem 2.3.7: If M and N are right D -modules, then ho m A ( M,N ) becomes a left D -module.

Proof: We just need to define the action of ξT by the formula:

( ξf )( m )=f( mξ )f( m )ξ,mM

Indeed, setting ( af )( m )=f( m )a=f( ma ) and introducing the bracket ( ξ,η )[ ξ,η ] of vector fields, we let the reader check that a( bf )=( ab )f,a,bA and that we have the formulas:

ξ( af )=( ξ( a )+aξ )f,( ξηηξ )f=[ ξ,η ]f,aA,ξ,ηT

in the operator sense.

Finally, if M is a left D -module, according to the comment following lemma 3.1.13, then M * =ho m D ( M,D ) is a right D -module and thus N= N r is a right D -module. However, we have the following technical proposition:

Proposition 2.3.8: Λ n T * has a natural right module structure over D .

Proof: If α=ad x 1 d x n T * is a volume form with coefficient aA , we may set α.P=ad( P )( a )d x 1 d x n . As D is generated by A and T , we just need to check that the above formula has an intrinsic meaning for any ξT . In that case, we check at once:

α.ξ= i ( a ξ i )d x 1 d x n =( ξ )α

by introducing the Lie derivative of α with respect to ξ , along the intrinsic formula ( ξ )=i( ξ )d+di( ξ ) where i( ) is the interior multiplication and d is the exterior derivative of exterior forms. According to well known properties of the Lie derivative, we get :

α.( aξ )=( α.ξ ).aα.ξ( a ), α.( ξηηξ )=[ ( ξ ),( η ) ]α=( [ ξ,η ] )α=α.[ ξ,η ].

According to the preceding theorem and proposition, the left differential module corresponding to ad( D ) is not N r but rather N l =ho m A ( Λ n T * , N r ) . When D is a commutative ring, this side changing procedure is no longer needed.

Of course, keeping the same module M but changing its presentation or even using an isomorphic module M (2 OD equations of order 2 or 4 OD equations of order 1 as in the case of the double pendulum), then N may change to N . The following result, totally unaccessible to intuition, justifies “a posteriori” the use of the extension functor by proving that the above results are unchanged and are thus “intrinsic[22]:

Theorem 2.3.9: N and N are projectively equivalent, that is to say one can find projective modules P and P such that NP N P .

Proof: According to Schanuel lemma, we can always suppose, with no loss of generality, that the resolution of M projects onto the resolution of M . The kernel sequence is a splitting sequence made up with projective modules because the kernel of the projection of F i onto F i is a projective module P i for i=0,1 . Such a property still holds when applying duality. Hence, if C is the kernel of the epimorphism from P 1 to P 0 induced by d 1 , then C is a projective module and the top short exact sequence splits in the following commutative and exact diagram:

0 0 0 0 C P 1 P 0 0 0 K F 1 d 1 F 0 M 0 0 K F 1 d 1 F 0 M 0 0 0 0 0

Applying ho m A ( ,A ) to this diagram while taking into account Corollary 3.1.15, we get the following commutative and exact diagram:

0 0 0 0 C P 1 P 0 0 0 N F 1 d 1 F 0 M 0 0 N F 1 d 1 F 0 M 0 0 0 0 0

In this diagram C * is also a projective module, the upper and left short exact sequences split and we obtain N N C * .

Accordingly, using the properties of the extension functor, we get:

Corollary 2.3.10: ex t i ( N )ex t i ( N ),i1 .

Remark 2.3.11: When A is a principal ideal ring, it is well known (See (Pommaret, 2001, Rotman, 1979) for more details) that any torsion-free module over A is free and thus projective. Accordingly, the kernel of the projection of F 0 onto M is free and we can always suppose, with no loss of generality, that d 1 and d 1 are monomorphisms [8]. In that case, there is an isomorphism P 0 P 1 in the proof of the preceding theorem and C=0 C * =0 , that is to say N N . This is the very specific situation only considered by OD control theory where the OD equations defining the control systems are always supposed to be differentially independent (linearly independent over D ).

Example 2.3.12: Revisiting the introductory example 1.6, we discover that, the only solution of the given system being y=0 , the differential modules defined by the systems ( A=0,B=0 ) or ( C=0 ) are isomorphic to M=Dy and we have the following commutative and exact diagram of operators:

0 0 0 0 y D 2 ( u,v ) D 1 2 C 0 2 0 y D 2 ( u,v ) D 1 4 ( A,B ) w 0 2 0 0 w = w 0 0 0

Translating this result in the language of differential modules, we obtain the commutative and exact diagram showing that CD :

0 0 0 C = D 0 0 0 D D 2 d 1 D 2 D 0 0 D d 1 D 2 D 0 0 0 0

Applying ho m D ( ,D ) we obtain the commutative and exact diagram:

0 0 0 0 D = D 0 0 0 N D 2 d 1 D 2 D 0 0 N D d 1 D 2 D 0 0 0 0 0

We obtain N N D and ex t 1 ( N )=ex t 1 ( N )=0 because ad( D 1 ) is an injective operator with N =0 , exactly like D is an injective operator, and the bottom horizontal sequence splits.

We are now ready for exhibiting the final desired link with operator theory.

Theorem 2.3.13: One has ad( D ¯ )= ad( D ) ¯ for any change x ¯ =φ( x ) of independent variables.

Proof: As the proof is rather technical, we shall divide it into three steps:

Step 1: We start providing the tricky computation for a change x ¯ j = φ j ( x ) on any ξ= ξ i ( x ) i T . Dealing with operators and no longer with vector fields, we

may set ξ= ξ i d i D , writing x ¯ j = i φ j ξ i = x ¯ j x i ξ i in order to keep the duality

existing between x and x ¯ . Using crucially Lemma ... with now Δ=det( i φ j ) , we obtain successively in the framework of operators:

ξ= ξ i d i Dad( ξ )= d i ξ i =ξ i ξ i D

ad( ξ ¯ )= d d x ¯ j ξ ¯ j = d d x ¯ j x ¯ j x i ξ i =ξ i ξ i ξ i x ¯ j ( ( 1 Δ x ¯ j x i )Δ ) =ξ i ξ i 1 Δ ξ i i Δ

and thus Δad( ξ ¯ )=ΔξΔ i ( Δξ )= d i ξ i Δ=ad( ξ )Δ , that is

ad( ξ ¯ )= ad( ξ ) ¯ .

Step 2: As any operator PD can be written as P= ξ 1 ξ r with ξ 1 ,, ξ r D , we obtain from the first step:

Δad( P ¯ )=Δad( ξ ¯ 1 ξ ¯ r )=Δad( ξ ¯ r )ad( ξ ¯ 1 ) =ad( ξ r )Δad( ξ ¯ r1 )ad( ξ ¯ 1 )=ad( ξ r )ad( ξ 1 )Δ

and thus the formula Δad( P ¯ )=ad( P )Δ , that is ad( P ¯ )= ad( P ) ¯ .

Step 3: With D=ξη and ad( D ):μν , using an integration by parts with contraction <> , we get:

μ,Dξ ad( D )μ,ξ = x i ( ) i

As any contraction is a n -form, we obtain in the new coordinate system:

μ ¯ , D ¯ ξ ¯ = 1 Δ μ,Dξ , ad( D ) ¯ μ ¯ , ξ ¯ = 1 Δ ad( D )μ,ξ

and thus:

μ ¯ , D ¯ ξ ¯ ad( D ) ¯ μ ¯ , ξ ¯ = 1 Δ x i ( ) i = 1 Δ x ¯ j x i x ¯ j ( ) i = x ¯ j ( 1 Δ x ¯ j x i ( ) i )

and thus ad( D ¯ )= ad( D ) ¯ as the adjoint of an operator is uniquely defined by such an identity.

Corollary 2.3.14: To any linear differential operator E D q F of order q we may associate another linear differential operator n T * E * ad( D ) q n T * F * of

order q , in such a way that ad( ad( D ) )=D but it is important to notice that its arrow is now going backwards, that is from right to left. We shall use to set ad( E )= n T * E * in order to simplify the notations for applications while keeping the same dimension.

Important Remark 2.3.15: In actual practice and in the operator framework, we may consider an operator matrix acting on the left of column vectors (sections of vector bundles).

Similarly, in the framework of left D -modules, we may use now row vectors and write:

D A F * D D A E * p M0

with D acting now by composition on the right of row vectors while D is acting on the left by usual composition of operators. We shall set D( E )=D A E * with E * =ho m A ( E,A ) and obtain therefore ho m D ( D( E ),A )=E . Applying ho m D ( ,D ) and using right D -modules or using the side changing functor ho m A ( n T * , ) and using left D -modules, we get:.

D( F )D( E )M0

In the dual situation, we shall obtain:

0ND A n T A F ad( D ) D A n T A E

in order to keep on going with left differential modules. Such a difficulty is explaining why adjoint operators have never been used in mathematical physics up to our knowledge (see « ideXlab» on the Net!).

We point out another difficulty existing because, in general, ad( D ) is far from being involutive or even formally integrable whenever D is involutive. This is particularly true even for OD systems like the Kalman systems or the double pendulum as we saw. For this reason, we shall rather suppose that the coefficients of the operators or systems are in a differential field K rather than in a differential ring A . In a word, one has to get used to a new language.

3. General Relativity

From standard results in continuum mechanics and the preceding formulas, we have [4]:

Proposition 3.1: The Cauchy operator is the adjoint of the Killing operator.

Proof: Let X be a manifold of dimension n with local coordinates ( x 1 ,, x n ) , tangent bundle T and cotangent bundle T * . If ω S 2 T * is a metric with det( ω )0 , we my introduce the standard Lie derivative in order to define the first order Killing operator:

D:ξTΩ=ξ ( ξ )ω=Ω( Ω ij = ω rj ( x ) i ξ r + ω ir ( x ) j ξ r + ξ r r ω ij ( x ) ) S 2 T *

Here start the problems because, in our opinion at least, a systematic use of the adjoint operator has never been used in mathematical physics and even in continuum mechanics apart through a variational procedure. As we have seen, the purely intrinsic definition of the adjoint can only be done in the theory of differential modules by means of the so-called side changing functor. From a purely differential geometric point of view, the idea is to associate to any vector bundle E over X a new vector bundle ad( E )= n T * E * where E * is obtained from E by patching local coordinates while inverting the transition matrices, exactly like T * is obtained from T . It follows that the stress σ=( σ ij )ad( S 2 T * )= n T * S 2 T is not a tensor but a tensor density, that is transforms like a tensor up to a certain power of the Jacobian matrix. When n=4 , the fact that such an object is called stress-energy tensor does not change anything as it cannot be related to the Einstein tensor which is a true tensor indeed. In any case, we may define as usual:

ad( D ): n T * S 2 T n T * T:σ=( σ ij )φ=( φ k )

Multiplying Ω ij by σ ij and integrating by parts, the factor of 2 ω kr ξ r is easly seen to be:

i σ ik = i σ ik + γ ij k σ ij = φ k

with well known Christoffel symbols γ ij k = 1 2 ω kr ( i ω rj + j ω ir r ω ij ) .

However, if the stress should be a tensor, we should get for the covariant derivative:

r σ ij = r σ ij + γ rs i σ sj + γ rs j σ is i σ ik = i σ ik + γ ri r σ ik + γ ij k σ ij

The difficulty is to prove that we do not have a contradiction because σ is a tensor density.

If we have an invertible transformation like in Lemma 2.3.6, we have successively:

τ kl ( f( x ) )= 1 Δ i f k ( x ) j f l ( x ) σ ij ( x )

τ kl y k = 1 Δ i f k y k ( j f l ) σ ij + 1 Δ i f k j f l y k σ ij = 1 Δ ( ij f l ) σ ij + 1 Δ j f l i σ ij

Now, we recall the transformation law of the Christoffel symbols, namely:

r f u ( x ) γ ij r ( x )= ij f u ( x )+ i f k ( x ) j f l ( x ) γ ¯ kl u ( f( x ) ) 1 Δ r f u γ ij r σ ij = 1 Δ ij f u σ ij + γ ¯ kl u ( y ) τ kl

Eliminating the second derivatives of f we finally get:

ψ u = τ ku y k + γ ¯ kl u τ kl = 1 Δ r f u ( i σ ir + γ ij r σ ij )= 1 Δ r f u φ r

This tricky technical result, which is not evident at all, explains why the additional term we had is just disappearing in fact when σ is a density.

One can prove, in a similar but even simpler fashion, that the two sets of Maxwell equations are invariant under any invertible transformation and that that the conformal group of spacetime is only the group of invariance of the Minkowski constitutive laws in vacuum [4].

Linearizing the Ricci tensor ρ ij over the Minkowski metric ω , we obtain the usual second order homogeneous Ricci operator ΩR with 4 terms (This result can be found in any textbook on general relativity but [24] [25] are elementary references using the same notations):

2 R ij = ω rs ( d rs Ω ij + d ij Ω rs d ri Ω sj d sj Ω ri )=2 R ji tr( R )= ω ij R ij = ω ij d ij tr( Ω ) ω ru ω sv d rs Ω uv

We may define the Einstein operator by setting E ij = R ij 1 2 ω ij tr( R ) and

obtain the 6 terms:

2 E ij = ω rs ( d rs Ω ij + d ij Ω rs d ri Ω sj d sj Ω ri ) ω ij ( ω rs ω uv d rs Ω uv ω ru ω sv d rs Ω uv )

We have the following crucial but purely mathematical theorem [24]:

Main Theorem 3.2: The Einstein operator S 2 T * Einstein S 2 T * is self-adjoint but the Ricci operator S 2 T * Ricci S 2 T * is NOT self-adjoint.

Proof: Multiplying on the left by (test functions) Lagrange multipliers λ ij , integrating by parts and changing the dumb indices if necessary in order to factor out Ω ij on the right of each adjoint, we get successively after numbering the 6 different terms:

1 λ ij ( ω rs d rs Ω ij ) ad ( ω rs d rs λ ij ) Ω ij 1 2 λ ij ( ω rs d ij Ω rs ) ad ( ω ij d rs λ rs ) Ω ij 6 3 λ ij ( ω rs d ri Ω sj ) ad ( ω ri d rs λ sj ) Ω ij 3 4 λ ij ( ω rs d sj Ω ri ) ad ( ω sj d rs λ ri ) Ω ij 4 5 λ ij ( ω ij ω rs ω uv d rs Ω uv ) ad ( ω uv ω rs d uv λ rs ) ω ij Ω ij 5 6 λ ij ( ω ij ω ru ω sv d rs Ω uv ) ad ( ω ui ω vj ω rs d uv λ rs ) Ω ij 2

As a byproduct, the operators 1,3+4,5 are self-adjoint while ad( 2 )=6 and thus ad( 6 )=2 . It follows that the sum 1+2+3+4+5+6 is self-adjoint. However, as a delicate point for explicit computations, one must not forget that λ ij E ij = λ 11 E 11 +2 λ 12 E 12 + λ 22 E 22 + and the factor “2” must be taken into account. It is important, in order to understand the confusion of Einstein, to notice

that we have S 2 T * Ricci S 2 T * but n T * S 2 T ad( Ricci ) n T * S 2 T backwards”.

We have the (exact and locally exact) differential sequence of operators acting on sections of vector bundles where the order of an operator is written under its arrow:

T Killing 1 S 2 T * Riemann 2 F 1 Bianchi 1 F 2 n D n( n+1 )/2 D 1 n 2 ( n 2 1 )/ 12 D 2 n 2 ( n 2 1 )( n2 )/ 24

Our purpose is now first to study the differential sequence onto which its right part is projecting:

S 2 T * Einstein 2 S 2 T * div 1 T * 0 n( n+1 )/2 n( n+1 )/2 n 0

and then the following adjoint sequence:

0ad( T ) Cauchy ad( S 2 T * ) Beltrami ad( F 1 ) Lanczos ad( F 2 )

In this sequence, if E is a vector bundle over the ground manifold X with dimension n , we may introduce, as we already said, the new vector bundle ad( E )= n T * E * where E * is obtained from E by inverting the transition rules exactly like T * is obtained from T . We have for example ad( T )= n T * T * n T * T n1 T * because T * is isomorphic to T by using the metric ω . The 10 × 10 Einstein operator matrix is induced from the 10 × 20 Riemann operator matrix and the 10 × 4 div operator matrix is induced from the 20 × 20 Bianchi operator matrix. We advise the reader not familiar with the formal theory of systems or operators to follow the computation in dimension n=2 with the 1 × 3 Airy operator matrix, which is the formal adjoint of the 3 × 1 Riemann operator matrix, and n=3 with the 6 × 6 Beltrami operator matrix which is the formal adjoint of the 6 × 6 Riemann operator matrix which is easily seen to be self-adjoint up to a change of basis. With more details, we have:

n=2 : The stress equations become d 1 σ 11 + d 2 σ 12 =0, d 1 σ 21 + d 2 σ 22 =0 . Their second order parametrization σ 11 = d 22 ϕ , σ 12 = σ 21 = d 12 ϕ , σ 22 = d 11 ϕ has been provided by George Biddell Airy in 1863 and is well known in plane elasticity. We get the second order system:

{ σ 11 d 22 ϕ=0 σ 12 d 12 ϕ=0 σ 22 d 11 ϕ=0 1 2 1 1

which is involutive with one equation of class 2, 2 equations of class 1 and it is easy to check that the 2 corresponding first order CC are just the Cauchy equations. Of course, the Airy function (1 term) has absolutely nothing to do with the perturbation of the metric (3 terms). With more details, when ω is the Euclidean metric, we may consider the only component:

tr( R )=( d 11 + d 22 )( Ω 11 + Ω 22 )( d 11 Ω 11 +2 d 12 Ω 12 + d 22 Ω 22 ) = d 22 Ω 11 + d 11 Ω 22 2 d 12 Ω 12

Multiplying by the Airy function ϕ and integrating by parts, we discover that:

Airy=ad( Riemann )Riemann=ad( Airy )

in the following adjoint differential sequences:

2 Killing 1 3 Riemann 2 1 0 0 2 Cauchy 1 3 Airy 2 1

n=3 : It is quite more delicate to parametrize the 3 PD equations:

d 1 σ 11 + d 2 σ 12 + d 3 σ 13 =0, d 1 σ 21 + d 2 σ 22 + d 3 σ 23 =0, d 1 σ 31 + d 2 σ 32 + d 3 σ 33 =0

A direct computational approach has been provided by Eugenio Beltrami in 1892, James Clerk Maxwell in 1870 and Giacinto Morera in 1892 by introducing the 6 stress functions ϕ ij = ϕ ji in the Beltrami parametrization. The corresponding system:

{ σ 11 d 33 ϕ 22 + d 22 ϕ 33 2 d 23 ϕ 23 =0 σ 12 d 33 ϕ 12 + d 12 ϕ 33 d 13 ϕ 23 d 23 ϕ 13 =0 σ 22 d 33 ϕ 11 + d 11 ϕ 33 2 d 13 ϕ 13 =0 σ 13 d 23 ϕ 12 + d 12 ϕ 23 d 22 ϕ 13 d 13 ϕ 22 =0 σ 23 d 23 ϕ 11 + d 11 ϕ 23 d 12 ϕ 13 d 13 ϕ 12 =0 σ 33 d 22 ϕ 11 + d 11 ϕ 22 2 d 12 ϕ 12 =0 1 2 3 1 2 3 1 2 3 1 2 1 2 1 2

is involutive with 3 equations of class 3, 3 equations of class 2 and no equation of class 1. We have dim( g 2 )=dim( S 2 T * S 2 T * )dim( S 2 T * )=( 6×6 )6=30 . The 3 CC are describing the stress equations which admit therefore a parametrization ... but without any geometric framework, in particular without any possibility to imagine that the above second order operator is nothing else but the formal adjoint of the Riemann operator, namely the (linearized) Riemann tensor with n 2 ( n 2 1 )/2 =6 independent components when n=3 .

Breaking the canonical form of the six equations which is associated with the Janet tabular, we may rewrite the Beltrami parametrization of the Cauchy stress equations as follows, after exchanging the third row with the fourth row, keeping the ordering { ( 11 )<( 12 )<( 13 )<( 22 )<( 23 )<( 33 ) } :

( d 1 d 2 d 3 0 0 0 0 d 1 0 d 2 d 3 0 0 0 d 1 0 d 2 d 3 )( 0 0 0 d 33 2 d 23 d 22 0 d 33 d 23 0 d 13 d 12 0 d 23 d 22 d 13 d 12 0 d 33 0 2 d 13 0 0 d 11 d 23 d 13 d 12 0 d 11 0 d 22 2 d 12 0 d 11 0 0 )0

as an identity where 0 on the right denotes the zero operator. However, if Ω is a perturbation of the metric ω , the standard implicit summation used in continuum mechanics is, when n=3 :

σ ij Ω ij = σ 11 Ω 11 +2 σ 12 Ω 12 +2 σ 13 Ω 13 + σ 22 Ω 22 +2 σ 23 Ω 23 + σ 33 Ω 33 = Ω 22 d 33 ϕ 11 + Ω 33 d 22 ϕ 11 2 Ω 23 d 23 ϕ 11 + + Ω 23 d 13 ϕ 12 + Ω 13 d 23 ϕ 12 Ω 12 d 33 ϕ 12 Ω 33 d 12 ϕ 12 +

because the stress tensor density σ is supposed to be symmetric. Integrating by parts in order to construct the adjoint operator, we get:

ϕ 11 d 33 Ω 22 + d 22 Ω 33 2 d 23 Ω 23 ϕ 12 d 13 Ω 23 + d 23 Ω 13 d 33 Ω 12 d 12 Ω 33

and so on. The identifications Beltrami=ad( Riemann ) , Lanczos=ad( Bianchi ) in the diagram:

3 Killing 1 6 Riemann 2 6 Bianchi 1 3 0 0 3 Cauchy 1 6 Beltrami 2 6 Lanczos 1 3

prove that the Cauchy operator has nothing to do with the Bianchi operator.

When ω is the Euclidean metric, the link between the two sequences is established by means of the elastic constitutive relations 2 σ ij =λtr( Ω ) ω ij +2μ Ω ij with the Lamé elastic constants ( λ,μ ) but mechanicians are usually setting Ω ij =2 ϵ ij . Using the standard Helmholtz decomposition ξ = φ+ ψ and substituting in the dynamical equation d i σ ij = ρ d 2 / d t 2 ξ j where ρ is the mass per unit volume, we get the longitudinal and transverse wave equations, namely

Δφ ρ λ+2μ d 2 d t 2 φ=0 and Δ ψ ρ μ d 2 d t 2 ψ =0 , responsible for earthquakes!

Then, taking into account the factor 2 involved by multiplying the second, third and fifth row by 2, we get the new 6 × 6 operator matrix with rank 3 which is clearly self-adjoint:

( 0 0 0 d 33 2 d 23 d 22 0 2 d 33 2 d 23 0 2 d 13 2 d 12 0 2 d 23 2 d 22 2 d 13 2 d 12 0 d 33 0 2 d 13 0 0 d 11 2 d 23 2 d 13 2 d 12 0 2 d 11 0 d 22 2 d 12 0 d 11 0 0 )( Ω 11 Ω 12 Ω 13 Ω 22 Ω 23 Ω 33 )=1( E 11 2 E 12 2 E 13 E 22 2 E 23 E 33 )

Surprisingly, the Maxwell parametrization is obtained by keeping ϕ 11 =A, ϕ 22 =B, ϕ 33 =C while setting ϕ 12 = ϕ 23 = ϕ 31 =0 but other parametrizations may exist like:

{ σ 11 d 33 ϕ 22 =0 σ 12 d 33 ϕ 12 =0 σ 22 d 33 ϕ 11 =0 σ 13 d 23 ϕ 12 d 13 ϕ 22 =0 σ 23 d 23 ϕ 11 d 13 ϕ 12 =0 σ 33 d 22 ϕ 11 + d 11 ϕ 22 2 d 12 ϕ 12 =0 1 2 3 1 2 3 1 2 3 1 2 1 2 1 2

When n=4 , the following crucial corollary is showing that the Einstein operator is useless, contrary to the classical GR literature [26].

Main Corollary 3.3: The GW equations are described by the adjoint of the Ricci operator which is not self-adjoint contrary to the Einstein operator which is self-adjoint.

Proof: Multiplying the Ricci operator by the Lagrange multipliers λ ij = λ ji used as test functions, setting = ω rs d rs and integrating by parts, we get the adjoint operator ad( Ricci ):σλ :

λ rs + ω rs d ij λ ij ω sj d ij λ ri ω ri d ij λ sj = σ rs

that is, exactly but backwards, the operator defining GW in the literature [26]. We also obtain:

d r σ rs = ω ij d rij λ rs + ω rs d rij λ ij ω sj d rij λ ri ω ri d rij λ sj =0

and finally the commutative diagram coherent with differential double duality:

4 Killing 1 10 Ricci 2 10 4 0 0 4 Cauchy 1 10 ad( Ricci ) 2 10

It follows that GW cannot exist as they cannot be considered as ripples of space-time because the Lagrange multiplier λ has nothing to do with the deformation Ω of the metric ω .

We finally prove that this result only depends on the second order jets of the conformal group of transformations of space-time, a result highly not evident at first sight for sure and not known. We need a few steps in order to show that the mathematical foundations of conformal geometry must be entirely revisited because the importance of acyclicity is not known in this framework.

4. Conformal Group

We start proving that the structure of the conformal with ( n+1 )( n+2 )/2 parameters may not be related to a classification of Lie algebras [27].

For n=1 , the simplest such group of transformations of the real line with 3 parameters is the projective group defined by the Schwarzian third order OD equation:

Φ( y, y x , y xx , y xxx ) y xxx y x 3 2 ( y xx y x ) 2 =ν( x )

with linearization the only third order Medolaghi equation with symbol g 3 =0 and no CC:

L( ξ 3 )ν ξ xxx +2ν( x ) ξ x +ξ x ν( x )=0

When ν=0 , the general solution is simply ξ= 1 2 a x 2 +bx+c with 3 parameters, that is to say 1 translation + 1 dilatation + 1 elation with respective generators { θ 1 = x , θ 2 =x x , θ 3 = 1 2 x 2 x } .

For n=2 , eliminating the conformal factor in the case of the Euclidean metric of the plane provides the two Cauchy-Riemann equations defining the infinitesimal complex transformations of the plane. The only possibility coherent with homogeneity is thus to consider the following system and to prove that it is defining a system of infinitesimal Lie equations, leading to 6 infinitesimal generators, namely: 2 translations + 1 rotation + 1 dilatation + 2 elations:

{ ξ ijr k =0 ξ 22 2 ξ 12 1 =0, ξ 22 1 + ξ 12 2 =0, ξ 12 2 ξ 11 1 =0, ξ 12 1 + ξ 11 2 =0 ξ 2 2 ξ 1 1 =0, ξ 2 1 + ξ 1 2 =0

{ θ 1 = 1 , θ 2 = 2 , θ 3 = x 1 2 x 2 1 , θ 4 = x 1 1 + x 2 2 , θ 5 = 1 2 ( ( x 1 ) 2 + ( x 2 ) 2 ) 1 + x 1 ( x 1 1 + x 2 2 ), θ 6 = 1 2 ( ( x 1 ) 2 + ( x 2 ) 2 ) 2 + x 2 ( x 1 1 + x 2 2 ) }

with [ θ 5 , θ 6 ]=0 . We have g ^ 3 =0 when n=1,2 .

Remark 4.1: (Special relativity): Though surprising it may look like, the conformal case when ( x 1 =x, x 2 =ct ) perfectly fits with the original presentation of Lorentz transformations if one uses the “ hyperbolic” notations. Indeed, setting th( ϕ )=u/c , th( ψ )=v/c , we obtain easily for the composition of speeds th( ϕ+ψ )= ( th( ϕ )+th( ψ ) )/ ( 1+th( ϕ )th( ψ ) ) = ( u/c +v/c )/ ( 1+( u/c )( v/c ) ) and a similar result still holds for the plane rotation with the usual tg instead of th .

Hence, one cannot distinguish between the time derivative of the position 4 ξ k and the infinitesimal rotation ξ 4 k , that is one can only give a meaning to the difference as a component of the Spencer operator. Indeed, an accelerometer in a rocket only measures the difference between the acceleration which is the time derivative of the speed 4 ξ 4 k and the gravitation ξ 44 k , that is another component of the Spencer operator, a reason for which elations were sometimes called accelerations at the beginning of the last century along with the following tabular:

gauging{ speed rotations 4 ξ k ξ 4 k acceleration elations 4 ξ 4 k ξ 44 k }Spenceroperator

Lemma 4.2: When there is a conformal factor, we have as in [28]:

g ^ 1 is finite type with g ^ 3 =0,n3 .

g ^ 2 is 2-acyclic when n4 .

g ^ 2 is 3-acyclic when n5 .

For n=3 , in order to convince the reader that classical and conformal differential geometry must be revisited, let us prove that the analogue of the Weyl tensor is made by a third order self-adjoint operator, a result which is neither known nor acknowledged today. We shall proceed by diagram chasing as the local computation done by using computer algebra does not provide any geometric insight (See arXiv:1603.05030 for the details). We have E=T and dim( F ^ 0 )=5 in the following commutative diagram providing F ^ 1 and where the vertical arrows are δ -maps:

0 0 0 0 g ^ 4 S 4 T * T S 3 T * F ^ 0 F ^ 1 0 δ δ δ 0 T * g ^ 3 T * S 3 T * T T * S 2 T * F ^ 0 0 δ δ δ 0 2 T * g ^ 2 2 T * S 2 T * T 2 T * T * F ^ 0 0 δ δ δ 0 3 T * g ^ 1 3 T * T * T 3 T * F ^ 0 0 0 0 0

0 0 0 45 50 5 0 0 90 90 0 0 9 54 45 0 0 4 9 5 0 0 0 0

A delicate double circular chase provides F ^ 1 = H 2 2 ( g ^ 1 ) in the short exact sequence:

0 F ^ 1 2 T * g ^ 2 δ 3 T * g ^ 1 0059 δ 40

We first notice that the map δ on the bottom left is surjective, a result that it is almost impossible to find in local coordinates. Let us prove it by means of circular diagram chasing in the preceding commutative diagram as follows. Lift any a 3 T * g ^ 1 3 T * T * T to b 2 T * S 2 T * T because the vertical δ -sequence for S 4 T * is exact. Project it by the symbol map σ 1 ( Φ ^ ) to c 2 T * T * F ^ 0 . Then, lift c to d T * S 2 T * F ^ 0 that we may lift backwards horizontally to e T * S 2 T * T to which we may apply δ to obtain f 2 T * S 2 T * T . By commutativity, both f and b map to c and the difference fb maps thus to zero. Finally, we may find g 2 T * g ^ 2 such that b=g+δ( e ) and we obtain thus a=δ( g )+ δ 2 ( e )=δ( g ) , proving therefore the desired surjectivity. We have 10 parameters: 3 translations + 3 rotations + 1 dilatation + 3 elations and the totally unexpected formally exact sequences on the jet level are thus, showing in particular that second order CC do not exist:

0 R ^ 3 J 3 ( T ) J 2 ( F ^ 0 )001060500

0 R ^ 4 J 4 ( T ) J 3 ( F ^ 0 ) F ^ 1 001010510050

0 R ^ 5 J 5 ( T ) J 4 ( F ^ 0 ) J 1 ( F ^ 1 ) F ^ 2 0 0101681752030

We obtain the minimum differential sequence, which is nervertheless not a Janet sequence:

0 Θ ^ T D ^ 1 F ^ 0 D ^ 1 3 F ^ 1 D ^ 2 1 F ^ 2 00 Θ ^ 3 D ^ 1 5 3 5 1 30

with D ^ the conformal Killing operator and vanishing Euler-Poincaré characteristic 35+53=0 .

We have proved in [5] that D 1 is self-adjoint.

For n=4 , we have 4 translations + 6 rotations + 1 dilatation + 4 elations = 15 parameters.

Also, g ^ 3 =0 g ^ 4 =0 g ^ 5 =0 in the conformal case, we have the commutative diagram with exact vertical long δ -sequences but the left one and where the second row proves that, contrary to what is still believed today:

There cannot exist first order Bianchi-like CC identities for the Weyl operator.

0 0 0 0 0 g ^ 4 S 4 T * T S 3 T * F ^ 0 T * F ^ 1 0 δ δ δ 0 T * g ^ 3 T * S 3 T * T T * S 2 T * F ^ 0 T * F ^ 1 0 δ δ δ 0 2 T * g ^ 2 2 T * S 2 T * T 2 T * T * F ^ 0 0 δ δ δ 0 3 T * g ^ 1 = 3 T * T * T 3 T * F ^ 0 0 δ δ 0 4 T * T = 4 T * T 0 0 0

0 0 0 0 140 180 40 0     δ     δ 0 320 360 40 0     δ     δ 0 24 240 216 0     δ     δ     δ 0 28 64 36 0     δ     δ 0 4 = 4 0 0 0

A diagonal snake chase proves that F ^ 1 H 2 ( g ^ 1 ) . However, we have the δ -sequence:

0 T * g ^ 2 δ 2 T * g ^ 1 δ 3 T * T0

We obtain dim( B 2 2 ( g ^ 1 ) )=4×4=16 and let the reader prove as before that the map δ on the bottom left is surjective, a result leading to dim( Z 2 2 ( g ^ 1 ) )=( 6×( 6+1 ) )( 4×4 )=4216=26 . The Weyl tensor has thus dim( F ^ 1 )=2616=10 components, a way that must be compared to the standard one that can be found in the GR literature. We obtain the minimum differential sequence, which is nervertheless not a Janet sequence:

0 Θ ^ T D ^ 1 F ^ 0 D ^ 1 2 F ^ 1 D ^ 2 2 F ^ 2 D ^ 3 1 F ^ 3 00 Θ ^ 4 1 9 2 10 2 9 1 40

Our purpose is to exhibit directly the Cauchy, Cosserat and Maxwell equations by computing with full details the adjoint of the first Spencer operator D 1 : R ^ 3 T * R ^ 3 for the conformal involutive finite type third order system R ^ 3 J 3 ( T ) for any dimension n1 . In general, one has n translations + n( n1 )/2 rotations + 1 dilatation + n nonlinear elations, that is a total of ( n+1 )( n+2 )/2 parameters, thus 15 when n=4 . As a byproduct, the Cosserat couple-stress equations will be obtained for the Killing involutive finite type second order system R ^ 2 J 2 ( T ) . It must be noticed that not even a single comma must be changed when n=3 when our results are compared to the original formulas provided by the bothers Cosserat in 1909 [16] [17]. We only need recall the specific features of the standard first order Spencer operator d: R ^ 3 T * R ^ 2 as follows by considering the multi-indices for the various parameters, separately as follows:

ξ 3 ( i ξ k ( x ) ξ i k ( x ), i ξ j k ( x ) ξ ij k ( x ), i ξ r r ( x ) ξ ri r ( x ), r ξ ij k ( x ) ξ ijr k ( x ) )

in the duality summation:

σ k i ( i ξ k ( x ) ξ i k ( x ) )+ μ k ij ( i ξ j k ( x ) ξ ij k ( x ) )+ ν i ( i ξ r r ( x ) ξ ri r ( x ) ) + π k ij,r ( r ξ ij k ( x ) ξ ijr k ( x ) )

We obtain a first simplification by noticing that the third order jets vanish, that is to say ξ rij k =0 . Indeed, starting with the Euclidean or Minkowski metric ω with vanishing Christoffel symbols γ=0 , the second order conformal equations can be provided in the parametric form:

ξ ij k = δ i k A j ( x )+ δ j k A i ( x ) ω ij ω kr A r ( x ) ξ ri r =n A i ( x )

The desired result follows from the fact that this system is homogeneous and g ^ 3 =0,n3 .

A second simplification may be obtained by using the (constant) metric in order to raise or lower the indices in the implicit summations considered. In particular, we have successively:

ω rj ξ i r + ω ir ξ i r =2A( x ) ω ij A( x )= ξ 1 1 ( x )= ξ 2 2 ( x )== ξ n n ( x )= 1 n ξ r r ( x )

In this situation, σ i,j ξ i,j = i<j ( σ i,j σ j,i ) ξ i,j + 1 n σ r r ξ r r and we may set

μ k ij ξ ij k = μ i A i where μ i is a linear (tricky) function of the μ k ij with constant coefficients only depending on ω . The new equivalent duality summation becomes:

σ i,r r ξ i + i<j ( μ ij,r r ξ i,j ( σ i,j σ j,i ) ξ i,j ) σ r r A( x ) μ i A i ( x ) + ν i ( i A( x ) A i ( x ) )+ π i,r ( r A i ( x ) )

When n=4 , the comparison with the Maxwell equations of electromagnetism is easily obtained as follows. Indeed, writing a part of the dualizing summation in the form:

J i ( i A A i )+ 1 2 ij ( i A j j A i ) = J i A i + ij ij ( i A j j A i )+ = J 1 A 1 ++ 12 ( 1 A 2 2 A 1 )+ F 13 ( 1 A 3 3 A 1 ) + 14 ( 1 A 4 4 A 1 )+ =( J 1 A 1 ++( 12 2 A 1 + 13 3 A 1 + 14 4 A 1 )+ ) =div( )+( J 1 + 2 12 + 3 13 + 4 14 ) A 1 +

Integrating by parts and changing the sign as usual, we obtain as usual the second set of Maxwell equations for the induction :

r ir J i =0 i J i = ij ij =0

Such a result is coherent with the virial equation on the condition to have σ r r =0 in a coherent way with the classical Maxwell stress tensor density:

σ j i = ir F rj + 1 4 δ j i rs F rs σ r r =0

which is traceless with a divergence producing the Lorentz force as we have indeed when n=4 :

dF=0 i σ j i = J r F rj + 1 2 rs ( r F sj + s F jr + j F rs )= J r F rj

The mathematical foundations of EM, that is both the first and second Maxwell equations, thus only depend on the group structure of the conformal group of space-time, a fact that can only be understood by using the Spencer operator and is therefore not even known. Our purpose at the end of this paper is to consider only the linearized framework. The crucial idea is to notice that the Poisson equation has only to do with the trace of the stress tensor density, contrary to the EM situation as we just saw.

Defining the vector bundle F ^ 0 = J 1 ( T )/ R ^ 1 T * g ^ 1 when n4 , another difficulty can be discovered in the following commutative and exact diagrams obtained by applying the Spencer δ -map to the symbol sequence with dim( g ^ 1 )=dim( g 1 )+1=( n( n1 )/2 )+1 :

0 g ^ 1 T * T F ^ 0 0

then to its first prolongation with dim( g ^ 2 )=n :

0 g ^ 2 S 2 T * T T * F ^ 0 0

and finally to its second prolongation in which g ^ 3 =0 :

0 0 0 0 g ^ 3 S 3 T * T S 2 T * F ^ 0 F ^ 1 0 δ δ δ 0 T * g ^ 2 T * S 2 T * T T * T * F ^ 0 0 δ δ δ 0 2 T * g ^ 1 2 T * T * T _ 2 T * F ^ 0 0 δ δ 0 3 T * T = 3 T * T 0 0 0

A snake chase allows to introduce the Weyl bundle F ^ 1 defined by the short exact sequence:

0 T * g ^ 2 δ Z 1 2 ( g ^ 1 ) F ^ 1 0

in which the cocycle bundle Z 1 2 ( g ^ 1 ) is defined by the short exact sequence:

0 Z 1 2 ( g ^ 1 ) 2 T * g ^ 1 δ 3 T * T0

We have of course dim( F ^ 1 )=10 when n=4 but more generally:

dim( F ^ 1 )=( n( n+1 )/2 )( n( n+1 )/2 1 ) n 2 ( n+1 )( n+2 )/6 =( ( n( n1 )/2 )( n( n1 )/2 +1 ) n 2 ( n1 )( n2 )/6 ) n 2 = n( n+1 )( n+2 )( n3 )/ 12

In the purely Riemannian case, as g 2 =0 , we have F 1 Z 1 2 ( g 1 ) and thus:

dim( F 1 )=( n( n+1 )/2 )( n( n+1 )/2 1 ) n 2 ( n+1 )( n+2 )/6 = n 2 ( ( n1 )/2 ) 2 n 2 ( n1 )( n2 )/6 = n 2 ( n 2 1 )/ 12

with the unexpected formula dim( F 1 )dim( F ^ 1 )=dim( S 2 T * )= n( n+1 )/2 .

No classical method can produce such results allowing to obtain the following Fundamental Diagram II provided as early as in 1983 and only valid whenever g ^ 3 =0 [29] [30].

0 0 Ricci 0 Z 1 2 ( g 1 ) Riemann 0 0 T * g ^ 2 δ Z 1 2 ( g ^ 1 ) Weyl 0 0 S 2 T * δ T * T * δ 2 T * 0 0 0

Theorem 4.3: This commutative and exact diagram splits and a diagonal snake chase proves that Ricci S 2 T * in a coherent way with the previous formulas.

Proof: The monomorphism δ: S 2 T * T * T * splits with

1 2 ( A i,j + A j,i ) A i,j while the epimorphism δ: T * T * 2 T * : A i,j A i,j A j,i splits with 1 2 F ij F ij = F ji . We explain

how the well known result T * T * S 2 T * 2 T * , which is coming from the elementary formula n 2 = n( n+1 )/2 + n( n1 )/2 , may be related to the Spencer δ -cohomology interpretation of the Riemann and Weyl bundles. For this, we have to give details on the “snake” chase:

S 2 T * T * T * T * g ^ 2 Z 1 2 ( g ^ 1 ) Z 1 2 ( g 1 )RiemannRicci

Starting with ( A ij = A i,j = A j,i = A ji ) S 2 T * T * T * , we may define:

ξ ri,j r =n A i,j =n A ij =n A ji = ξ rj,i r ( ξ lj,i k = δ l k A j,i + δ j k A l,i ω ij ω kr A r,i ) T * g ^ 2

( R l,ij k = ξ li,j k ξ lj,i k ) Z 1 2 ( g ^ 1 ) 2 T * g ^ 1 2 T * T * T

R r,ij r = ξ zi,j r ξ rj,i r =n( A i,j A j,i )=0( R l,ij k ) Z 1 2 ( g 1 )

n R l,ij k =( δ l k ξ ri,j r + δ i k ξ rl,j r ω li ω ks ξ rs,i r )( δ l k ξ rj,i r + δ j k ξ rl,i r ω lj ω ks ξ rs,i r )

R l,ij k =( δ i k A lj δ j k A li ) ω ks ( ω li A sj ω lj A si )

Introducing tr( A )= ω ij A ij and R ij = R i,rj r =( n A ij A ij )( A ij ω lj tr( A ) ) , we get:

R ij =( n2 ) A ij + ω ij tr( A )= R ji tr( R )= ω ij R ij =2( n1 )tr( A )

Substituting, we finally obtain A ij = 1 n2 R ij 1 2( n1 )( n2 ) ω ij tr( R ) and the tricky formula:

R l,ij k = 1 n2 ( δ i k R lj δ j k R li ω ks ( ω li R sj ω lj R si ) ) 1 ( n1 )( n2 ) ( δ i k ω lj δ j k ω li )tr( R )

totally independently from the standard elimination of the derivatives of a conformal factor, contrary to the way used in most textbooks.

Contracting in k and i , we obtain indeed the lift:

Riemann= H 1 2 ( g 1 ) S 2 T * Ricci: R l,ij k R i,rj r = R ij = R ji

in a coherent way. Using a standard result of homological algebra [22] or section 2, we obtain therefore a splitting Weyl= H 1 2 ( g ^ 1 ) H 1 2 ( g 1 )=Riemann :

W l,ij k = R l,ij k 1 n2 ( δ i k R lj δ j k R li ω ks ( ω li R sj ω lj R si ) ) 1 ( n1 )( n2 ) ( δ i k ω lj δ j k ω li )tr( R )

in such a way that W i,rj r =0 , a result leading to the isomorphism RiemannRicciWeyl .

We are now ready to apply the previous diagrams by proving the following crucial Theorem:

Theorem 4.4: When n=4 , the linear Spencer sequence for the Lie algebra Θ ^ of infinitesimal conformal group of transformations projects onto a part of the Poincaré sequence for the exterior derivative with a shift by one step according to the following commutative and locally exact diagram:

0 Θ ^ j 2 R ^ 2 D 1 T * R ^ 2 D 2 2 T * R ^ 2 T * d 2 T * d 3 T *

This purely mathematical result also contradicts classical gauge theory because it proves that EM only depends on the structure of the conformal group of space-time but not on U( 1 ) .

Proof: Restricting our study to the linear framework, we introduce a new system R ˜ 1 J 1 ( T ) of infinitesimal Lie equations defined by L( ξ 1 )ω=2Aω with prolongation defined by by L( ξ 2 )γ=0 in such a way that R 1 R ˜ 1 = R ^ 1 with a strict inclusion and the strict inclusions R 2 R ˜ 2 R ^ 2 .

Indeed, from the definitions there is an isomorphism R ^ 2 / R ˜ 2 g ^ 2 and a first problem to solve is to construct vector bundles from the components of the image of D 1 . Using the corresponding capital letter for denoting the linearization, let us introduce:

i ξ μ k ξ μ+ 1 i k = X μ,i k B μ,i k ( tensors )

( B l,i k = X l,i k + γ ls k X ,i s ) T * T * T( B r,i r = B i ) T *

( B lj,i k = X lj,i k + γ sj k X l,i s + γ ls k X j,i s γ lj s X s,i k + X ,i r r γ lj k ) T * S 2 T * T ( B ri,j r B rj,i r = F ij ) 2 T *

We obtain from the relations i γ rj r = j γ ri r and the previous results:

F ij = B ri,j r B rj,i r = X ri,j r X rj,i r + γ rs r X i,j s γ rs r X j,i s + X ,j r r γ si s X ,i r r γ sj s = i X r,j r j X r,i r + γ rs r ( X i,j s X j,i s )+ X ,j r i γ sr s X ,i r j γ sr s = i ( X r,j r + γ rs r X ,j s ) j ( X r,i r + γ rs r X s,i s ) = i B j j B i

Now, using the contracted formula ξ ri r + γ rs r ξ i s + ξ s s γ ri r =n A i , we obtain:

B i =( i ξ r r ξ ri r )+ γ rs r ( i ξ s ξ i s ) = i ξ r r + γ rs r i ξ s + ξ s s γ ri r n A i = i ( ξ r r + γ rs r ξ s )n A i =n( i A A i )

and we finally get F ij =n( j A i i A j ) which is no longer depending on A , a result fully solving the dream of Weyl [31]. Of course, when n=4 and ω is the Minkowski metric, then we have γ=0 in actual practice and the previous formulas become particularly simple.

It follows that dB=FndA=F in 2 T * and thus dF=0 , that is i F jk + j F ki + k F ij =0 , has an intrinsic meaning in 3 T * . It is finally important to notice that the usual EM Lagrangian is defined on sections of C ^ 1 killed by D 2 but not on C ^ 2 . Finally, the south west arrow in the left square is the composition:

ξ 2 R ^ 2 D 1 X 2 T * R ^ 2 π 1 2 X 1 T * R ^ 1 ( γ ) ( B i ) T * ξ 2 R ^ 2 ( n A i ) T *

More generally, using the Lemma, we have the composition of epimorphisms:

C ^ r C ^ r / C ˜ r r T * ( R ^ 2 / R ˜ 2 ) r T * g ^ 2 r T * T * δ r+1 T *

Accordingly, though A and B are potentials for F , then B can also be considered as a part of the field but the important fact is that the first set of (linear) Maxwell equations dF=0 is induced by the (linear) operator D 2 because we are only dealing with involutive and thus formally integrable operators, a fact justifying the commutativity of the square on the left of the diagram.

5. Conclusions

Summarizing the results obtained in the preceding sections, we can only refer to the Zentralblatt review Zbl 1079.93001 for comments on the new mathematical methods that can be found in the corresponding book ([18],1000 pages !). We have successively obtained:

1) Our first task has been to revisit the concept of controllability still existing in classical control theory through a reference to the engineering choice of inputs and outputs. It is a pleasure to thank J.-L. Lions from INRIA who died too early in 2001 for measuring the importance of the help he provided me in 1990, allowing me to start at INRIA a solo intensive one week European ERCIM course (30 hours) held during five consecutive years in Paris, Bonn or Amsterdam and ending with the publication of [11], then [18]. The reason is that he understood for the first time that the possibility to extend control theory from OD systems to PD systems had to do with the injectivity of the adjoint operator, even though he was essentially interested by functional analysis. This paper is presenting in a rather self-contained way the elements allowing to prove that an OD or PD control system is controllable if and only if it is parametrizable by means of a certain number of functions and their derivatives up to a certain order through a constructive algorithm, a difficult problem even for elementary examples like the double pendulum or the RCL electrical circuit. Such a new structural approach to controllability will lead to revisit many founding points of control theory or even mathematical physics, though striking it may look like for apparently well established theories as we saw. From a purely historical point of view, we may say that the use of homological algebra in physics will bring the same revolution as the one it brought in mathematics after 1950, not only an evolution ! However, when I had to learn about control theory through textbooks, my first surprise has ben to find a lot a examples fully treated or provided as exercises, often depending on a few constant parameters, like the double pendulum or the RCL circuits and such that the controllability was depending on a few equalities or inequalities between the parameters. It seemed to me that nobody did ever get in mind to exchange a few inputs with a few outputs, just for fun, as he should have discovered that controllability is indeed a “built-in” structural property. However, researchers through the world are not “adventurers” in general, perhaps because they are too much engaged into contracts with dead lines !

2) The Einstein operator has been written down for the first time 25 years before A. Einstein by the Italian mechanician E. Beltrami in dimension n=3 for parametrizing the Cauchy stress equations by the Beltrami stress functions known in elasticity where they are used as potentials through the Beltrami=ad( Riemann ) operator. The explicit comparison, that has never been done, needs no comment (see (4.1) in [4]) and the adjoint operator has never been used. Also, the Einstein operator is self-adjoint (who knows such a property even in dimension n=3 ) and Einstein made two dual confusions, one between Beltrami stress functions and the deformation of the metric, both having n( n+1 )/2 components, but also between the Cauchy=ad( Killing ) operator and the “div” operator induced from the Bianchi operator, by far the worst confusion, disappearing of course when n=2 (see for a recent summary). These two confusions can only be understood through homological algebra, because the Einstein operator goes from the variation of the metric to another symmetric tensor having nothing to do with stress but the adjoint of the Ricci operator goes from Lagrange multipliers λ=( λ ij = λ ij ) , used as stress functions having also nothing to do with the metric, to the stress tensor density σ=( σ ij = σ ji ) . Einstein equations in GR are thus not coherent with differential duality, contrary to Maxwell equations in EM. Also, according to Poincaré, as the (geometrical) left member is a tensor, the (physical) right member must be a tensor density. Hence, GW cannot be ripples of space-time produced by merging binary black holes and cannot thus exist for purely mathematical reasons. This is why Einstein hesitated so many times all along his life as he could not quote Beltrami for sure ! These results could have been found since 20 years because the double pendulum and the impossibility to parametrize the Einstein operator are already in ([8], p. 201). The reader may notice that the main results presented in did appear as early as in 2017 and that the reason for which black holes cannot exist [32] just appeared in 2025. We have also vainly tried to warn the European Space Agency (ESA) not to engage into its LISA project for the above reasons but future will judge !

3) Electromagnetism and gravitation only depend on the elations of the conformal group of space-time by chasing in the Fundamental diagram II along the dream of H. Weyl in 1918 [31]. However, such a result is not coherent with classical gauge theory because U( 1 ) is not acting on space-time contrary to the conformal group and also because the EM field is a section of the first Spencer bundle, not of the second Spencer bundle. Paraphrasing Shakespeare as in [30], we may say:

“TO ACT OR NOT TO ACT, THAT IS THE QUESTION !”

Conflicts of Interest

The author declares no conflicts of interest regarding the publication of this paper.

References

[1] Oberst, U. (1990) Multidimensional Constant Linear Systems. Acta Applicandae Mathematicae, 20, 1-175.[CrossRef
[2] Zerz, E. (2000) Topics in Multilinear Systems Theory. Springer Lecture Notes in Control and Information Sciences Vol. 256. Springer.
[3] Kashiwara, M. (1995) Algebraic Study of Systems of Partial Differential Equations. Mémoires de la Société Mathématique de France, 63, (Translation from Japanese of His 1970 Master Thesis).
[4] Pommaret, J. (2023) Gravitational Waves and Parametrizations of Linear Differential Operators. In: Frajuca, C., Ed., Gravitational WavesTheory and Observations, IntechOpen, 3-39.[CrossRef
[5] Pommaret, J.-F. (2024) Gravitational Waves and the Foundations of Riemannian Geometry. Advances in Mathematical Research, 35, 95-161.
[6] Pommaret, J.-F. (2024) From Control Theory to Gravitational Waves. Advances in Pure Mathematics, 14, 49-100.[CrossRef
[7] Pommaret, J.-F. (2025) Why Gravitational Waves Cannot Exist! Open Access Government, 45, 294-296.[CrossRef
[8] Pommaret, J.-F. (2005) 5 Algebraic Analysis of Control Systems Defined by Partial Differential Equations. In: Lamnabhi-Lagarrigue, F., Loría, A. and Panteley, E., Eds., Advanced Topics in Control Systems Theory, Springer, 155-223.[CrossRef
[9] Pommaret, J.-F. (2023) General Relativity and Gauge Theory: Beyond the Mirror.
[10] Pommaret, J.-F. (1995) Dualité Differentielle et Applications. Comptes Rendus de l’’Académie des Sciences Paris, Série I, 320, 1225-1230.
[11] Pommaret, J.-F. (1994) Partial Differential Equations and Group Theory. Kluwer.
[12] Macaulay, F.S. (1916) The Algebraic Theory of Modular Systems, Cambridge Tract 19. Cambridge University Press.
[13] Pommaret, J.F. (2011) Macaulay Inverse Systems Revisited. Journal of Symbolic Computation, 46, 1049-1069.[CrossRef
[14] Spencer, D.C. (1965) Overdetermined Systems of Partil Differential Equations. Bulletin of the American Mathematical Society, 75, 1-114.
[15] Pommaret, J.-F. (1978) Systems of Partial Differentia Equations and Group Theory. Gordon and Breach.
[16] Cosserat, E. and Cosserat, F. (1909) Théorie des Corps Déformables. Hermann.
[17] Pommaret, J. (2010) Parametrization of Cosserat Equations. Acta Mechanica, 215, 43-55.[CrossRef
[18] Pommaret, J.-F. (2001) Partial Differential Control Theory. Kluwer.
[19] Pommaret, J. (2019) The Mathematical Foundations of Elasticity and Electromagnetism Revisited. Journal of Modern Physics, 10, 1566-1595.[CrossRef
[20] Pommaret, J.-F. (2016) Deformation Theory of Algebraic and Geometric Structures. Lambert Academic Publisher (LAP).
[21] Pommaret, J. (2013) Relative Parametrization of Linear Multidimensional Systems. Multidimensional Systems and Signal Processing, 26, 405-437.[CrossRef
[22] Rotman, J.J. (1979) An Introduction to Homological Algebra, Pure and Applied Mathematics. Academic Press.
[23] Schneiders, J.-P. (1994) An Introduction to D-Modules. Bulletin de la Société Royale des Sciences de Liège, 63, 223-295.
[24] Pommaret, J.-F. (2013) The Mathematical Foundations of General Relativity Revisited. Journal of Modern Physics, 4, 223-239.[CrossRef
[25] Pommaret, J.-F. (2014) The Mathematical Foundations of Gauge Theory Revisited. Journal of Modern Physics, 5, 157-170.[CrossRef
[26] Foster, J. and Nightingale, J.D. (1979) A Short Course in General Relativity. Longman.
[27] Pommaret, J.-F. (2021) The Conformal Group Revisited. Journal of Modern Physics, 12, 1822-1842.[CrossRef
[28] Pommaret, J.-F. (2016) Chapter I. From Thermodynamics to Gauge Theory: The Virial Theorem Revisited. In: Bailey, L., Ed., Gauge Theories and Differential Geometry, Nova Science Publishers, 1-44.
[29] Pommaret, J.-F. (1983) La Structure de l’Electromagnetisme et de la Gravitation. Comptes Rendus de lAcadémie des Sciences, 297, 493-496.
[30] Pommaret, J.-F. (2012) Spencer Operator and Applications: From Continuum Mechanics to Mathematical Physics. In: Gan, Y., Ed., Continuum MechanicsProgress in Fundamentals and Engineering Applications, Intech, 1-32.[CrossRef
[31] Weyl, H. (1918+1952) Space, Time, Matter. Dover.
[32] Pommaret, J.-F. (2025) From Differential Sequences to Black Holes. Journal of Modern Physics, 16, 410-440.[CrossRef

Copyright © 2026 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.