The Partial Jacobian Matrix and the Chain Rule for Functions on Products of Real n-Spaces

Abstract

The usual chain rule is for compositions of functions from R n to R m , for some m and n . We consider functions whose domains and codomains are products of various R n ’s, and introduce the notion of a partial Jacobian matrix, using which we formulate a chain rule in this setting that should be useful for calculations as well as for some theoretical purposes.

Share and Cite:

Agashe, A. (2025) The Partial Jacobian Matrix and the Chain Rule for Functions on Products of Real n-Spaces. Applied Mathematics, 16, 420-427. doi: 10.4236/am.2025.164022.

1. Introduction

The goal of this article is to fill some potential gaps in the literature and introduce some notation that should be useful in applications. To motivate the main result, which may look abstract, we consider a concrete example. The reader who just wants to see and use the main result of this paper may jump directly to Section 3.

Let f: R 2 R 2 be given by x( u,t )=u+t , y( u,t )=ut , and g: R 2 R be given by z= x 2 +y . Thinking of z as a function of ( u,t ) via f , if one wants to find the rate of change of z with respect to u at the point ( 1,1 ) , we can use the chain rule

z u = z x x u + z y y u , (1)

to get z u =2x1+11=2( u+t )+1=2u+2t+1 , which evaluated at ( 1,1 ) gives the answer 5. Note that found the answer without computing the function gf , which is (u+t) 2 +( ut ) , which we could have differentiated with respect to u to get the same answer. The chain rule is especially useful if one wants to find the rate of change of z with respect to u , knowing the other rates of change in Equation (1); this happens often in real life situations, where one may not know the functions f and g explicity, but knows the instantaneous rates of change on the right side of Equation (1). Such problems are called related rates problems.

Now suppose we are given functions f: R 2 × R 3 R 4 × R 5 and g: R 4 × R 5 R . We can compose the two to get the function gf: R 2 × R 3 R . Let z denote the coordinate of the codomain R , and suppose we want to find the directional derivative of z relative to a change in the coordinates of the R 2 in the domain of gf (keeping the coordinates of R 3 fixed) at some given point in the domain R 2 × R 3 . We could, of course, write down the function gf in terms of the coordinates of R 2 × R 3 and then take partial derivatives, etc., but that is more work than needed. In this article, we derive a chain rule (formula (8) below) where the calculation can be done in a manner similar to the chain rule in the previous paragraph, thus making the calculation much easier (see Remark 3.3). Note that in the example above, we did not give formulas to define f and g since the number of variables is too much; the reader can easily make up an explicit example. Our chain rule is also useful for related rate problems in this context, either for computations or for theoretical use. In fact, the reason for the author to write up this article was that a colleague in the Economics department asked for a chain rule in a context similar to the one above, and neither of us could find it in the form that was needed in the literature. So it is hoped that this article fills a potential gap in the literature, and can be quoted for the sake of completeness in papers that need this version of the chain rule.

The organization of the rest of this article is as follows. In Section 2, we recall the classical setting alluded to two paragraphs above, and introduce a notation for the Jacobian matrix that may be new, and is very convenient. In Section 3, we consider a more general setting of the type discussed in the previous paragraph, and give the more general chain rule alluded to above, which involves a new object that we call a partial Jacobian matrix. Finally, in Section 4, we give the proof of our generalization of the chain rule.

2. The Classical Setting

Suppose f: R n R m is differentiable. Let x denote the vector of coordinate functions on the domain (so x 1 ,, x n , the components of x , are the coordinates of R n ) and let y denote the vector of coordinate functions on the codomain. A standard notation for the Jacobian matrix of f is ( f 1 ,, f m ) ( x 1 ,, x n ) and J f ; let us introduce another notation:

Definition 2.1. Let

dy dx

denote the Jacobian matrix of f .

The notation above is clearly related to the earlier notation (our notation may not be new, but we have not seen it before), and also makes notational sense: the

( j,i ) -th entry of dy dx is y j x i .

Let g: R m R k be differentiable, and let z denote the vector of coordinate functions on the codomain. Letting D denote the derivative as usual, the chain rule says that for a R n , we have

D( gf )( a )=Dg( f( a ) )Df( a ). (2)

This looks nice in theory and is very convenient for theoretical purposes, but has to be unraveled in applications and may be too abstract for people working in applied areas. A more concrete restatement is the matrix version:

J gf ( a )= J g ( f( a ) ) J f ( a ), (3)

where the product on the right is matrix multiplication. Equation (3) still has to be unraveled, as we shall do soon, but first note that in the notation of Definition 2.1, the equation becomes

dz dx = dz dy dy dx , (4)

where again the product on the right is the matrix product. The rule looks much nicer with the new notation, and moreover, if n=m=k=1 , then it becomes the usual chain rule in the Leibnitz notation, namely

dz dx = dz dy dy dx .

As usual, for j=1,,m , let f j denote the j -th component function of f ; thus f=( f 1 ,, f m ) . Recall that for i=1,,n , and for a R n , the symbol D i f j ( a ) denotes the partial deriviative of f j at a with respect to the i -th coordinate; thus D i f j ( a ) is y j x i evaluated at a . The following version of the chain rule is very explicit; it is well known, and the only reason that we are stating it as a result is that it does not seem to be easy to find the statement in the literature (for example, in ([1], Theorem~2.9) and ([2], pp. 61-62), it is only stated for k=1 , and in [3], it is only mentioned in the proof of Proposition 3.10).

Proposition 2.2. For i=1,,n , and =1,,k , we have for a R n ,

D i ( gf ) ( a )= j=1 m D j g ( f( a ) ) D i f j ( a ) , and(5)

z x i = j=1 m z y j y j x i . (6)

Proof. This follows from Equation (3), the definition of the Jacobian matrix, and the definition of matrix multiplication. □

Equation (6) above may be the most useful version of the chain rule in applications.

3. A More General Setting

For i=1,,n and j=1,,m , let X i and Y j each be R p for some p (so for example, we could have X 1 = R 2 , Y 2 = R 5 , etc.), and let u i and v j denote the dimensions of X i and Y j respectively (so X i = R u i and Y j = R v j ). Let f: X 1 ×× X n Y 1 ×× Y m be differentiable (regarding the domain and the codomain as finite dimensional vector spaces in the natural way). Note that if each of X i and Y j are R , then X 1 ×× X n = R n and Y 1 ×× Y m = R m as vector spaces, so we are in the classical setting of Section 2; thus our setup generalizes the classical setting. The function f above can be thought of as a function whose inputs and outputs, instead of being tuples of numbers, are tuples of vectors (not necessarily in the same vector space). Such functions arise often in applied areas such as Economics and Engineering.

For i=1,,n and j=1,,m , let x i denote the vector of coordinate functions on X i (so x i,1 ,, x i, u i , the components of x i , are the coordinates of X i ), and let y j denote the vector of coordinate functions on Y j .

Definition 3.1. For each i,j , denote by

y j x i

the matrix whose ( r,s ) -th entry for s=1,, u i and r=1,, v j is y r x s ; we call it a partial Jacobian matrix.

The reason for this notation and terminology is that while this object looks like the Jacobian matrix in the classical setting of Section 2, which might suggest the notation d y j d x i that was used for the Jacobian matrix therein, it plays the role of a partial dervative in the chain rule analog of (6), as we shall soon see (Equation (8) below).

For =1,,k , let Z each be R p for some p , and let z denote the vector of coordinate functions on Z . Let g: Y 1 ×× Y m Z 1 ×× Z k be differentiable. Thinking of each domain and codomain as real vector spaces in the natural way and letting D denote the derivative as usual, the chain rule again says that for a in the domain of f ,

D( gf )( a )=Dg( f( a ) )Df( a ). (7)

This equation is mostly of theoretical use, and is difficult to use in practice. Also, we are not distinguishing between the components of each domain (e.g., the components X 1 ,, X n of the domain of f ) in any way or treating an individual component special. A similar criticism holds for the analogs of equations (3) and (4), of which Equation (3) holds verbatim even in this situation.

We have the following version of the chain rule, which we prove in the next section:

Theorem 3.2. In the setup above, for =1,,k , and i=1,,n , we have

z x i = j=1 m z y j y j x i . (8)

The equation above distinguishes between the components of each domain, unlike Equation (7) above, and is the analog of Equation (6) from the classical setting, which it agrees with when each of X i , Y j , and Z is R . Equation (8) may be the most concrete form of the chain rule in this general setting, and should be especially useful for calculations, but also for some theoretical purposes.

Remark 3.3. As an example of an application, if u X i is a unit vector, then the directional derivative1 of the composite map X i X 1 ×× X n gf Z 1 ×× Z k Z (where the first map is the natural inclusion, and the last map is the natural projection) at a point a X 1 ×× X n in the direction of u is

z x i u=( j=1 m z y j y j x i )u= j=1 m ( z y j y j x i u )

evaluated at a (this follows from Theorem 3.2 and ([2], Theorem~5.1), for example).

Next we give a more theoretical version of Theorem 3.2, for which we need a generalization of the usual partial derivative.

Definition 3.4. For each j , let f j denote the composite function

X 1 ×× X n f Y 1 ×× Y m Y j ,

where the second map is the natural projection. Let a be an element of X 1 ×× X n . For each i , let ι a,i denote the map X i X 1 ×× X n that takes x X i to the element of X 1 ×× X n whose p -th component is the p -th component of a , except for p=i , when it is x . Denote by D i f j ( a ) the derivative at the i -th component of a of the composite map

X i ι a,i X 1 ×× X n f Y 1 ×× Y m Y j ,

where the last map is the natural projection; we call it the i -th partial derivative of f j at a .

An alternate definition, which is perhaps more conceptual, is given after Definition 4.1 in the next section. Note that the matrix of D i f j ( a ) in the standard bases is y j x i evaluated at a . If each of X i and Y j are R , then it is the usual partial derivative D i f j ( a ) ; we remark that even in this situation, our conceptual definition above (or the equivalent one in the next section) in terms of maps seems to be missing in the literature (e.g., it is not there in [1] [2], and [3]), and could be useful for theoretical purposes (e.g., it is immediate that the partial derivative of a sum is the sum of the partial derivatives, since the partial derivative inherits that property of the derivative, from our definition). Note that in what follows, the definitions of D j g and D i ( gf ) are analogous to that of D i f j .

Theorem 3.5. In the setup above, for =1,,k , and i=1,,n , we have for a in X 1 ×× X n ,

D i ( gf ) ( a )= j=1 m D j g ( f( a ) ) D i f j ( a ). (9)

The equation above also distinguishes between the components of each domain, unlike Equation (7) above, and is the analog of Equation (5) from the classical setting, which it agrees with when each of X i , Y j , and Z is R .

Remark 3.6. As mentioned in the second proof of Theorem 3.5 in the next section, the proof does not use the fact that the X i , Y j , and Z k are finite dimensional vector spaces over the real numbers, just that they are normed vectors spaces, so in particular Theorem 3.5 applies even if any of them are normed vector spaces of infinite dimension and/or over the complex numbers (over the complex numbers, the derivative is in the Fréchet sense, for example, as in ([4], Section VIII.1)).

Remark 3.7. Recall that given a smooth map f:XY between two smooth manifolds, the derivative induces a map df:TXTY between the tangent bundles of the manifolds. The classical chain rule translates to a chain rule for such maps. Similarly our chain rule (9) applies to maps between products of manifolds, and gives a chain rule for products of manifolds, which we now state more precisely. For each i , j , as above, let X i , Y j , and Z be manifolds2 and let f: X 1 ×× X n Y 1 ×× Y m and g: Y 1 ×× Y m Z 1 ×× Z k be morphisms. Then the classical chain rule says that d( gf )=dgdf . Now let d i f j denote the map induced by X i X 1 ×× X n f Y 1 ×× Y m Y j (where the first map is the natural inclusion and the second map is the natural projection) on the corresponding tangent bundles; similar definitions apply to d j g and d i ( gf ) . Then our discussion above gives the following application and analog of Theorem 3.5: d i ( gf ) = j=1 m d j g d i f j . In fact, by the previous remark, this will apply to certain infinite-dimensional manifolds, in particular to Banach manifolds.

4. Proofs of Theorems 3.2 and 3.5

We continue using the notation of the previous section. First note that Theorem 3.2 and Theorem 3.5 follow from each other by considering the basis of a linear transformation in the standard basis. We now prove Theorem 3.2; later, we shall give an independent proof of Theorem 3.5.

Let p be an indexing variable for the entries of z and q an indexing variable for the entries of x i . By definition, the ( p,q ) -th entry in the matrix on the left side of Equation (8) is z ,p x i,q . Now z ,p is a function of x i,q via the inter-mediate variables y j,r for j=1,,m and r=1,, v j for each j . So by the classical version (6) of the chain rule, we see that the left side of Equation (8) is

j=1 m r=1 v j z ,p y j,r y j,r x i,q . (10)

The ( p,q ) -th entry of the term z y j y j x i on the right side of Equation (8) is the product of the p -th row of z y j and the q -th column of y j x i . Now the p -th row of z y j is

[ z ,p y j,1 z ,p y j, v j ]

and the q -th column of y j x i is

[ y j,1 x i,q y j, v j x i,q ].

So the product of the row and column becomes

r=1 v j z ,p y j,r y j,r x i,q  .

As we sum these terms over j on the right side of Equation (8), we get the expression (10), which recall is the left side of Equation (8). This proves that the two sides of Equation (8) are equal, which proves Theorem 3.2, and thus Theorem 3.5 as well.

We now give a second proof of Theorem 3.5, which gives another proof of Theorem 3.2 as well. In fact, the proof of Theorem 3.5 that we give below does not use the fact that X i , Y j , and Z are finite dimensional, and so applies to normed vector spaces (even over the complex numbers) that need not be finite dimensional. Another reason for giving a second proof of Theorem 3.5 is to note how the second proof is conceptual and illustrates where the linearity of the derivative is used, while the first proof is computational and hides some of the ideas of the proof.

The following definitions will be convenient in the proof.

Definition 4.1. If h X i , then let h i denote the vector in X 1 ×× X n whose components are all 0 except the i -component, which is h ; so h i is the image of h under the natural inclusion X i X 1 ×× X n , and h= i h i . Let a X 1 ×× X n . If W is a vector space and ψ: X 1 ×× X n W is a linear map, then define D i ψ( a ): X 1 W by

D i ψ( a )( h )=Dψ( a )( h i ); (11)

it is a linear map.

Note that the definition of D i ψ( a ) above is consistent with Definition 3.4: when W= Y j and ψ= f j , we get D i ψ( a )= D i f j ( a ) .

Let h X i . Then

D i ( gf ) ( a )( h )=D ( gf ) ( a )( h i )( byequation( 11 ) ) =( D g ( f( a ) )Df( a ) )( h i )( bytheusualchainrule( 7 ) ) =D g ( f( a ) )( Df( a )( h i ) ) (12)

Now the j -th component of Df( a )( h i ) is D f j ( a )( h i ) , so

Df( a )( h i )= j ( D f j ( a )( h i ) ) j (by the first sentence in Definition 4.1)

Putting this in Equation (12) above, we get

D i ( gf ) ( a )( h )=D g ( f( a ) )( j ( D f j ( a )( h i ) ) j ) = j D g ( f( a ) ) ( D f j ( a )( h i ) ) j ( since the derivative is linear ) = j D j g ( f( a ) )( D f j ( a )( h i ) )( by equation( 11 ) ) = j D j g ( f( a ) )( D i f j ( a )( h ) )( by equation( 11 )again ) = j ( D j g ( f( a ) ) D i f j ( a ) )( h ) =( j D j g ( f( a ) ) D i f j ( a ) )( h )

Since h was arbitrary, this shows that D i ( gf ) ( a )= j D j g ( f( a ) ) D i f j ( a ) , as needed.

Acknowledgements

We are grateful to R. Vijay Krishna for several useful suggestions on this article.

NOTES

1As defined on page 42 in ([2], Section 5), for example.

2We are not using X i , etc., to denote the manifolds, since the X i ’s correspond to local charts on the manifolds.

Conflicts of Interest

The author declares no conflicts of interest regarding the publication of this paper.

References

[1] Spivak, M. (1965) Calculus on Manifolds. A Modern Approach to Classical Theorems of Advanced Calculus. W. A. Benjamin, Inc.
[2] Munkres, J.R. (1991) Analysis on Manifolds. Addison-Wesley Publishing Company, Advanced Book Program.
[3] Coleman, R. (2012) Calculus on Normed Vector Spaces. Universitext, Springer.
[4] Dieudonné, J. (1969) Foundations of Modern Analysis, Vol. 10-I of Pure and Applied Mathematics. Academic Press.

Copyright © 2025 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.