Received 27 April 2014; accepted 4 December 2015; published 7 December 2015

1. Introduction
In this paper, we find a formula depending on determinants for the solutions of the following linear equation
(1)
or
(2)
Now, if we define the column vectors

then the system (2) also can be written as follows:
(3)
where
denotes the innerproduct in
and A is
real matrix. Usually, one can apply Gauss Elimination Method to find some solutions of this system, and this method is a systematic procedure for solving systems like (1); it is based on the idea of reducing the augmented matrix
(4)
to the form that is simple enough such that the system of equations can be solved by inspection. But, to my knowledge, in general there is not formula for the solutions of (1) in terms of determinants if
.
When
and
, the system (1) admits only one solution given by
, and from here one can deduce the well known Cramer Rule which says:
Theorem 1.1. (Cramer Rule 1704-1752) If A is
matrix with
, then the solution of the system (1) is given by the formula:
(5)
where
is the matrix obtained by replacing the entries in the ith column of A by the entries in the matrix
![]()
A simple and interested generalization of Cramer Rule is done by Prof. Dr. Sylvan Burgstahler ( [1] ) from University of Minnesota, Duluth, where he taught for 20 years. This result is given by the following Theorem:
Theorem 1.2. (Burgstahler 1983) If the system of equations
(6)
has(unique) solution
, then for all
one has
. (7)
Using Moore-Penrose Inverse Formula and Cramer’s Rule, one can prove the following Theorem. But, for better understanding of the reader, we will include here a direct proof of it.
Theorem 1.3. For all
, the system (1) is solvable if, and only if,
(8)
Moreover, one solution for this equation is given by the following formula:
(9)
where
is the transpose of A (or the conjugate transpose of A in the complex case).
Also, this solution coincides with the Cramer formula when
. In fact, this formula is given as follows:
(10)
where
is the matrix obtained by replacing the entries in the jth column of
by the entries in the matrix
![]()
In addition, this solution has minimum norm, i.e.,
(11)
and
.
The main results of this work are the following Theorems.
Theorem 1.4. The solutions of (1)-(3) given by (9) can be written as follows:
(12)
Theorem 1.5. The system (1) is solvable for each
, if, and only if, the set of vectors
formed by the rows of the matrix A is lineally independent in
.
Moreover, a solution for the system (1) is given by the following formula:
(13)
(14)
where the set of vectors
is obtain by the Gram-Schmidt process and the numbers
are given by
(15)
and
.
2. Proof of the Main Theorems
In this section we shall prove Theorems 1.3, 1.4, 1.5 and more. To this end, we shall denote by
the Euclidian innerproduct in
and the associated norm by
. Also, we shall use some ideas from [2] and the following result from [3] , pp 55.
Lemma 2.1. Let W and Z be Hilbert space,
and
the adjoint operator, then the following statements holds,
[(i)]
such that
![]()
[(ii)]
.
We will include here a direct proof of Theorem 1.3 just for better understanding of the reader.
Proof of Theorem 1.3. The matrix A may also viewed as a linear operator
; therefore
and its adjoint operator
is the transpose of A and
.
Then, system (1) is solvable for all
if, and only if, the operator A is surjective. Hence, from the Lemma 2.1 there exists
such that
![]()
Therefore,
![]()
This implies that
is one to one. Since
is a
matrix, then
.
Suppose now that
. Then
exists and given
we can see that
is a solution of
.
Now, since
is the only solution of the equation
![]()
then from Theorem 1.1 (Cramer Rule) we obtain that:
![]()
where
is the matrix obtained by replacing the entries in the ith column of
by the entries in the matrix
![]()
Then, the solution
of (1) can be written as follows
![]()
Now, we shall see that this solution has minimum norm. In fact, consider w in
such that
and
![]()
On the other hand,
![]()
Hence,
.
Therefore,
, and
if
.
Proof of Theorem 1.5. Suppose the system is solvable for all
. Now, assume the existence of real numbers
such that
![]()
Then, there exists
such that
.
In other words,
![]()
Hence,
![]()
So,
![]()
Therefore,
, which prove the independence of
.
Now, suppose that the set
is linearly independent in
. Using the Gram-Schmidt process we can find a set
of orthogonal vectors in
given by the formula:
. (16)
Then, system (1) will be equivalent to the following system:
, (17)
where
. (18)
If we denote the vectors
’s by
![]()
and the
matrix
by
![]()
then, applying Theorem 1.3 we obtain that system (17) has solution for all
if, and only if,
. But,
.
So,
![]()
From here and using the formula (9) we complete the proof of this Theorem.
Examples and Particular Cases
In this section we shall consider some particular cases and examples to illustrate the results of this work.
Example 2.1. Consider the following particular case of system (1)
(19)
In this case
and
. Then, if we define the column vector
![]()
![]()
Then,
and
![]()
Therefore, a solution of the system (19) is given by:
(20)
Example 2.2. Consider the following particular case of system (1)
(21)
In this case
and
![]()
Then, if we define the column vectors
![]()
then
![]()
Hence, from the formula (10) we obtain that:
.
Therefore, a solution of the system (21) is given by:
(22)
(23)
(24)
. (25)
Now, we shall apply the foregoing formula or (12) to find the solution of the following system
. (26)
If we define the column vectors
![]()
then
and
,
, and
.
Example 2.3. Consider the following general case of system (1)
. (27)
Then, if
is an orthogonal set in
, we get
![]()
and the solution of the system (1) is very simple and given by:
(28)
Now, we shall apply the formula (28) or (12) to find solution of the following system:
. (29)
If we define the column vectors
![]()
Then,
is an orthogonal set in
and the solution of this system is given by:
,
,
and
.
3. Variational Method to Obtain Solutions
Theorems 1.3, 1.4 and 1.5 give a formula for one solution of the system (1) which has minimum norma. But it is not the only way allowing to build solutions of this equation. Next, we shall present a variational method to obtain solutions of (1) as a minimum of the quadratic functional
,
(30)
Proposition 3.1. For a given
the Equation (1) has a solution
if, and only if,
(31)
It is easy to see that (31) is in fact an optimality condition for the critical points of the quadratic functional j define above.
Lemma 3.1. Suppose the quadratic functional j has a minimizer
. Then,
(32)
is a solution of (1).
Proof. First, we observe that j has the following form:
![]()
Then, if
is a point where j achieves its minimum value, we obtain that:
![]()
So,
and
is a solution of (1).
Remark 3.1. Under the condition of Theorem 1.3, the solution given by the formulas (32) and (9) coincide.
Theorem 3.1. The system (1) is solvable if, and only if, the quadratic functional j defined by (30) has a minimum for all
.
Proof. Suppose (8) is solvable. Then, the matrix A viewed as an operator from
to
is surjective. Hence, from Lemma 2.1. there exists
such that
![]()
Then,
![]()
Therefore,
![]()
Consequently, j is coercive and the existence of a minimum is ensured.
The other way of the proof follows as in proposition 3.1.
Now, we shall consider an example where Theorems 1.3, 1.4 and 1.5 can not be applied, but proposition 3.1 does.
Example 3.1. It considers the system with linearly independent rows
.
In this case
and
![]()
Then
![]()
Therefore, the critical points of the quadratic functional j given by (30) satisfy the equation:
![]()
i.e.,
.
So, there are infinitely many critical points given by
![]()
Hence, a solution of the system is given by
.