1. Introduction
Function data are infinite dimensional vectors in functional space, the study of functional data helps people to further understand data changes in finance, medicine, etc. The study of fucntional response can date back to the work Ramsay and Silverman (2005) [1] where the model the functional concurrent model considers current response relate to the current values of the covariates. Wong et al. (2018) [2] investigated a class of partially linear functional additive models (PLFAM) that predicts a scalar response by both parametric effects of a multivariate predictor and nonparametric effects of a multivariate functional predictor. Further an additive function-on-function regression is established using principal component basis functions and b-spline basis functions by Janet S. Kim et al. (2016) [3]. Luo, R., Qi, X. (2017) [4] consider functional linear regression models with a functional response and multiple functional predictors, with the goal of finding the best finite-dimensional approximation to the signal part of the response function. This paper extends function on function regression, considering functional partially linear additive models that predict a functional response by a scalar predictor and functional predictors. Specifically we consider the model
(1)
where
is the response function defined in an interval
, For convenience, we assume that the response has zero-mean, the predictive curves.
are defined in
. where
and
are compact intervals.
is scalar predictor defined in an interval
.
is scalar coefficient,
is the noise function with mean zero and unknown autocovariance function
and is independent of the predictor and the response. In the previous papers, these focused on function-on-function, scalar-on-function, and scalar-mixed data. There is no paper devoted to function-on-mixed data. That is to say, the functional dependent variable analyzes the regression model of the mixed data. This paper considers the actual situation is closer to the actual demand. Considering that the real world data cannot be all the functional data, there will definitely be a part of the scalar in the model. From this perspective, expand the new model. This paper uses a different estimation method from the previous. The general method adopted by these articles is to use non-parametric methods such as spline function to approximate the model generation estimation parameters, and at the same time minimize the objective function. In the past, the model was generally complicated and inconvenient to calculate. They consider fitting with B-splines in three directions at the same time, and there will be over-fitting when doing so. In this paper, the problem is optimized by using the eigenbasis and orthogonal B-spline, that is, using the special feature base of
and an orthogonal B-spline to fit the model. While ensuring the fitting effect, the complexity of the model is reduced, and the model is not over-fitting. We use the classical formula of probability to express the error term of the model from the perspective of variance, and the method of resampling is used to approximate the error of the model. At the same time, assume from the article hypothesis, we establish the asymptotic convergence rate for estimator.
2. Model
In this section, we consider the estimation of coefficients in a partial linear regression model of a function. In this paper, the model is estimated by using the eigenbasis spline and the B-spline function. Note
is the corresponding eigenfunctions of
, and
(2)
Let
(3)
we project
onto the eigenbasis of
,
(4)
then we can express
as
(5)
For the function coefficient term part of (1), we estimate it by means of B-spline and feature base. For convenience we write model (1) as
(6)
where
is the p-dimensional vector of predictive functions,
is the p-dimensional vector of coefficient functions. Let
(7)
Transform
in the same way as
(8)
So we can put
similar expression of Karhunen-Loeve expansion
(9)
In this paper, on spectral decomposition method projecting the corresponding function to the orthogonal eigenbasis of y, we estimate the parameter of the model can be performed without punished the complexity of the t direction. Generally, we preset the maximum variance number similar to the method in the functional principal component analysis, usually 85%. Next we will use the Orthogonal b-spline method (2010) [5] [6] to represent the parameter items in the s direction.
(10)
(11)
where
are or-thogonalized B-spline bases of dimensions M,
is unknown parameters, we get
(12)
In the t direction perform truncation K for the preset maximum variance, we get
(13)
We denote
,
as
, let
.
Then we can get the simplified expression of
(14)
3. Parameter Estimation
We assume that with probability 1, the trajectory of
is contained in a Hilbert space
, with inner product. We will focus on the case that
’s are
functional spaces and the inner products are
for any
. We use quadratic penalties for the direction s and control the roughness in the direction t by the preset number of orthogonal basis functions. so the loss function is as follow
(15)
We use the least squares penalty to punish the curvature of the functional parameter, considering the penalty term
(16)
(17)
where
. Then we can get the penalized criterion
(18)
Similarly
and
to process
. We have Karhunen-Loeve expansion on y(t)
(19)
where
is zero-mean error, and the variables
are the FPCA (Functional principal component) scores of Y. And note that
, so the criterion can be simplify written as
(20)
So the estimation process for the parameters
,
is as follows.
1) We get the function curve
for smooth sample points, and we demean
for get a curve
with a mean of 0. The point K is determined by the pre-specified percentage of variance.
2) We use FPCA(Functional Principal Component) to estimate eigenbasis
, and we project
onto the corresponding eigenbasis
to obtain an estimation of the principal component score
.
3) We get a two-step estimation for
,
a) Obtain estimation of
,
, of the coefficients by minimizing the penalized criterion(19) with respect to
and
, The estimated value we get for
is
(21)
b) Obtain estimation of
by minimizing
(22)
So the estimation of
is
and
is
. We can get
(23)
4. Error Variance Decomposition
In this section, we get error variance decomposition. Let
be the variance function of
.we estimate
.
1) obtain residuals by fit (14),
.
2) Use FPCA for the residual function, to approximate the infinite covariance function using the estimated finite covariance function.
Let
,
be new observation independent of X and
, and fit function-on-partially linear functional additive models,then we measure the uncertainty in the prediction by predicting error
.Base on Ruppert et al. (2009) [7]
(24)
We estimate the expression of
by estimating
Then we estimate
using the classical variance formula
(25)
In this paper, we assume that
is a known parameter set containing estimated arbitrary parameter values and corresponding parameter variance and covariance functions. So we will expand
of the formula (24) as follows
(26)
If we write
as
,
is recorded as
, then
and
. So we expand the expression (25) as follow
(27)
It can be seen that all the variances and covariance functions in the expression are represented by the variance and covariance functions of the existing estimates. Then we focus on the second item of expression (24), the second term is
. This paper hopes to obtain an approximation form of the finite term of the variance function under finite samples by resampling the bootstrap method. We perform the bootstrap algorithm through the following steps;
1) For
.
2) Resample the samples to get Q groups different sample.
3) Smooth the resample points and get the corresponding curve
.
4) Obtain the de-mean
and the corresponding feature basis function
, and obtain the corresponding principal component score
.
5) Fit the function-on-partially linear functional additive models to obtain the corresponding parameter estimation.
6) Get prediction function
to any new observation curve
and calculate
.
7) Get
by expression (24).
Thus the approximate value of the finite term of
is found
. and the variance estimate of the error
(28)
5. Basic Assumptions and Convergence Properties
In this section,we study the asymptotic properties of
,
, where
(29)
(30)
5.1. Basic Assumption
1) For the square integrable random function
satisfies
;
2) For each k, there is
, where c is a constant and c can take different values in different expressions;
3) For the eigenvalue
, satisfied
;
4) For the coefficient
, assume there is a constant t, and
, which make
satisfied
;
5) For the parameter K, assume
6) The number of knots
, with
.
7) The distribution of
and
is absolutely continuous and its density is bounded on
;
8) The joint density function of
is bounded on
;
9) Covariance function
of
is bounded on interval
Conditions (1)-(5) are common in functional PCA literature and conditions (6)-(9) are common in nonparametric regression literature.
5.2. Related Lemma
Lemma 1. If
is spline expansion of
, it is known by Chen [8] that there is
, we can get
(31)
5.3. Convergence Properties
Consider the model (6)
(32)
Use the principal component of
for the model (6) to represent the following expression
(33)
where
and
. We can get an estimation by least squares estimation of
, let
, where
is spline expansion of
.
Theorem 2. Under the conditions of (1)-(5), we can get
(34)
Proof. Use Cauchy-Buniakowsky-Schwarz Inequality we can get
(35)
Assume from the article hypothesis, we have
(36)
☐
Theorem 3. Available under the conditions of (6)-(9), and let
is spline expansion of
we can get
(37)
Proof.
(38)
Consider the latter item in the text, because
is a spline expansion of
, by lemma we have
(39)
Now consider the former item
(40)
Then consider
(41)
Since
,
(42)
Get
by the above formula.☐
6. Conclusion
This paper is a model extension of the function-to-function regression. The scalar variable is added to the dependent variable, which extends the application scope of the model. For real-life data, real data should include scalar data and function data, and the model used in this paper can be better explained. In this paper, the model is estimated by using the eigenbasis spline function and the orthogonal B-spline function. When the loss function is punished, the complexity of the t direction is controlled by controlling the preset principal component in the t direction, which reduces the complexity. It is more practical. At the same time, the variance decomposition of the error under the finite sample size is given. The approximation is performed by the resampling bootstrap method. Finally, the convergence properties of the estimated parameters are studied.