Robustness of Variable Selection Based on Functional Response Regression Model

Abstract

Function-on-scalar regression is a type of function response regression used to analyze the relationship between function response and a set of scalar predictor factors. The variable selection methods of FOSR models mostly focus on the linear effects of scalar predictor factors. Therefore, in this paper, we perform robust variable selection for nonlinear FOSR models with the presence of multiple continuous covariates to further explain the behavior of function response over time. We project functional data into a low dimensional principal component space via the principal component score of the function response, in order to use the principal component score for variable selection and regression modeling. In this paper, we develop a regularized iterative algorithm based on exponential squared loss and group smoothly clipped absolute for predicting estimates of scalar factors and function coefficients, and use tuning parameters chosen by data-driven methods to achieve robustness in variable selection. The robustness of the proposed method is verified through simulation studies and demonstrated on real datasets.

Share and Cite:

Yan, Z.Q. and Ye, W.Z. (2025) Robustness of Variable Selection Based on Functional Response Regression Model. Advances in Pure Mathematics, 15, 220-234. doi: 10.4236/apm.2025.153010.

1. Introduction

With the advancement of modern technology, more and more data are recorded continuously at intervals of time, or intermittently at several discrete points in time, often stored in the form of curves and images. Functional data are essentially infinite dimensional its basic unit is the function. A common type of functional data are curves recorded during a time interval. Functional Data Analysis (FDA) includes statistical methods for such data. FDA deals with the analysis of data which is in the form of functions. The functions could be curves, images, or other types of forms, depending on the set on which the functions are recorded. Functional data analysis is mainly divided into the following three regression models: 1) scalar-on-function regression: a continuous response variable regressed on functional covariates; 2) function-on-function regression: a functional response regressed on functional predictors; 3) Function-on-scalar regression: a functional response regressed on scalar predictors. The aim of this paper is to provide robust variable selection for function-on-scalar regression models containing nonlinear scalar predictors.

Function-on-scalar regression (FOSR) is an increasingly popular area of research in the analysis of functional data, e.g. Barber et al. [1] for genome-wide association studies (GWAS), Chen et al. [2] for the effect of stroke severity on motor control, Goldsmith et al. [3] for physical activity (PA), and so on. In FOSR, a continuous functional response is typically modeled by the additive effects of a set of scalar predictors of interest, which are captured using smooth univariate coefficient functions. This has led to several methods of variable selection in the FOSR model. Chen et al. [2] used a pre-whitening approach to account for temporal correlation within curves and proposed a variable selection procedure via group minima and maxima concave penalties (MCP) regularization with a penalized least squares technique. Barber et al. [1] used a penalized least squares objective function with the LASSO penalty on the basis coefficients and proposed a function-on-scalar LASSO (FSL). Fan and Reimherr [4] further developed an adaptive FSL in high dimensions. Parodi and Reimherr [5] proposed a functional linear adaptive mixed estimation method to achieve variable selection and smoothing for function-on-scalar regression.

The vast majority of FOSR methods focus on linear effects of scalar predictors, but this may not necessarily hold true in many applications. Ghosal and Maity [6] proposed a nonlinear function-on-scalar regression approach considering the additional nonparametric effects of the scalar predictors into the FOSR model. This paper is based on this nonlinear model for robust variable selection. It is well known that the least squares method is sensitive to outliers in the data, however, the variable selection methods discussed above are based on the least squares criterion. Therefore, it is necessary to propose a more robust variable selection method to replace the least squares method in the presence of outliers. Many loss functions are resistant to outliers, such as minimum absolute deviation loss, quantile loss, or Huber loss, which can be used for robust estimation and variable selection. In this paper, we introduce the exponential squared loss (ESL) introduced by Wang et al. [7] in nonlinear FOSR to reduce the effect of outliers. ESL is a bounded continuous function, and we can control the robustness of the estimator by selecting the appropriate adjustment parameter h , that is, reducing the influence of outliers that lead to large errors. The ESL function is defined as ϕ h ( t )=1exp( t 2 /h ) . When h is very large, ϕ h ( x ) x 2 /h , and therefore the proposed estimator is similar to the least squares estimator in the extreme case. When h is small, observations with larger absolute values will have a small impact on the loss function ϕ h ( x ) and therefore have a smaller impact on the estimator.

In this paper, we investigate the use of functional principal component analysis (FPCA) to project the functional response into a low-dimensional principal component space in a nonlinear FOSR model. The principal component scores are used as target variables in a regression model to analyze the effect of scalar covariates on the functional response and perform robust variable selection for the nonlinear covariates. The nonlinear covariates are approximated by a B-spline basis function, and the estimation of the coefficient function combines the FPCA eigenfunctions and the B-spline basis function. We integrate the exponential squared loss function into the nonlinear FOSR model and propose a robust variable selection process using the group SCAD regularization method. By locally linearly approximating the SCAD penalty function, it is replaced by a set of weighted LASSO penalty functions. We utilize a data-driven approach to select the smoothing parameters, which achieves robustness and establishes consistency of the proposed estimators. The estimated coefficients are projected back using univariate basis functions and eigenfunctions of the functional response to recover the effects of the scalar predictors. Finally, we show through simulations and real data applications that the proposed method performs well. When outliers are present in the dataset, our method is robust to outliers. Meanwhile, in the absence of outliers, the proposed estimation method is comparable to the least squares estimation method.

The rest of this paper is organized as follows. In Section 2, we describe the estimation methods for nonlinear function-on-scalar regression and discuss the theoretical properties of the proposed estimators. In Section 3, we present the implementation algorithm for computing the estimator based on nonlinear FOSR and the criterion used to select the smoothing parameter. To demonstrate the superiority of the proposed method, in Section 4, we report and compare the simulation results and illustrate the proposed method with a real diffusion tensor imaging data example. Finally, in Section 5, the summary and conclusion of the paper are given in Section 5.

2. Estimation Method and Theoretical Properties

2.1. Estimation Method

Suppose that the observed data for the ith subject is { Y i ( t ), M i1 , M i2 ,, M iq } ( i=1,2,,n ) where Y i ( t ) represents the functional response and M i1 , M i2 ,, M iq are the corresponding scalar predictors of interest. Sometimes there might be additional control or confounding variables X i1 , X i2 ,, X ip which we want to adjust for. In this paper, we assume the functions are observed on a dense and regular grid of points S={ t 1 , t 2 ,, t m }T=[ a,b ] for some a,bR where the functional response is observed on an irregular and sparse domain. Now, the linear FOSR model is extended to the nonlinear dynamic effect of scalar predictors, and we use the following generalized function on scalar regression (GFOSR) model proposed by Ghosal and Maity [6]:

Y i ( t )=μ( t )+ j=1 p X ij β j ( t ) + l=1 q θ l ( M il ,t ) + ε i ( t ), (1)

Here, β j ( t ) represents the functional effect of the scalar predictor X ij , and θ l ( , ) is an unknown smooth function on R×T (twice differentiable with respect to both parameters), used to capture the functional effect of the predictor M il . The coefficient functions β j ( ) , θ l ( , ) , functional response Y i ( ) , and the error process ε i ( ) are assumed to lie in a real separable Hilbert space [1]. In this paper, we focus our attention to L 2 ( T ) . We assume that the covariates are centered and μ( ) is a smooth function capturing marginal mean of the functional response Y i ( ) , where the error functions ε i ( ) are i.i.d. of ε i ( ) which is a mean zero stochastic process with unknown nontrivial covariance structure.

Let Y 1 ( t ),, Y n ( t ) be independent realizations of a smooth random function Y( t ) with mean function μ( t ) and covariance function cov{ Y( t ),Y( s ) }= C( t,s ) . By the Karhunen-Loève theorem by Ash and Gardner [8], the random

function Y i ( t )=μ( t )+ k=1 ξ ik ϕ k ( t ) where μ( t )=E{ Y i ( t ) } is the overall mean

function; ϕ k ( t ) is the kth orthonormal eigenfunction of the covariance function, satisfying ϕ k ( t ) ϕ j ( t )dt =1 if j=k and 0 otherwise; ξ ik denotes FPC scores with E( ξ ik )=0 , var( ξ ik )= ρ k and cov( ξ ij , ξ ik )=0 if jk , and ρ k is the eigenvalue corresponding to the eigenfunction ϕ k ( t ) . Since

sup t[ 0,1 ] E { k=1 ξ ik ϕ k ( t ) k=1 K n ξ ik ϕ k ( t ) } 2 0 , one can reasonably suppose that

Y i ( t )μ( t )+ k=1 K n ξ ik ϕ k ( t ) (2)

as K n . The approximation of (2) with a fixed K n is commonly adopted for longitudinal and functional data analysis (see Yao et al. [9]). To be more flexible, Hall and Mohammad [10] considered model (2) with K n as n . In this paper, we adopt the FPCA method of Yao et al. [9], which is used to estimate the FPCA score ξ ik , the mean function μ( t ) , and the eigenfunction ϕ k ( t ) . Specifically, a spectral decomposition of the covariance function yields μ( t ) and ϕ k ( t ) , and thus ξ ik is estimated by numerical integration. Thus, the model in (1) becomes:

ξ ik = T ( Y i ( t )μ( t ) ) ϕ k ( t )dt = j=1 p X ij T β j ( t ) ϕ k ( t )dt + l=1 q T θ l ( M il ,t ) ϕ k ( t )dt + T ε i ( t ) ϕ k ( t )dt = j=1 p X ij β jk + l=1 q θ lk ( M il ) + ε ik (3)

for k=1,,K , where we define β jk = s β j ( t ) ϕ k ( t )dt and

θ lk ( u )= T θ l ( u,t ) ϕ k ( t )dt .

We project the nonlinear variable M il onto the B-spline basis function space, and expand the nonparametric function θ l ( ) with B-spline basis functions as

θ lk ( u )= T θ l ( u,t ) ϕ k ( t )dt . Let B j ( t )= ( B j ( t ):1j J n ) T denote the qth-order

B-spline basis functions, where J n = N n +q , and N n is the number of interior knots for a knot sequence α 1 ==0= α q < α q+1 << α N n +q <1= α N n +q+1 == α N n +2q . So, model (3) becomes as follows:

ξ ik = X i T β K + l=1 q j=1 J n B j ( M il ) δ l,k,j + ε ik (4)

where we define X i T =( X i1 ,, X ip ) , β k T =( β 1k ,, β pk ) , δ k T =( δ 1,k,1 ,, δ 1,k, J n ,, δ q,k, J n ) , β k T =( β 1k ,, β pk ) , and

η i T =( B 1 ( M i1 ),, B 1 ( M iq ), B 2 ( M i1 ),, B J n ( M iq ) ) . Let Z i [ U i , V i ] dimension

of K×( Kp+Kq J n ) , γ=( β,δ ) , where U i = I K×K X i T dimension of K×( Kp ) , V i = I K×K η i T dimension of K×( Kq J n ) , δ= ( δ 1 T ,, δ k T ) T , ξ 1 =( ξ i1 , ξ i2 ,, ξ ik ) , and ε 1 =( ε i1 , ε i2 ,, ε ik ) . Thus, we can write the FOSR model as,

ξ i = Z i γ+ ε i . (5)

Then, the robust estimator of γ=( β,δ ) can be obtained by minimizing

Q( γ )= i=1 n ϕ h ( ξ i Z i γ ), (6)

where ϕ h ( ) is the ESL function. We select variables for all scalar predictors, including linear and nonlinear variables, and group them. We group the variables γ to be selected into, γ=( γ 1 ,, γ l+1 ) where each γ i represents a group of variables. Specifically, when l=0 , γ 1 =( β 1 ,, β p ) , the effect of all control or confounding variables X j ,j=1,,p , is divided into a group that is either simultaneously selected or simultaneously 0. When l=1,,q , γ l+1 =( δ l,1,1 ,, δ l,k, J n ) , l=0,1,,q each nonlinear predictor M l is divided into a group, with the projection coefficients under the B-spline basis forming a group for variable selection. We use the group penalty function p λ ( γ l 2 ) for group variable selection, where

γ l 2 = j=1 d γ l,j 2 , where d is the number of variables within theth group. In this

paper, we use the group smoothly clipped absolute deviation (SCAD) regularization method, thus proposing to minimizing:

L( γ )= i=1 n ϕ h ( ξ i Z i γ )+n l=0 q p λ ( γ l ) (7)

The SCAD penalty of Fan and Li [11] is defined in the following way,

p λ ( θ )={ λ| θ |,| θ |λ, θ 2 2aλ| θ |+ λ 2 2( a1 ) ,λ<| θ |aλ, ( a+1 ) λ 2 2 ,| θ |>aλ.

for some a>2 . Where γ j is the Euclidean norm of γ j , and p λ ( γ j ) is the penalty function with λ as a regularization parameter. When the loss function ϕ h ( x )= x 2 is the least square loss, it is the least squares (LS) method proposed by Wang, et al. [12]. When the loss function ϕ h ( x )=| x | is the least absolute deviation loss, there is a LAD estimation method.

2.2. Theoretical Properties

In this subsection, we discuss the theoretical properties of the proposed estimators. We apply local linear smoothing to the mean function μ . Let C ^ ( s,t ) be the estimated smooth surface of C( t,s )=cov{ Y( t ),Y( s ) } obtained through local quadratic fitting, and h μ and h G be the bandwidths of the estimated values μ ^ and G ^ , respectively. We use the symbol to denote the Euclidean norm for vectors and the L 2 ( T ) norm for functions defined on T . Let f and f present the first and second derivatives of f , respectively. We impose the following regularity conditions:

(C1) The mean function μ( t ) and the covariance function G( s,t ) are

smooth, and the estimation error satisfies: sup tT | μ ^ μ( t ) |= O p ( 1/ n h μ ) , sup tT | G ^ ( s,t )G( s,t ) |= O p ( 1/ n h G 2 ) , thus FPCA estimation characteristic function

ϕ ^ k ( t ) converge to the real characteristic function:

sup| ϕ ^ k ( t ) ϕ k ( t ) |= O p ( 1/ n h μ ) .

(C2) Error function ε( t ) is a finite variance with a mean of zero, satisfy the cov( ε i ( t ), ε j ( t ) )=0 , ji , so that the FPCA score ξ ik satisfies consistency: ξ ^ ik p ξ ik .

(C3) The matrix E( Z Z T ) is positively definite and the eigenvalues are bounded, where the dimension of Z is fixed. There exists a positive constant M such that | Z j |M , for all 1jp .

(C4) The functional coefficient δ l ( ) C r [ 0,1 ] for some integer r2 , and the spline order q satisfies qr . C r [ 0,1 ]={ φ| φ ( r ) C[ 0,1 ] } means the space of r -th order smooth function.

(C5) The distance between neighboring knots H i = τ i+1 τ i , and H= max 1i1+ N n H i . For some constant 0<C< such that

H/ min 1i1+ N n H i C and max 1i N n { H i+1 H i }=o( N n 1 ) .

(C6) E[ ϕ h ( ε( t ) ) ]=0 , and E[ ϕ h ( ε( t ) ) ]>0 , for all t and any h>0 .

Remark 1 The conditions (C1) and (C2) are the consistency and asymptotic results derived by Yao et al. [9] for estimating the FPC score ξ ^ ik . Therefore, in this paper, the function FPCA scores ξ ik can be used as the target variable, and the estimated coefficients can be projected back using the univariate basis functions and the eigenfunctions of the functional response to recover the influence of the scalar predictor. The conditions (C3)-(C5) are standard in spline estimation literature, where the number of nodes J n for B-splines should satisfy

lim n J n / n 1/ ( 2n+1 ) =C , where C is some nonzero constant. Condition (C6) is introduced by Yao et al. [13] as an additional tuning parameter to automatically

select based on observed data, ensuring that the objective function has a local minimum.

Theorem 1 Suppose that the conditions (C1)-(C6) are satisfied, if λ n and n r/ ( 2r+1 ) λ n as n , then we have

γ ^ i ( t ) γ i ( t ) = O p ( n r/ ( 2r+1 ) ),j=1,,p+q.

Remark 2 Based on the convergence rate in Theorem 1, we can see that our proposed method provides the optimal rate of convergence O p ( n r/ ( 2r+1 ) ) established by Stone [14] for estimating the functional coefficients. When r=2 , the optimal convergence rate becomes O p ( n 2/5 ) . In the studies of Fan and Li [11] and Wang et al. [7], similar variable selection methods have been proven to be consistent and asymptotically normal. Therefore, the asymptotic property of our method still holds if the above conditions are satisfied.

3. Computational Algorithm and Tuning Parameter Selection

3.1. Computational Algorithm

In this paper, we develop a new iterative algorithm using the local linear approximation (LLA) [15] to the group SCAD method and combining it with the exponential square loss function. The minimization problem of the objective function is solved through iterative reweighted least squares. Since the SCAD penalty function itself is non-convex, the LLA approximation is used to transform it into a weighted LASSO, which penalizes different groups of variables with different weights, and the weights are updated at each iteration, thereby enhancing the robustness of variable selection. The ESL function is applied to the residuals to reduce the effect of outliers on the objective function.

In the LLA method, we perform a local linear approximation on the SCAD

penalty function: p λ ( γ l 2 ) p λ ( γ l ( m ) 2 ) γ l 2 ,where:

w l = p λ ( γ l ( m ) 2 )={ λ,if γ l ( m ) 2 λ, aλ| γ | a1 ,ifλ< γ l ( m ) 2 aλ, 0,if γ l ( m ) 2 >aλ,

Thus, the SCAD penalty function is approximated as a weighted group LASSO:

l=0 q p λ ( γ l 2 ) l=0 q w l γ l 2 where w l is the weight updated at each iteration, let W SCAD =diag( w 1 γ l 2 ,, w l γ l 2 ) . Then our objective Function (7) becomes:

L( γ )= i=1 n ϕ h ( ξ i Z i γ )+n l=0 q w l γ l 2 (8)

Derive this equation with respect to γ and set it to 0, resulting in the following equation:

i=1 n Z i T ϕ h ( ξ i Z i γ )+n W SCAD γ=0. (9)

Under ESL loss, let the weight matrix W ESL be calculated from the derivatives

of ESL: w i = ϕ h ( r i ) r i = 2 h exp( r i 2 h ) , where r i = ξ i Z i T γ , i=1,,n . With

W ESL =diag( w 1 ,, w n ) . Then Equation (9) becomes:

( Z T W ESL Z+n W SCAD )γ= Z T W ESL ξ (10)

Update γ as in each iteration step:

γ ^ = ( Z T W ESL Z+n W SCAD ) 1 Z T W ESL ξ.

Therefore, the iterative calculation process for γ ^ = ( γ ^ 0 T ,, γ ^ q T ) T is as follows:

Step 1: Obtain initial values using least squares estimation γ ^ ( 0 ) = ( Z T Z ) 1 Z T ξ .

Step 2: Given γ ^ ( m ) : Calculate the residual weight matrix W ^ ( m ) , where the residual is γ ^ i ( m ) = ξ i Z i T γ ^ ( r ) .

Step 3: Given γ ^ ( m ) , calculate the LLA approximation weights for SCAD w ^ ( m ) .

Step 4: Update γ ^ ( m+1 ) through iterative reweighted least squares: γ ^ ( m+1 ) = ( Z T W ESL Z+n W SCAD ) 1 Z T W ESL ξ .

Step 5: Repeat steps 2 to 4 until convergence. If γ ( m+1 ) γ ( m ) is less than a predetermined threshold, the final estimate of γ is obtained.

3.2. The Choice of Tuning Parameters

The parameter K is the number of eigenbases in the function principal component analysis, and the parameter J is the number of B-spline bases used to approximate the nonlinear parameters as a linear model. To select the number of these two basis functions more robustly, we use the following weighted generalized cross-validation (GCV) statistic:

GCV( K,J )= ( ξZη ) T W ( YZη )/ nk ( 1 t( K,J )/ nk ) 2 ,

where t( K,J )=trace{ Z ( Z T W ^ Z ) 1 Z T W } . Thus, the optimal number of K and

J is obtained by ( K ^ , J ^ )=minGCV( K,J ) . In order to achieve consistency in variable selection, we use the extended Bayesian information criterion (EBIC) [16] of a corresponding Gaussian likelihood for choosing the penalty parameter λ . For the parameter a in the SCAD penalty function, following the suggestion of Fan and Li [11], we adopt a=3.7 for implementation. The tuning parameter h impacts the efficiency of the proposed estimator. To attain high efficiency, we adopt the approach of Yao et al. [13] to select h .

4. Simulation and Application

4.1. Simulation Studies

In this subsection, we implement several simulation studies to investigate the finite sample performance of the proposed method. We take the sample sizes of n=50,100,200 and generate the data from the nonlinear FOSR model:

Y i ( t )=μ( t )+ j=1 2 X ij β j ( t )+ l=1 20 θ l ( M il ,t )+ ε i ( t ),

Here, we have the coefficient functions given by μ( t )=8sin( πt/ 50 ) , β 1 ( t )=3+ 5t/ 100 , β 2 ( t )=4sin( πt/ 50 )+4cos( πt/ 50 ) , the nonparametric functions are given by θ 1 ( x,t )=2 x 3 exp( t/ 100 ) , θ 2 ( x,t )=5( x+ x 3 )sin( πt/ 100 ) , θ 3 ( x,t )=16sin( xt/ 50 ) . The rest of the nonlinear functions θ l ( x,t ) are set to zero. The exogenous covariates X ij ~Unif( 2j,2i ) , and the predictors of interest M ij ~N( 0, 1 2 ) . The error process ε i ( t ) is generated as ε i ( t )= ξ i1 cos( t )+ ξ i2 sin( t )+N( 0, 1 2 I m ) , where ξ i1 ~N( 0, 0.5 2 ) and ξ i2 ~N( 0, 0.75 2 ) . The response Y i ( t ) is observed on a grid of m=80 . The values of K and J (number of B-spline basis) were chosen from a two-dimensional grid with K{ 3,4,5,6,7 } and J{ 5,7,9 } .

In order to examine the accuracy of the proposed method, we use the standard deviation of the integrated squared error (ISE), which is defined as

ISE= j=1 p T { η ^ j ( t ) η j ( t ) } 2 dt . Besides, we use the mean square prediction error

(MSPE) to assess the accuracy of prediction. MSPE is given by

MSPE= 1 N i=1 N T { Y ^ i * ( t ) j=1 p η j ( t ) Z ij * } 2 dt , where the independent samples

Z i =[ X i , M i ] are generated for the test set, Y ^ i * ( t )= j=1 2 X ij β j ( t )+ l=1 20 θ l ( M il ,t )

are estimated from the training data.

To demonstrate the effectiveness of the algorithm described in Section 4, we generated data with a sample size of n=100 and used cubic splines of order q=4 to approximate the nonlinear predictor variables. The tuning parameters K,J and h selected by the proposed process are approximately 5, 7, and 3.5, respectively.

In order to show the robustness of the proposed approach, we simulated three types of outlier contamination: outliers in the function response, outliers in the predictor variables, and outliers in both the function response and scalar predictor. To introduce outliers in the function response, we assumed that 10% of the response curves of the original sample were contaminated by outlier curves, denoted as Y ^ i * ( t ) , so that the contaminated function response data Y i 0 ( t )=( 1 R i ) Y i ( t )+ R i Y i * ( t ) is treated as a numerical simulation, where R i is a Bernoulli random variable such that P( R i =1 )=0.1 . To introduce outliers in the predictor variables, we randomly selected subsets of the predictors from the original sample and contaminated them with peaks values. The detail mechanisms for producing the outlier response Y i * ( t ) and outliers in scalar predictors are described as follows:

Scenario 1. No outliers. In this setting, Y i * ( t )= Y i ( t )

Scenario 2. When the outlier is in the function response. The outlier curve Y i * ( t ) is generated as Y i * ( t k )= Y i ( t k )+ r ik I [ U i , U i +v ] ( t k ),k=1,,m , where r ik are random values taken from a uniform distribution on [ r u , r l ][ r l , r u ] and I( ) is the indicator function, [ U i , U i +v ] be an interval in [ 0,1 ] for the ith subject, where U i is randomly chosen from [ 0,1v ] , and v is a constant. The r ik control the strength of outliers while the constant v determines the contamination length of each outlier curve. We set r l =3 , r u =10 and v=0.5 in the scenario.

Scenario 3. When the outliers are in the scalar predictors. We randomly select 5% of the predictors that are contaminated by the high leverage outliers ( X i1 * ,, X ip * )=( X i1 ,, X ip )+5 .

Scenario 4. When the outliers are both in the function response and in the scalar predictor. We consider the contaminated error distribution 0.8N( 0, 0.5 2 )+0.2Cauchy( 0,1 ) , which is used to produce outliers with a heavy-tailed error distribution for data set.

For each scenario, our method is compared with the other two methods. Since the method proposed in this paper is based on the exponential square loss function ϕ h ( t )=1exp( t 2 /h ) , our model is called ESL. When ϕ h ( x ) is the square loss function ϕ h ( x )= x 2 , it is the least squares (LS) method proposed by Wang et al. [12]. When ϕ h ( x )=| x | , it leads to the least absolute deviation method (LAD). For each scenario, 100 repetitions are used, and the numerical results are recorded in the following four tables. First, as shown in Table 1, when there are no outliers in the dataset and the error distribution is normal, the ISE and MSPE values of the ESL estimator are close to those of the LS estimator. This means that for datasets without outliers, the performance of these two methods is comparable. Moreover, the LS estimator performs best because its sample mean and the standard deviations of ISE and MSPE are smaller than those of the proposed ESL and LAD estimators. Second, from Tables 2-4, we can see that the ESL estimator produces smaller ISE and MSPE values than the LS and LAD estimators. This indicates that for datasets with outliers, the proposed ESL method outperforms the LS and LAD methods. This is mainly because when there are some very large outliers in the dataset, the proposed ESL method places more weight on the “most likely” data around the true value, thereby achieving a robust estimator. Finally, the ISE and MSPE values of the ESL estimator decrease with increasing sample size n , which confirms the consistency result established in Theorem 1. In summary, the proposed ESL method has certain advantages in variable selection and estimation.

4.2. Application

In this subsection, we apply the proposed method to analyze the real DTI dataset collected from the NIH Alzheimer’s Disease Neuroimaging Initiative (ADNI) study, which is available at http://adni.loni.usc.edu/. The DTI data includes 213 subjects. For each subject, we calculate the fractional anisotropy (FA) curve along the midsagittal skeleton of the corpus callosum at m = 83 positions, as shown in

Table 1. Simulation results of the averages (×102) and the standard deviations (×102) of integrated squared error (ISE) and mean square prediction error (MSPE) under Scenario 1.

n

Methods

ISE

MSPE

Aver.

std

Aver.

std

50

ESL

0.532

0.214

1.763

0.473

LS

0.438

0.186

1.516

0.432

LAD

0.732

0.278

1.928

0.675

100

ESL

0.273

0.106

0.316

0.127

LS

0.258

0.103

0.308

0.119

LAD

0.376

0.145

0.531

0.193

200

ESL

0.124

0.032

0.078

0.026

LS

0.121

0.032

0.076

0.025

LAD

0.183

0.054

1.117

0.041

Table 2. Simulation results of the averages (×102) and the standard deviations (×102) of integrated squared error (ISE) and mean square prediction error (MSPE) under Scenario 2.

n

Methods

ISE

MSPE

Aver.

std

Aver.

std

50

ESL

0.574

0.203

1.837

0.413

LS

3.486

2.693

13.723

7.516

LAD

0.904

0.523

3.841

1.967

100

ESL

0.241

0.098

0.564

0.139

LS

2.773

1.392

5.103

2.877

LAD

0.339

0.124

1.072

0.438

200

ESL

0.134

0.045

0..341

0.113

LS

1.747

1.219

2.914

1.485

LAD

0.202

0.074

0.517

0.149

Table 3. Simulation results of the averages (×102) and the standard deviations (×102) of integrated squared error (ISE) and mean square prediction error (MSPE) under Scenario 3.

n

Methods

ISE

MSPE

Aver.

std

Aver.

std

50

ESL

0.728

0.165

2.137

0.413

LS

87.72

31.46

486.7

103.2

LAD

14.47

16.82

63.59

89.74

100

ESL

0.341

0.105

0.536

0.153

LS

83.45

16.82

253.6

41.67

LAD

12.97

8.183

39.26

23.92

200

ESL

0.179

0.063

0.218

0.072

LS

76.81

11.74

165.3

17.23

LAD

11.34

3.769

14.35

6.138

Table 4. Simulation results of the averages (×102) and the standard deviations (×102) of integrated squared error (ISE) and mean square prediction error (MSPE) under Scenario 4.

n

Methods

ISE

MSPE

Aver.

std

Aver.

std

50

ESL

0.931

0.273

2.928

0.773

LS

673.5

703.4

2156

2013

LAD

32.18

132.6

107.3

388.5

100

ESL

0.452

0.156

0.763

0.248

LS

756.1

514.9

1235

891.4

LAD

15.14

14.42

42.37

28.36

200

ESL

0.237

0.083

0.153

0.076

LS

548.6

413.7

409.6

314.9

LAD

12.16

6.297

13.54

9.433

Figure 1. We refer to Smith et al. [17] for a detailed description of the application of this data.

Figure 1. For 213 subjects, fractional anisotropy (FA) curves were calculated for m=83 locations along the sagittal skeleton in the corpus callosum.

We applied the model proposed in this paper,

Y i ( t )=μ( t )+ j=1 p X ij β j ( t ) + l=1 q θ l ( M il ,t ) + ε i ( t ) , to this dataset. Here, X ij are the

categorical predictors (dummy encoded) with linear effects β j ( t ) , and M il are the continuous predictors. In this study, we regarded gender (coded by a dummy variable indicating males), handedness (coded by a dummy variable indicating left-handedness), education level, Alzheimer’s disease (AD) status, and mild cognitive impairment (MCI) status as categorical predictors, and age as a continuous variable. We centered all the functional responses and scalar predictors at zero mean. Then, we use the proposed ESL method to analyze the ADNI data, which approximates the nonlinear scalar function using cubic splines of order q=4 . The first five panels of Figure 2 show the estimated function coefficients of these five linear variables. We note that the function coefficient corresponding to gender takes positive values for the most part, while the functional coefficients of the other four predictors take negative values. This indicates that males tend to have higher FA values than females, and that patients with AD and MCI have lower FA values. Left-handedness or higher education level may lead to smaller FA values. The last graph shows the effect of continuous predicted age on FA values, which decrease with age and decline more rapidly than FA values during normal aging. These results are consistent with the previous analysis by Cai et al. [18].

Figure 2. Estimated functional coefficients (×103) for gender, handedness, education, AD, age and MCI for the ADNI data.

To evaluate the predictive performance of the model, we randomly divided 213 samples into a training set containing 149 samples and a testing set containing 64 samples. We use the training set to estimate the function coefficients and then predict the response in the test set. We use mean squared prediction error (MSPE) to verify the quality of the prediction, and the smaller the MSPE, the better the prediction. Based on 100 random splits, we obtained LS estimation average MSPEs and their standard deviations of 0.74× 10 2 and 0.21× 10 2 , respectively. We obtained the average MSPEs and standard deviation of our proposed ESL estimator, which are 0.69× 10 2 and 0.18× 10 2 , respectively. These results indicate that our proposed method is preferred when considering the robustness of predictive performance.

5. Conclusions

In order to better explain the behavior of the function response over time, this paper develops a robust variable selection procedure for the nonlinear FOSP model containing continuous type. The function response is projected into a low-dimensional principal component space by using functional principal component analysis, and the principal component scores are used for variable selection and regression modeling. A robust variable selection for all scalar predictors is achieved by combining the ESL function with the group SCAD regularization method. To implement the computational procedure of the proposed method, we propose an iterative algorithm based on the LLA method. The tuning parameters are selected through a data driven program to achieve the robustness of variable selection. Simulation studies and real data analysis show that when there are outliers in the function response or scalar predictors, the proposed method can select the relevant predictors.

There are still some problems worth solving in the future research. In this paper, all linear covariates are divided into a group, which can either be selected or rejected at the same time. This condition is too strict and has limitations in practical application. Therefore, the variable selection of paper for scalar predictors with nonlinear covariates needs further study. When nonlinear covariates have high dimensional characteristics, robust variable selection for functional response regression is a significant topic for continued research.

Conflicts of Interest

The authors declare no conflicts of interest regarding the publication of this paper.

References

[1] Barber, R.F., Reimherr, M. and Schill, T. (2017) The Function-on-Scalar LASSO with Applications to Longitudinal GWAS. Electronic Journal of Statistics, 11, 1351-1389.
https://doi.org/10.1214/17-ejs1260
[2] Chen, Y., Goldsmith, J. and Ogden, R.T. (2016) Variable Selection in Function-on-Scalar Regression. Stat, 5, 88-101.
https://doi.org/10.1002/sta4.106
[3] Goldsmith, J., Liu, X., Jacobson, J.S. And Rundle, A. (2016) New Insights into Activity Patterns in Children, Found Using Functional Data Analyses. Medicine & Science in Sports & Exercise, 48, 1723-1729.
https://doi.org/10.1249/mss.0000000000000968
[4] Fan, Z. and Reimherr, M. (2017) High-Dimensional Adaptive Function-on-Scalar Regression. Econometrics and Statistics, 1, 167-183.
https://doi.org/10.1016/j.ecosta.2016.08.001
[5] Parodi, A. and Reimherr, M. (2018) Simultaneous Variable Selection and Smoothing for High-Dimensional Function-On-Scalar Regression. Electronic Journal of Statistics, 12, 4602-4639.
https://doi.org/10.1214/18-ejs1509
[6] Ghosal, R. and Maity, A. (2021) Variable Selection in Nonlinear Function-on-Scalar Regression. Biometrics, 79, 292-303.
https://doi.org/10.1111/biom.13564
[7] Wang, X., Jiang, Y., Huang, M. and Zhang, H. (2013) Robust Variable Selection with Exponential Squared Loss. Journal of the American Statistical Association, 108, 632-643.
https://doi.org/10.1080/01621459.2013.766613
[8] Ash, R.B. and Gardner, M.F. (1978) Topics in Stochastic Processes. Academic Press.
[9] Yao, F., Müller, H. and Wang, J. (2005) Functional Data Analysis for Sparse Longitudinal Data. Journal of the American Statistical Association, 100, 577-590.
https://doi.org/10.1198/016214504000001745
[10] Hall, P. and Hosseini-Nasab, M. (2005) On Properties of Functional Principal Components Analysis. Journal of the Royal Statistical Society Series B: Statistical Methodology, 68, 109-126.
https://doi.org/10.1111/j.1467-9868.2005.00535.x
[11] Fan, J. and Li, R. (2001) Variable Selection via Nonconcave Penalized Likelihood and Its Oracle Properties. Journal of the American Statistical Association, 96, 1348-1360.
https://doi.org/10.1198/016214501753382273
[12] Wang, L., Chen, G. and Li, H. (2007) Group SCAD Regression Analysis for Microarray Time Course Gene Expression Data. Bioinformatics, 23, 1486-1494.
https://doi.org/10.1093/bioinformatics/btm125
[13] Yao, W., Lindsay, B.G. and Li, R. (2012) Local Modal Regression. Journal of Nonparametric Statistics, 24, 647-663.
https://doi.org/10.1080/10485252.2012.678848
[14] Stone, C.J. (1982) Optimal Global Rates of Convergence for Nonparametric Regression. The Annals of Statistics, 10, 1040-1053.
https://doi.org/10.1214/aos/1176345969
[15] Zou, H. and Li, R. (2008) One-Step Sparse Estimates in Nonconcave Penalized Likelihood Models. The Annals of Statistics, 36, 1509-1533.
https://doi.org/10.1214/009053607000000802
[16] Chen, J. and Chen, Z. (2008) Extended Bayesian Information Criteria for Model Selection with Large Model Spaces. Biometrika, 95, 759-771.
https://doi.org/10.1093/biomet/asn034
[17] Smith, S.M., Jenkinson, M., Johansen-Berg, H., Rueckert, D., Nichols, T.E., Mackay, C.E., et al. (2006) Tract-Based Spatial Statistics: Voxelwise Analysis of Multi-Subject Diffusion Data. NeuroImage, 31, 1487-1505.
https://doi.org/10.1016/j.neuroimage.2006.02.024
[18] Cai, X., Xue, L., Pu, X. and Yan, X. (2020) Efficient Estimation for Varying-Coefficient Mixed Effects Models with Functional Response Data. Metrika, 84, 467-495.
https://doi.org/10.1007/s00184-020-00776-0

Copyright © 2025 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.