1. Introduction
Generalized linear models (GLM) encompass several parametric regression models by studying parametric modeling between a (often canonical or known) link function between the mean response and certain covariates; check [1,2]. This is not always desirable because the link function is not always known and may be more complicated.
Several models have been developed to overcome this problem, such as non-parametric and semi-parametric regression models. However, the great scourge arises and limits the use of such models. Thus, efforts have been made to circumvent this difficulty. The remedy is available according to two approaches: the approximation of the functions of link and the reduction of dimension. To do this, the generalized additive model (GAM), in which the non-parametric component is replaced by a sum of functions with only one variable, was recommended by Hastie et al. [3] and detailed by Wood [4]. The only criticism of this type of model is that it does not take into account the interactions between the explanatory variables. Thus, the single index model (SIM) was developed by Hardle et al. [5] and Hristache et al. [6] because it makes it possible to reduce the dimensions and soften the restrictive parametric assumptions by replacing several covariates by a linear combination of them. Subsequently, the partially linear single-index model (PLSIM), making it possible to model discrete explanatory variables in the linear part, were developed by Liang et al. [7] and Chen et al. [8] for the case of longitudinal data. Generalized partial linear single index (GPLSIM) models that are based on kernel smoothing to estimate the single index link function first appeared in the work of Caroll et al. [9], while in that of Cao and Wang [10], we use penalized spline smoothing of the quasi-likelihood and the technique of Fisher scoring, which is theoretically more reliable and relevant.
Notice also that some covariates may be functional and are not taken into account by this model. It should be remembered that several works have focused on the study of functional variables (see, for example, Ramsay and Silverman [11] and Ferraty and Vieu [12]). Note also that semi-functional partial linear regression has been studied by Aneiros-Perez and Vieu [13], then partial linear modeling with multi-functional covariates by Aneiros-Perez and Vieu [14]. We can also refer to several books on this subject, such as Horvath and Kokoszka [15], Kokoszka and Reimherr [16], Schumaker [17], Ould-Said et al. [18], Ouassou and Rachdi [19,20], Laksaci et al. [21], Cao et al. [22], Li and Lu [23] and Yao et al. [24]. We specifically cite, for example, Yu et al. [25] for the study of the partially functional single-index linear regression model and Yu and Ruppert [26] for a comprehensive review of the penalized spline smoothing methodology for PLSIM, in which the sub-regression function underlying is assumed to be a spline function with a fixed number of knots. Partially linear generalized single index models for functional data (PLGSIMF) has been studied by Rachdi et al. [27] and Alahiane et al. [28] using B-spline expansion and the quasi-likelihood function where the functional part is linear.
In this paper, we study the generalized non-parametric functional partially linear single-index models GNPFPLSIM where the functional variables, which froze there, are taken into account. Notice that in this model (I) the link function is unknown, (II) the number of knots increases with the size of the sample, (III) the functional regression is estimated in parallel with the unknown link function and the simple nonparametric index function using an iterative algorithm based on smoothing by spline functions and maximizing the quasi-likelihood function.
In addition, we use Fisher’s algorithm to solve the maximization problem iteratively. In addition, we also provide a generalized cross-validation method to select the number of knots in the spline approximation and use kernel methods for functional data.
We also provide the convergence rates of our different estimators of the different parameters of GNPFPLSIM.
The rest of this paper is organized as follows. In Section 2 and Section 3, we develop an estimation methodology, give some asymptotic properties of the proposed estimators and we present an iterative algorithm based on the maximization of the quasi-likelihood function for the computation of the proposed estimators. A simulation study is conducted in Section 5. The technical lemmas allowing to prove Theorems 1, 2 and 3 are presented in Appendix A.
Notice finally that in order to save space, proofs of various results obtained are grouped in a supplementary file to this paper.
2. Some Preliminaries
Let Y be a scalar response variable and be the predictor vector where and Z belongs to H where is a semi-metric space of functions defined on , i.e., Z is a functional random variable and is a semi-metric.
For a fixed , we assume that the conditional density function of the response Y given belongs to a canonical exponential family, which is given by
(1)
where B and C are two known functions, and where is the unknown natural parameter in the generalized parametric linear models, which is linked to the dependent variable by(2)
where denotes the first derivative of the function B (see [10,29]).In what follows, we modelize as a generalized non-parametric functional partially linear single-index model (GNPFPLSIM) by
(3)
where is the single-index coefficient vector of dimension d, is the unknown non-parametric function in the functional component, and is the unknown single-index link function which will be assumed to be sufficiently smooth.Notice the following:
-
For identifiability reasons, we assume that and the first component of α is non-negative, i.e., , where denotes the Euclidean norm on .
-
In order to identify the function , we define its support as , where and .
-
The GNPFPLSIM includes as special cases the linear model (LM), the single-index model (SIM), as well as the partially linear model (PLM), the PLSIM, and the non-parametric models.
-
In the definition of the real canonical link function g, we assume that the functional random variable is valued in H and such that
-
If the conditional variance where is an unknown positive function, then the estimation of the mean function may be obtained by replacing the log-likelihood , given by (1), by the quasi-likelihood which is given, for any real numbers u and v, by
and which may be written as follows:
-
The regression operator , which is a nonlinear operator from H into , satisfies
-
•. where
or
-
•. There exists such that , where
-
3. Estimation Methodology
Let for , be an independent and identically distributed (i.i.d.) n-sample of . Then, for each
(4)
Let and such that . We designate by the collection of functions g defined on whose vth derivative, exists and satisfies the following Lipschitz condition of order
We introduce a knots sequence in the interval with J interior knots, such that
where increases as the sample size n increases.A function is said to belong to the space of polynomial splines, , on an interval of order , if
-
is a polynomial of degree on each sub-interval for and .
-
is -times continuously differentiable on .
Let be the number of knots, and let , be the B-spline basis functions of order . Moreover, let be the distance between the neighboring knots. Then, a function (which will be defined in Section 3) may be approximated by a function with where is a vector of cubic B-splines of order (see DeBoor [30]).
So, our estimation process consists of two steps as follows.
3.1. The First Step
Let
(5)
A non-parametric estimator of the regression operator is defined by
(6)
where , , and the kernel which is supported within , is of-
Type 1: if for constants ,
or
-
Type 2: if , K is derivable on and for some constants
and the sequence is the bandwidth (the smoothing parameter).
3.2. The Second Step
By plugging in the non-parametric estimator in (6), we consider the model
(7)
The mean function estimator is given by the evaluation of the parameters and inverting Equation (7). In fact, is determined by maximizing the following quasi-likelihood
(8)
where with and where and denote the true values, respectively, of and .In order to overcome the constraints and of the d-dimensional index , we proceed by a re-parametrization (see Yu and Ruppert [26])
The true value , of , must satisfy . Then, we assume that .
The Jacobian matrix of of dimension is
Notice that is unconstrained and of one dimension lower than the dimension.
Recall that since , then there exists such that and . Thus, let
(9)
Then
(10)
whereWe define , such that
(11)
Notice that is -dimensional, while is -dimensional. Then, becomes
For , we denote by
Then
The score vector is
(12)
whereThen, the Hessian matrix of the quasi-likelihood function is therefore
We have
Then
By replacing by , we obtain
and thenElsewhere, the Fisher scoring update equations become
(13)
where, forThen and is the estimator of the single index coefficient vector of the GNPFPLSIM model.
4. Some Asymptotic Properties
In this section, we present the asymptotic properties of the estimators for the non-parametric components, the parametric components, the single-index and the almost complete convergence of the functional regression operator of the GNPFPLSIM model. For this aim, we will need some assumptions.
4.1. Assumptions
Let , and be measurable functions on . We define the empirical inner product and its corresponding norm as follows
If , and are -integrable, we define the theoretical inner product and its corresponding norm as follows
We assume that
-
(C1).
-
(C2). For all and for all y in the range of the response variable Y, we have, for that
for some positive constants and -
(C3). The th order partial derivative of the joint density function of X satisfies the Lipschitz’ condition of order
The marginal density function of is continuous and bounded away from zero and supported within .
-
(C4). For any vector , there exists positive constants and such that
where -
(C5). The number of knots satisfies
-
(C6). The variable where is semi-metric space.
-
(C7)
-
–. The operator , where
-
–. There exists such that .
-
–. For all
-
–. The bandwidth satisfies when .
-
–. The kernel K is of Type 1 or Type 2.
-
–. For all , where is a continuous function in z.
-
-
(C8). There exist some positive constants such that
-
(C9). There exist positive constants such that the link function g satisfies
for all -
(C10). There exists a positive constant such that almost surely (a.s.)
Remark 2.
-
1.
If the kernel K is of Type 1, then there are two generic constants and such that
-
2.
If the kernel K is of Type 1 and if there exists and such that for all , , then there are two generic constants and such that for small enough
4.2. The Consistency Study
Under assumptions (C1)–(C5), we have
and
where (respectively, ) denotes the convergence in distribution (respectively, in probability), and and A are two matrices that will be defined in the Appendix A.
Under assumptions (C1)–(C5), we have
(14)
The proofs of the previous results are based on the following lemmas and among others on the papers of Pollard [31] and Stone [32].
(Lemma A.1. in Huang [33]). For any , let Then, for any
(Lemma A.2. in Wang and Yang [29] and Lemma A.4. in Xue and Yang [34]). Under assumptions (C1)–(C5), we have
where denotes the “O” Lanadau symbol, for the almost-complete convergence.
Let , and denote by
Then, we have the following lemma.
(Lemma A.3 in the Supplementary Material of Wang and Yang [29]). Under assumptions (C1)–(C8), there exists such that
where
4.2.1. Almost Complete Convergence of the Functional Kernel Estimator of
Under assumptions (C1)–(C7), we have
In order to show Theorem 1, we will need the following lemmas for which the proofs are given in the Supplementary Material. In fact, the proof is based on the following decomposition:
whereUnder assumptions (C1)–(C6), (C7)-1 and (C7)-4, we have
Under assumptions (C1)–(C7), we have the following:
-
(i)
If assumptions (C7)-3 to (C7)-5 are satisfied, then we have
-
(II)
If assumptions (C7)-3 and (C7)-4 are satisfied, then we have
(Corollary of Bernstein’s inequality).
- (i)
If for all there exists such that , then for all ,
- (II)
If depends on n, and for all there exists , such that and , then
4.2.2. Estimation of the Non-Parametric Function
Under assumptions (C1)–(C7), we have
and
The proof of Theorem 2 is given in Appendix A.
4.2.3. Estimation of the Parametric Components
Under assumptions (C1)–(C10), the quasi-likelihood estimator with the constraint is asymptotically normal, i.e.,
where
The proof of Theorem 3 and the used technical lemmas are given in Appendix A.
5. A Simulation Study
We aim to illustrate numerically the convergence of different estimators of the parameters , , the non-parametric function and the regression operator r of Y on Z. We conduct this numerical study in the Gaussian and in the logistic cases.
The conditional density of is given by
We deduce that
We consider that the model is given by the following equation
(15)
The responses are simulated according to Equation (15) (see Figure 1), is taken uniformly over the interval , whereas the errors . Moreover, we take the following coefficients
The functional real variable is taken as where and .
The regression operator r of Y on Z is defined as follows:
The knots are selected according to formula where (see Cao and Wang [10]). We choose and we make 2000 sample replications of sizes .
Then, the computed bias, the standard deviation (SD) and the mean squared error (MSE) with respect to (I) the parameter , and (II) the parameter are summarized in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10.
Notice that in the first step, we estimate the regression operator r using the functional kernel regression estimator by using the R-code routine
By the plug-in process (the second step) we estimate the parameters of the following model by using our algorithm GNFPLSIM as described before
(16)
To compute the bias, SD and MSE, we recorded 2000 replications of the GNFPLSIM algorithm in the Gaussian case (Table 1, Table 2, Table 3, Table 4 and Table 5) and in the logistic case (Table 6, Table 7, Table 8, Table 9 and Table 10) with as follows.
Notice that it is obvious to see that the quality of the estimators is illustrated through these simulations. The method works quite well. The bias, SD and MSE are generally reasonably low. The parametric and non-parametric components, simple index and nonlinear regression operator r of Y over Z are calculated by the procedure described above. The two tables therefore indicate the consistency of such that the bias, SD and MSE decrease as the sample size increases.
We developed our algorithm in both cases: the identity link function and the logistic link function. The simulations show that the GNFPLSIM algorithm works well in both cases.
In Figure 2, we illustrate 500 realizations of the functional random variable Z and the predicted response versus the true response.
We present below, in Figure 3, the single index estimated by the model in both cases: Gaussian and logistic cases.
We observe that the single-index estimated by our model fits well with the single-index.
We present below, in Figure 4, the systematic component estimated by the model in both cases: Gaussian and logistic cases.
We consider that our model approximated to the best non-parametric function . We use the square root of the average square errors criterion (RASE, see Lai et al. [35]) in the Gaussian case and in the logistic case as follows:
The following Table 11 and Table 12 summarize the samples’ means, medians and variances of the with different sample sizes in the Gaussian case.
The following table summarizes the samples means, medians and variances of the with different sample sizes in the logistic case.
We conclude that as the sample size n increases from 500 to the sample mean, median and variance of decrease.
Application to Tecator Data
In this paragraph, we apply the GNFPLSIM model for Tecator data, popularly known in the functional data analysis. These data can be downloaded in the following link
Given 215 finely chopped pieces of meat, (see Figure 5) Tecator’s data contain their corresponding fat contents (), near-infrared absorbance spectra () observed on 100 equal wavelengths in the range 850–1050 nm, the protein and the moisture content .
We try to predict the fat content of the finely chopped meat samples.
The following figure shows the absorbance curves.
We divide the sample randomly into sub-samples: the training of size 160 and the test of size The training sample is used to estimate the parameters and the test sample is employed to verify the quality of predictors. To perform our model, we use the mean square error of prediction (MSEP) like in Aneiros-Pérez and Vieu [13] defined as the following:
where is the predicted value based on the training sample and is variance of the response variables test sample.The following Table 13 and Table 14 show the performance of our GNPFPLSIM model by comparing it with other models. We can conclude that GNPFPLSIM is competitive for such data.
The following Figure 6, shows us the estimator of the non-parametric function by the model in both cases: Gaussian and logistic cases.
The following Figure 7 compares the content of fatness and its estimation by the model in both cases: Gaussian and Logistic cases.
We can see that our model well fits the content of fatness of “215 pieces of meat”.
6. Summary
In this paper, we introduce estimates for the non-parametric generalized functional partially linear single-index model (GNPFPLSIM). Our estimates are obtained via the kernel methods and the Fisher scoring update equation derived from the quasi likelihood function and the normalized B-splines basis with their derivatives.
We prove the n-consistency and asymptotic normality of our estimates and therefore, first, we define estimates of the estimator , which converges almost completely to the true operator regression r. Second, we define estimates, with rates, of the estimator , which still converge at the rate to the true non-parametric function . Finally, we define estimates, with rates, of the estimator , which still converge at the rate to the non-parametric parameter .
A numerical study reveals that our estimation performs well in higher dimensions. The quality of the estimators is illustrated via simulations and real data.
Formal analysis, M.A.; Investigation, I.O.; Methodology, M.R. and P.V. All authors have read and agreed to the published version of the manuscript.
Not applicable.
Not applicable.
Not applicable.
The authors declare no conflict of interest.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Figure 2. On the left plot: 500 realizations of the functional random variable Z, and on the right plot: the predicted response (in abscissa) compared to the true response (in ordinate).
Figure 3. On the left plot: single-index versus predicted single-index, Gaussian case. On the right plot: single-index versus predicted single-index, logistic case.
Figure 4. On the left plot: the non-parametric function [Forumla omitted. See PDF.] versus its estimator [Forumla omitted. See PDF.], Gaussian case. On the right plot: the non-parametric function [Forumla omitted. See PDF.] versus its estimator [Forumla omitted. See PDF.], logistic case.
Figure 6. On the left plot: estimated non-parametric function [Forumla omitted. See PDF.], Gaussian case. On the right plot: estimated non-parametric function [Forumla omitted. See PDF.], logistic case.
Figure 7. On the left plot: the content of fatness and its estimation, Gaussian case. On the right plot: the content of fatness and its estimation, logistic case.
Bias, SD and MSE according to the parameter
|
|
|
---|---|---|
Bias | 0.0004 | −0.0005 |
SD | 0.0006 | 0.0013 |
MSE | 5.20 |
1.94 |
Bias, SD and MSE evolutions with respect to the parameter
|
|
|
|
|
|
---|---|---|---|---|---|
Bias | −0.0054 | 0.0258 | −0.0387 | 0.0289 | −0.0093 |
SD | 0.0123 | 0.0165 | 0.0214 | 0.0152 | 0.0064 |
MSE | 1.8045 |
9.3789 |
1.9556 |
1.0662 |
1.2745 |
Bias, SD and MSE evolutions with respect to the parameter
|
|
|
|
|
|
---|---|---|---|---|---|
Bias | 0.0072 | 0.0006 | −0.0043 | 0.0006 | −0.0056 |
SD | 0.0062 | 0.0046 | 0.0024 | 0.0036 | 0.0042 |
MSE | 9.028 |
2.152 |
2.425 |
1.332 |
4.900 |
Bias, SD and MSE evolutions with respect to the parameter
|
|
|
|
|
|
---|---|---|---|---|---|
Bias | 0.0009 | 0.0027 | −0.0057 | 0.0031 | 0.0092 |
SD | 0.0056 | 0.0034 | 0.0028 | 0.0057 | 0.0051 |
MSE | 3.217 |
1.885 |
4.033 |
4.210 |
1.1065 |
Bias, SD and MSE evolutions with respect to the parameter
|
|
|
|
|
|
|
---|---|---|---|---|---|---|
Bias | 0.0009 | 0.0027 | −0.0057 | 0.0031 | 0.0092 | 0.0051 |
SD | 0.0046 | 0.0084 | 0.0142 | 0.0154 | 0.0232 | 0.0131 |
MSE | 2.792
|
8.017 |
2.2373 |
4.1140 |
7.1780 |
6.3386 |
Bias, SD and MSE according to the parameter
|
|
|
---|---|---|
Bias | −0.0084 | 0.0047 |
SD | 0.0103 | 0.0108 |
MSE | 1.7665 |
1.3873 |
Bias, SD and MSE evolutions with respect to the parameter
|
|
|
|
|
|
---|---|---|---|---|---|
Bias | −0.0043 | 0.0352 | −0.0389 | 0.0383 | −0.0065 |
SD | 0.0107 | 0.0234 | 0.0223 | 0.0136 | 0.0058 |
MSE | 1.3298 |
1.7866 |
2.0105 |
1.6518 |
7.589 |
Bias, SD and MSE evolutions with respect to the parameter
|
|
|
|
|
|
---|---|---|---|---|---|
Bias | 0.0087 | 0.0006 | 0.0467 | 0.0003 | −0.0054 |
SD | 0.0070 | 0.0061 | 0.0051 | 0.0047 | 0.0026 |
MSE | 1.2469 |
3.757 |
2.2069 |
2.2180 |
3.5920 |
Bias, SD and MSE evolutions with respect to the parameter
|
|
|
|
|
|
---|---|---|---|---|---|
Bias | 0.0006 | 0.0053 | −0.0083 | 0.0036 | −0.0072 |
SD | 0.0041 | 0.0027 | 0.0072 | 0.0064 | 0.0052 |
MSE | 1.7170 |
3.5380 |
1.2073 |
5.3920 |
7.888 |
Bias, SD and MSE evolutions with respect to the parameter
|
|
|
|
|
|
|
---|---|---|---|---|---|---|
Bias | 0.0027 | 0.0035 | −0.0048 | 0.0215 | −0.0187 | −0.0214 |
SD | 0.0063 | 0.0078 | 0.0127 | 0.0215 | 0.0254 | 0.0213 |
MSE | 4.698 |
7.309 |
1.8433 |
9.2450 |
9.9485 |
9.1165 |
The RASE criterion with the non-parametric function
Gaussian Case | Mean | Median | Variance |
---|---|---|---|
|
0.028 | 0.024 | 0.004 |
|
0.027 | 0.022 | 0.002 |
The RASE criterion with the non-parametric function
Logistic Case | Mean | Median | Variance |
---|---|---|---|
|
0.038 | 0.043 | 0.027 |
|
0.029 | 0.039 | 0.016 |
The MSEPs for different models: Gaussian case.
Functional Models | MSEP |
---|---|
Model 1 (GNPFPLSIM) |
0.019 |
Model 2 (GNPFPLM) |
0.059 |
Model3 (SIM) |
1.102 |
Model 4 (FM) |
1.831 |
The MSEPs for different models: logistic case.
Functional Models | MSEP |
---|---|
Model 1 (GNPFPLSIM) |
0.039 |
Model 2 (GNPFPLM) |
0.093 |
Model 3 (SIM) |
1.102 |
Model 4 (FM) |
1.831 |
Supplementary Materials
The following supporting information can be downloaded at:
References
1. McCullagh, P.; Nelder, J.A. Generalized Linear Models; 2nd ed. Chapman and Hall: London, UK, 1972; [DOI: https://dx.doi.org/10.1007/978-1-4899-3242-6]
2. Nelder, J.A.; Wedderburn, R.W.M. Generalized Linear Models. J. R. Stat. Soc. Ser. A; 1972; 135, pp. 370-384. [DOI: https://dx.doi.org/10.2307/2344614]
3. Hastie, T.J.; Tibshirani, R.J. Generalized Additive Models; Chapman and Hall: London, UK, 1990; [DOI: https://dx.doi.org/10.1002/sim.4780110717]
4. Wood, S.N. Generalized Additive Models. An Introduction with R; Chapman and Hall/CRC: Boca Raton, FL, USA, 2017; Volume 2017, [DOI: https://dx.doi.org/10.1201/9781315370279]
5. Härdle, W.; Hall, P.; Ichimura, H. Optimal Smoothing in Single-Index Models. Ann. Stat.; 1993; 21, pp. 157-178. [DOI: https://dx.doi.org/10.1214/aos/1176349020]
6. Hristache, M.; Juditsky, A.; Spokoiny, V. Direct Estimation of the Index Coefficient in a Single-Index Model. Ann. Stat.; 2001; 29, pp. 595-623. [DOI: https://dx.doi.org/10.1214/aos/1009210682]
7. Liang, H.; Wang, N. Partially Linear Single-Index Measurement Error Models. Stat. Sin.; 2005; 15, pp. 99-116.
8. Chen, J.; Li, D.; Liang, H.; Wang, S. Semiparametric GEE Analysis of Partially Linear Single-Index Models for Longitudinal Data. Ann. Stat.; 2015; 43, pp. 1682-1715. [DOI: https://dx.doi.org/10.1214/15-AOS1320]
9. Caroll, R.J.; Fan, J.; Gijbels, I.; Wand, M.P. Generalized Partially Linear Single-Index Models. J. Am. Stat. Assoc.; 1997; 92, pp. 477-489. [DOI: https://dx.doi.org/10.1080/01621459.1997.10474001]
10. Wang, L.; Cao, G. Efficient Estimation for Generalized Partially Linear Single-Index Models. Bernoulli; 2018; 24, pp. 1101-1127. [DOI: https://dx.doi.org/10.3150/16-BEJ873]
11. Ramsay, J.O.; Silverman, B.W. Functional Data Analysis; Springer: New York, NY, USA, 2005; [DOI: https://dx.doi.org/10.1007/b98888]
12. Ferraty, F.; Vieu, P. Nonparametric Functional Data Analysis: Theory and Practice; Springer Series in Statistics; Springer: Berlin, Germany, 2006.
13. Aneiros-Perez, G.; Vieu, P. Semi Functional Partial Linear Regression. Stat. Probab. Lett.; 2006; 76, pp. 1102-1110. [DOI: https://dx.doi.org/10.1016/j.spl.2005.12.007]
14. Aneiros-Perez, G.; Vieu, P. Partial Linear Modelling with Multi-Functional Covariates. Comput. Stat.; 2015; 30, pp. 647-671. [DOI: https://dx.doi.org/10.1007/s00180-015-0568-8]
15. Horváth, L.; Kokoszka, P. Inference for Functional Data with Applications; Springer Series in Statistics; Springer: New York, NY, USA, 2012.
16. Kokoszka, P.; Reimherr, M. Introduction to Functional Data Analysis; Chapman and Hall/CRC Press: Boca Raton, FL, USA, 30 June 2021.
17. Schumaker, L.L. Spline Functions: Basic Theory; Wiley: New York, NY, USA, 1981; Volume 14.
18. Ould-Said, E.; Ouassou, I.; Rachdi, M. Functional Statistics and Applications; Contributions to Statistics Springer: Berlin, Germany, 2013.
19. Ouassou, I.; Rachdi, M. Stein Type Estimation of the Regression Operator for Functional Data. Adv. Appl. Stat. Sci.; 2010; 1, pp. 233-250.
20. Ouassou, I.; Rachdi, M. Regression Operator Estimation by Delta-Sequences Method for Functional Data and its Applications. AStA Adv. Stat. Anal.; 2012; 92, pp. 451-465. [DOI: https://dx.doi.org/10.1007/s10182-011-0175-0]
21. Laksaci, A.; Kaid, Z.; Alahiane, M.; Ouassou, I.; Rachdi, M. Non parametric estimations of the conditional density and mode when the regressor and the response are curves. Commun. Stat.—Theory Methods; 2022; [DOI: https://dx.doi.org/10.1080/03610926.2021.1998831]
22. Cao, R.; Du, J.; Zhou, J.; Xie, T. FPCA-based Estimation for Generalized Functional Partially Linear Models. Stat. Pap.; 2020; 61, pp. 2715-2735. [DOI: https://dx.doi.org/10.1007/s00362-018-01066-8]
23. Li, C.S.; Lu, M. A Lack-of-fit Test for Generalized Linear Models via Single-Index Techniques. Comput. Stat.; 2018; 33, pp. 731-756. [DOI: https://dx.doi.org/10.1007/s00180-018-0802-2]
24. Yao, D.S.; Chen, W.X.; Long, C.X. Parametric estimation for the simple linear regression model under moving extremes ranked set sampling design. Appl. Math. J. Chin. Univ.; 2021; 36, pp. 269-277. [DOI: https://dx.doi.org/10.1007/s11766-021-3993-1]
25. Yu, P.; Du, J.; Zhang, Z. Single-Index Partially Functional Linear Regression Model. Stat. Pap.; 2020; 61, pp. 1107-1123. [DOI: https://dx.doi.org/10.1007/s00362-018-0980-6]
26. Yu, Y.; Ruppert, D. Penalized Spline Estimation for Partially Linear Single-Index Models. J. Am. Stat. Assoc.; 2002; 16, pp. 1042-1054. [DOI: https://dx.doi.org/10.1198/016214502388618861]
27. Rachdi, M.; Alahiane, M.; Ouassou, I.; Vieu, P. Generalized functional partially linear single-index models. Functional and High-Dimensional Statistics and Related Fields; Springer International Publishing: Cham Switzerland, 2020; pp. 221-228.
28. Alahiane, M.; Ouassou, I.; Rachdi, M.; Vieu, P. Partially Linear Generalized Single Index Models for Functional Data (PLGSIMF). Stats; 2021; 4, pp. 793-813. [DOI: https://dx.doi.org/10.3390/stats4040047]
29. Li, W.; Yang, L. Spline Estimation of Single-Index Models. Stat. Sin.; 2009; 19, pp. 765-783.
30. De Boor, C. A Practical Guide to Splines; Revised ed. Applied Mathematical Sciences Springer: Berlin, Germany, 2001; Volume 27.
31. Pollard, D. Asymptotics for Least Absolute Deviation Regression Estimators. Econ. Theory; 1991; 7, pp. 186-199. [DOI: https://dx.doi.org/10.1017/S0266466600004394]
32. Stone, C.J. The Dimensionality Reduction Principle for Generalized Additive Models. Ann. Stat.; 1986; 14, pp. 590-606. [DOI: https://dx.doi.org/10.1214/aos/1176349940]
33. Huang, J. Efficient Estimation of the Partly Linear Additive Cox Model. Ann. Stat.; 1999; 27, pp. 1536-1563. [DOI: https://dx.doi.org/10.1214/aos/1017939141]
34. Xue, L.; Yang, L. Additive Coefficient Modeling via Polynomial Spline. Stat. Sin.; 2006; 16, pp. 1423-1446.
35. Lai, P.; Tian, Y.; Lian, H. Estimation and Variable Selection for Generalised Partially Linear Single-Index Models. J. Nonparametr. Stat.; 2014; 26, pp. 171-185. [DOI: https://dx.doi.org/10.1080/10485252.2013.841156]
36. Van der Vaart, A.W.; Wellner, J.A. Weak Convergence and Empirical Processes with Applications to Statistics; Springer: New York, NY, USA, 1996.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
We study the non-parametric estimation of partially linear generalized single-index functional models, where the systematic component of the model has a flexible functional semi-parametric form with a general link function. We suggest an efficient and practical approach to estimate (I) the single-index link function, (II) the single-index coefficients as well as (III) the non-parametric functional component of the model. The estimation procedure is developed by applying quasi-likelihood, polynomial splines and kernel smoothings. We then derive the asymptotic properties, with rates, of the estimators of each component of the model. Their asymptotic normality is also established. By making use of the splines approximation and the Fisher scoring algorithm, we show that our approach has numerical advantages in terms of the practical efficiency and the computational stability. A computational study on data is provided to illustrate the good practical behavior of our methodology.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details



1 Ecole Nationale des Sciences Appliquées, Université Cadi Ayyad, Marrakech 40000, Morocco;
2 Laboratoire AGEIS EA 7407, Université Grenoble Alpes, AGIM Team, UFR SHS, BP. 47, CEDEX 09, 38040 Grenoble, France
3 Institut de Mathématiques de Toulouse, Université Paul Sabatier, CEDEX 09, 31062 Toulouse, France;