1. Introduction and Motivations
U-statistics, first introduced in [1] and building on earlier work in [2], represent an essential class of statistical tools serving as unbiased estimators. Specifically, U-statistics of order m with kernel h are constructed from a sequence of random variables defined on a measurable space , using a measurable function . They are given by
where Notice that serves as the nonparametric uniformly minimum variance estimator of , minimizing the following expression: where is the minimizer. Common estimators based on U-statistics include the empirical variance, Gini’s mean difference, and Kendall’s rank correlation coefficient. A notable example is the Wilcoxon signed-rank test, a classical nonparametric method for testing zero location, as discussed in [3], Example 12.4. Asymptotic properties of U-statistics for independent and identically distributed (i.i.d.) random variables were initially developed in [1] and subsequently advanced in [4,5,6,7]. Parallel results for V-statistics were presented in [8,9]. Comprehensive reviews of U-statistics literature can be found in works [5,10,11], with a more extensive discussion in [12]. U-processes, a generalization of U-statistics where the statistics are indexed by a family of kernels, extend U-statistics to infinite-dimensional settings. These processes serve as nonlinear extensions of empirical processes and are crucial for tackling complex problems in areas such as density estimation, nonparametric regression, and goodness-of-fit testing. Transitioning from empirical processes to U-processes involves specialized techniques, particularly in stationary contexts, and finds wide applications in estimator analysis, particularly for functions of varying smoothness. Notable applications include testing qualitative features in nonparametric statistics [13,14], cross-validation in density estimation [15], and deriving limiting distributions for M-estimators [10,16,17]. Ref. [10] provided necessary and sufficient conditions for the law of large numbers and the central limit theorem for U-processes. More recent applications include normality tests using U-processes [18] and the new normality tests proposed in [19], which utilize weighted distances between the standard normal density and local U-statistics derived from standardized observations. Using a characterization of symmetry in terms of extremal order statistics, the authors of [20] developed several new nonparametric tests of symmetry based on novel U-empirical processes. In a related work, ref. [21] considered a class of U-statistics with kernels indexed by a multivariate parameter and, as an application, discussed an empirical characteristic function-based transformation to symmetry. The median-of-means approach, which leverages U-statistics, was introduced in [22] for estimating the mean of multivariate functions in heavy-tailed distributions. A thorough overview of the theory behind U-processes was provided in [17]. The broad utility of U-statistics extends to various fields, including random graph theory for counting subgraphs such as triangles [23] and machine learning applications like clustering, image recognition, and graph-based learning. Even with random kernels of diverging orders, U-statistics remain relevant, as explored in [24,25,26,27]. Infinite-order U-statistics are also employed to construct simultaneous prediction intervals, addressing uncertainties in ensemble methods like subbagging and random forests [28]. Specific applications of U-statistics include the MeanNN method for estimating differential entropy [29] and novel test statistics for goodness of fit [30]. In genetics, ref. [31] used U-statistics for model-free clustering and classification, while [32] applied them to analyze random compressed sensing matrices. Ref. [33] evaluated independence tests in functional data using Kendall statistics, a specific type of U-statistics. In high-dimensional clustering, ref. [34] developed a U-statistics-based framework for group classification and partition significance, while [35] focused on dimension-agnostic inference methods using variational representations of test statistics. Other innovations include U-statistics-based empirical risk minimization [36] and asymmetric U-statistics for pattern matching in random strings and permutations [37]. Ref. [38] proposed U-statistics under left truncation and right censoring for nonparametric independence tests between time to failure and failure cause in competing risks, while [39] explored quadruplet U-statistics for network analysis. The extension of U-statistics to conditional empirical U-processes presents both practical advantages and technical challenges.We introduce Stute’s estimators for a sequence of random elements , where and , a Polish space, with . For a measurable function , our goal is to estimate the conditional expectation, or regression function, for , given by
whenever it exists, i.e., when . We let be a kernel function with support in , , satisfying Stute’s class of estimators for , known as conditional U-statistics, is defined for each as where is the set of all m-tuples of distinct integers between 1 and n, and is a sequence of positive constants converging to zero such that . In the particular case where , the regression function simplifies to , and Stute’s estimator corresponds to the well-known Nadaraya–Watson estimator of [40,41], as elaborated in [42,43]. The seminal work of [44] concentrated on determining the uniform convergence rate of to as a function of . Subsequently, ref. [45] investigated the asymptotic distribution of , drawing comparisons with the findings of Stute. Under appropriate mixing conditions, ref. [46] extended Stute’s framework to weakly dependent data, demonstrating the Bayes risk consistency of the corresponding classification rules. As an alternative to traditional kernel-based approaches, ref. [47] introduced symmetrized Nearest Neighbor conditional U-statistics. In a complementary vein, ref. [48] analyzed functional conditional U-statistics, establishing the asymptotic normality of their finite-dimensional distributions. Despite the critical role of nonparametric estimation of conditional U-statistics in functional data analysis, this area has remained relatively underexplored. Recent developments in this field, highlighted in [49,50,51,52,53], address key challenges, particularly regarding the uniform consistency of bandwidth selection.There has been increasing interest in regression models where the response variable is real-valued and the explanatory variables consist of smooth functions that can vary arbitrarily across observations. This form of data, referred to as functional data, is encountered across diverse fields, including climatology, medicine, economics, and linguistics. Functional time series emerge when continuous processes are divided into smaller segments such as daily intervals where each intraday curve is treated as a functional random variable. This paper delves into functional data by investigating the theory of U-processes. For an in-depth introduction to functional data analysis, we direct readers to foundational works [54,55,56], which present case studies from various disciplines alongside core analytical methods. It is important to note that the extension of probability theory to random variables within normed vector spaces, such as Banach and Hilbert spaces, precedes many contemporary developments in the field of functional data, as highlighted in [57]. Ref. [58] explored density and mode estimation in normed vector spaces, addressing the challenges posed by the curse of dimensionality in functional data. In the context of regression, ref. [56] examined nonparametric models, while key foundational contributions to the theoretical framework and practical applications can be found in works [59,60,61]. Recent advances in functional data analysis include [62], the authors of which provided consistency rates for various functionals of the conditional distribution, uniformly over a subset of the explanatory variable. Building on this, ref. [63] extended these results by establishing uniform in bandwidth (UIB) consistency rates for nonparametric models, covering key functionals such as the regression function, conditional distribution, conditional density, and conditional hazard function. Further contributions of [64] investigated local linear estimation of the regression function when the regressor is functional, demonstrating strong convergence uniformly in bandwidth parameters. Additionally, ref. [65] explored k-Nearest Neighbors (kNN) estimation for nonparametric regression models using strongly mixing functional time series data, establishing uniform nearly complete convergence rates under moderate conditions. For more recent advancements in the field, works [66,67,68,69,70,71] provide further insights and developments. Recent literature has increasingly advocated for the integration of dimension reduction techniques in regression models. One prominent approach is the use of single-index models which reduce the influence of multiple predictors to a single index—a projection in a specific direction—coupled with a nonparametric link function. These models effectively capture the essential relationship between predictors and the response while mitigating the curse of dimensionality by focusing on a one-dimensional index. The nonparametric link function, applied to this index, allows for greater flexibility than traditional linear models, extending the scope of linear regression, where the link function is simply the identity function (see [72,73,74,75,76]). Recent advances in Functional Data Analysis (FDA) have further highlighted the importance of addressing high-dimensionality in functional regression models (see [77,78,79] for comprehensive reviews). Semiparametric models, particularly functional single-index models (FSIMs), have emerged as powerful tools in this context. Refs. [80,81,82] explored the FSIM framework, extending single-index models to accommodate functional data. Ref. [83] introduced a functional single-index composite quantile regression model, utilizing B-spline basis functions to estimate both the unknown slope and link functions. Recent papers include [84], the authors of which proposed a compact FSIM where the coefficient function is restricted to be non-zero in a subregion, and [85], the authors of which developed methods for estimating general FSIMs where the conditional distribution of the response is governed by a functional predictor through a single-index structure. Ref. [86] advanced this framework by combining functional principal component analysis (FPCA) for the functional predictors with B-spline modeling for the parameters, applying profile estimation techniques for unknown functions and parameters. Additionally, refs. [87,88] investigated FSIMs in the presence of randomly missing responses in strongly mixing time series data. Ref. [89] introduced a functional single-index varying coefficient model, wherein the functional predictor forms the single index. By applying FPCA and basis function approximations, the authors proposed an iterative estimation procedure for slope and coefficient functions. In a similar vein, ref. [90] developed an automatic, location-adaptive estimation procedure for FSIM using the kNN techniques. Inspired by imaging data analysis, ref. [91] introduced a novel functional varying-coefficient single-index model to analyze functional response data with covariates of interest. Meanwhile, ref. [92] explored the nonparametric estimation of conditional cumulative distribution functions using a functional Hilbertian regressor in the single-index framework, later extending this methodology to the multi-index case in [93]. This approach avoids fixing the true parameter within a pre-specified sieve and offers a rigorous theoretical analysis of a direct kernel-based estimation method, demonstrating polynomial convergence rates. These contributions reflect the ongoing evolution of FSIMs as versatile and robust tools in the analysis of complex, high-dimensional functional data. The principal objective of this study is to establish a comprehensive framework and rigorously investigate the weak and uniform convergence properties of k-Nearest Neighbor (kNN) single-index conditional U-processes within a regular sequence of random functions. This research is driven by the fundamental statistical significance of the kNN method, a widely adopted and versatile approach. The kNN method identifies the k closest neighbors of a given point to x using a prescribed distance metric . A notable feature of the kNN approach is its random, locally adaptive bandwidth, which adjusts to the underlying structure of the data—a critical aspect, especially in infinite-dimensional settings. The kNN method, originally introduced in [94] and further examined in [95], has its roots in nonparametric discrimination and was later explored in [96]. For a comprehensive exposition of the method, readers are referred to [97]. In practice, the kNN method is widely used, as demonstrated in [56], and is favored for its simplicity as it involves only one parameter, k, which dictates the number of Nearest Neighbors. This parameter is typically selected from a finite set, and the method’s local adaptability enables it to respond to the specific data structure at any point. Extensive research has been conducted on the kNN method in finite-dimensional spaces. Notable contributions include works [98,99,100,101,102]. However, in infinite-dimensional spaces, particularly within functional frameworks, three primary approaches to kNN regression estimation have emerged. The first, introduced in [103], focuses on a kNN kernel estimate where the functional variable resides in a separable Hilbert space . In this setting, ref. [103] established weak consistency by projecting the infinite-dimensional space onto a finite-dimensional subspace. This was achieved by considering the first m coefficients of an expansion of X in an orthonormal basis of and subsequently applying multivariate kNN regression techniques to the projected data. The second approach integrated kNN methods with functional local linear estimation. Consistency and convergence rate results for this method were presented in [104,105], further enriching the theoretical underpinnings of kNN in infinite-dimensional contexts. These advances underscore the growing importance of kNN methods in functional data analysis and their adaptability to high-dimensional and complex data structures.
In this paper, we first aim to establish almost sure uniform consistency and almost sure uniform-in-the-number-of-neighbors (UINN) consistency for nonparametric functional single-index regression estimators and functional single-index conditional U-processes in the dependent setting. The concept of uniform-in-bandwidth (UIB) consistency was originally introduced in [15] for kernel density estimators using empirical process methods. This investigation is motivated by a series of foundational works, including [49,51,106,107], which established UIB consistency for similar estimators in the i.i.d. finite-dimensional framework, with varying within intervals indexed by n. In the realm of Functional Data Analysis (FDA), numerous studies have focused on nonparametric functional estimators. For example, ref. [62] provided consistency rates for several functionals of the conditional distribution, such as the regression function, conditional cumulative distribution, and conditional density, uniformly over specific subsets of the explanatory variables. Ref. [63] extended these results by establishing uniform consistency rates for conditional models, including regression functions, conditional distributions, densities, and hazard functions. Refs. [68,108] demonstrated almost complete convergence of k-Nearest Neighbor (kNN) estimators—uniform in the number of neighbors—under classical assumptions on kernel functions, small ball probabilities of the functional variable, and an entropy condition to control space complexity. Similarly, ref. [64] studied local linear estimation of the regression function in the context of functional covariates and established strong UIB convergence. Ref. [65] explored kNN estimation for nonparametric regression models with strongly mixing functional time series data, achieving uniform almost sure convergence rates under mild conditions, while [90] provided new uniform asymptotic results for kernel estimates within functional single-index models. Despite this rich body of literature, most studies have focused exclusively on either UIB or UINN consistency or on uniform consistency over specific functional subsets. However, the simultaneous investigation of both UIB and UINN consistency, particularly in a dependent setting, remains an open problem. This gap, which was addressed in [51] in the independent case, forms the central focus of our research. By integrating results from both Functional Data Analysis and empirical process theory, we aim to advance the understanding of consistency in more general settings. A second major challenge we address concerns the weak convergence of these estimators. This problem is inherently complex, as it requires controlling asymptotic equicontinuity under minimal conditions—a challenge that remains unresolved in the current literature. Our approach draws on seminal works [109,110], combined with strategies outlined in [111,112], to handle functional data. However, our contribution extends beyond simply merging existing concepts; it involves developing intricate mathematical derivations tailored to the specific nature of functional data. The effective application of large sample theory, particularly results related to dependent empirical processes, is essential to this endeavor, and we rely heavily on the theoretical foundations laid out in [109,110,112]. Notably, even in the i.i.d. setting, weak convergence for kNN single-index conditional U-processes has not yet been established—a gap we aim to fill with this work.
In addition, this work focuses on estimating the regression function via the kNN method in the single index nonparametric regression model under the mixing dependence condition. We emphasize that the kNN method is a fundamental statistical technique with several advantages. Generally, it is computationally fast and does not require extensive parameter tuning. One of the key features is its nonparametric nature, enabling it to adapt automatically to any continuous underlying distributions without relying on specific models. For significant statistical problems, including density estimation, classification, and regression, kNN methods are proven consistent when k is appropriately selected. kNN methods are one of the main paradigms in machine learning. Historically, the kNN method was first introduced in [94]—see also [95]—in the context of nonparametric discrimination and further investigated in [96]. For more details, refer to [97]. The investigation of single index models, popular in econometrics, is motivated by two primary concerns: dimension reduction and the interpretability of index in these models. For more on this, refer to [72,80,113] in infinite-dimensional settings. Therefore, the single functional index model accumulates the advantages of single index models and the potential of functional linear models in applications (see [114,115,116]). In essence, the single functional index model reduces dimensionality effects while capturing as much information as possible from the data. Functional data analysis is a challenging research field in the statistical community. The kNN method considers the k neighbors of nearest to x with respect to distance . The local bandwidth of the kNN is random and depends on data , respecting the local structure of the data, which is essential in infinite dimensions. It is commonly used in practice (see [56]) and is simple to handle because the user controls only one parameter: the number k of Nearest Neighbors, valued in a finite set. Additionally, it allows building a neighbor adapted to the data at any point. The kNN method is widely studied if the explanatory variable is in a finite-dimensional space (see [98,99,100,101,102]). In an infinite dimensional space, i.e., a functional framework, three approaches exist for kNN regression estimation. The first, published in [103], examines a kNN kernel estimate when the functional variable is an element of a separable Hilbert space . In [103], a weak consistency result was established by reducing the infinite dimension of through projection onto a finite dimension subspace, considering only the first ℓ coefficients of an expansion of X in an orthonormal system of and then applying multivariate techniques on the projected data for kNN regression. The second approach, based on the kNN procedure and functional local linear estimation, achieved consistency with convergence rate (see [104,105,117]). The third approach, in [118], is a purely functional method.
1.1. Contribution
This paper addresses challenges in high-dimensional functional data analysis by advancing classical kernel estimators for stationary random processes. It focuses on the k-Nearest Neighbor (kNN) single-index kernel estimator, examining its regression performance and uniform consistency within functional single-index regression. Introducing the concepts of Uniform In Bandwidth (UIB) and Uniform In Nearest Neighbor (UINN) consistency, the study aims to provide robust estimates for accurate relative error predictions. It extends the kNN methodology to functional single-index conditional U-statistics, analyzing their uniform and UINN consistency, asymptotic properties, and limiting behavior. A key contribution is establishing a uniform central limit theorem for function classes meeting certain moment conditions, applicable to both bounded and unbounded functions. The methodologies employed include advanced techniques such as the kNN framework, small-ball probability, Hoeffding decompositions, decoupling methods, and modern empirical process theory indexed by function classes. The results are derived under general conditions, enhancing their practical relevance. The findings have versatile applications in statistics, including time series forecasting, set-indexed conditional U-statistics, and estimation of the Kendall rank correlation coefficient. The theoretical framework utilizes maximal moment inequalities for U-processes and -mixing results as established in [119], providing a robust foundation for further exploration of high-dimensional functional data and sophisticated statistical models.
1.2. Organization of the Paper
The structure of this article is organized as follows. Section 2 introduces the functional framework and presents the necessary definitions for our analysis, alongside a thorough discussion of the assumptions underlying our asymptotic results. In Section 3, we investigate strong uniform convergence rates, providing a detailed exploration of the conditions that ensure such convergence. Section 4.1 is devoted to the weak convergence of empirical processes within the functional data setting, laying the foundation for subsequent theoretical developments. The main theoretical contributions, including the uniform Central Limit Theorem (CLT) for conditional U-processes, are articulated in Section 4.2, where we establish the key results of the study. In Section 5, we explore potential applications of our findings, covering set-indexed conditional U-statistics (Section 5.1), the Kendall rank correlation coefficient (Section 5.2), and discrimination problems (Section 5.3). Section 6 addresses practical concerns related to bandwidth selection, offering insights into its implementation in empirical studies. Concluding remarks, along with recommendations for future research directions, are provided in Section 7. For the sake of clarity and coherence, all proofs, which primarily draw upon modern empirical process theory, are consolidated in Section 8, with a focus on the core arguments given the technical length of the demonstrations. Additionally, a collection of pertinent technical results is included in the Appendix section to support the main text.
2. The Functional Framework
2.1. Generality on the Model
Statistical data often exhibit some degree of dependence, which can profoundly impact the accuracy of statistical inference. Neglecting to account for this dependence during analysis can lead to erroneous conclusions. The concept of mixing serves as a valuable tool for quantifying the level of dependence within a sequence of random variables, enabling the extension of classical results for independent sequences to those that exhibit weak dependence or mixing behavior. This study focuses specifically on -mixing, or absolute regularity, a measure of dependence originally introduced in [120,121]. Sequence is defined as -mixing if the conditional probabilities converge to the unconditional probabilities in a specific manner as the lag between observations increases. In the present analysis, we assume that the sequence of random elements is absolutely regular, which ensures that the dependence structure of the data is controlled and behaves predictably. For instance, Markov chains are known to be -mixing under the mild Harris recurrence condition, particularly in scenarios where the underlying state space is finite [27,122,123,124]. We let denote the inner product and the corresponding norm in the Hilbert space . We utilize as a complete orthonormal system for . We consider a sequence of stationary random copies of the random vector , where X assumes values in the abstract space and Y takes values in the abstract space . The Hilbert space is equipped with a semi-metric , which induces a topology to measure the proximity between two elements within . Importantly, this metric is independent of the specific definition of X to circumvent issues related to measurability. Furthermore, we define a semi-metric associated with a single index given by for . This construction facilitates the analysis of functional data through a single-index structure while accounting for the dependence introduced by the -mixing condition. We assume that fixed values exist, where observations satisfy the following relation:
(1)
where is a real random variable satisfying almost surely. To ensure identifiability, we impose the assumption that the regression function is differentiable, and for , we require that , where denotes the first element of the orthonormal basis in . This condition ensures the identifiability of the single functional index model. For a comprehensive discussion on the identifiability issues related to single functional index models, we refer to [80]. The primary objective of this study is to establish the weak convergence of the single-index conditional U-process, which is constructed from the single-index conditional U-statistic in the kNN framework introduced in [51]. This investigation is pivotal for advancing the theoretical understanding of nonparametric methods in high-dimensional functional data, particularly within the context of U-processes indexed by single indices. This kNN estimator is defined as(2)
which serves as an estimator for the multivariate regression function: The kernel function and is a symmetric measurable function from class . Further, is a vector of positive random variables depending on , defined for by where is a ball in and denotes the indicator function. This kNN estimator generalizes the functional conditional U-statistic defined as(3)
where are positive real numbers decreasing as . This estimator traces its origins to the Akaike–Parzen–Rosenblatt kernel density estimation [125,126,127], with early applications discussed in [94] and later republished in [95]. The extensive body of literature surrounding this topic includes seminal works such as [128,129,130,131], among others. The k-Nearest Neighbor (kNN) method remains a fundamental statistical tool, valued for its computational efficiency and minimal requirement for parameter tuning. Our objective is to extend the results from [49,112,132,133] to a multivariate framework, with a particular focus on VC-subgraph classes and small-ball probabilities, as introduced in [56,58,111]. For a comprehensive understanding of VC-theory, we refer readers to the foundational work in [134] and related literature. This extension is critical for advancing the theoretical understanding and practical application of nonparametric methods in multivariate functional data contexts.A class of subsets on a set C is called a VC-class if there exists a polynomial such that for every set of N points in C, the class selects at most distinct subsets.
A class of functions is called a VC-subgraph class if the graphs of the functions in form a VC-class of sets. Specifically, if we define the subgraph of a real-valued function f on S as the subset on such that
then the class forms a VC-class of sets on . Informally, a VC-class of functions is characterized by having a polynomial covering number, which is the minimal number of functions required to cover the entire class.
Let be a subset of a semi-metric space , and let be a positive integer. A finite set of points is called an ε-net of for a given if
where represents a ball of radius ε around the point . If is the cardinality of the smallest ε-net, i.e., the minimal number of open balls of radius ε needed to cover , then the Kolmogorov entropy (metric entropy) of the set is defined as
The concept of metric entropy, introduced by Kolmogorov (cf. [135]), has been the subject of extensive research across various metric spaces. Dudley [136] applied this concept to establish sufficient conditions for the continuity of Gaussian processes, thereby laying the foundation for significant generalizations of Donsker’s theorem concerning the weak convergence of empirical processes. Let and represent two subsets of the semi-metric space , and denote the Kolmogorov entropy for a given radius by and , respectively. The Kolmogorov entropy for the product subset within the semi-metric space is expressed as
Consequently, denotes the Kolmogorov entropy of the subset within the semi-metric space . Given a semi-metric d on , define a semi-metric on as where and are elements of . Furthermore, it is possible to define another semi-metric, , on , parameterized by in , and based on the individual semi-metrics . The selection of the appropriate semi-metric is critical in such analyses, as it significantly impacts the theoretical properties and results. For an in-depth discussion on these considerations, refer to [56], particularly Chapters 3 and 13, which provide comprehensive insights into the role of semi-metrics in functional data analysis.2.2. Conditions and Comments
Let us present the conditions that we need in our analysis in the case of known .
- (C.1.)
On the distributions/small-ball probabilities
- (C.1.1)
For and , we have
- (C.1.2)
For let , we have
where is a non-negative function, and as satisfying is bounded.
- (C.1.1)
- (C.2.)
On the smoothness of the model
- (C.2.1)
The regression satisfies for and some for ,
- (C.2.2)
The conditional variance, defined for is continuous in some neighborhood of ,
Further, we assume that for some and
is continuous in some neighborhood of - (C.2.3)
For the function does not depend on and is continuous in some neighborhood of ,
- (C.2.1)
- (C.3)
On the kernel function
- (C.3.1)
The kernel functions is supported within , and there exist some constants such that
and - (C.3.2)
The kernel is a positive function and differentiable function on with derivative such that
(4)
- (C.3.1)
- (C.4)
On the classes of functions
- (C.4.1)
The class of functions is bounded and its envelope function satisfies, for some ,
- (C.4.2)
The class of functions is unbounded and its envelope function satisfies, for some ,
- (C.4.3)
The metric entropy of the class satisfies, for some ,
where - (C.4.4)
The class of functions is supposed to be of VC-type with envelope function previously defined. Hence, there are two finite constants b and such that
for any and each probability measure such that .
- (C.4.1)
- (C.5)
On the dependence of the random variables
- (C.5.1)
Absolute regularity
for some and - (C.5.2)
There is a sequence of positive integers such that, as ,
- (C.5.1)
- (C.6)
On the entropy
For n large enough and for some , Kolomogorov’s entropy satisfies
(5)
(6)
where and is the minimal number of open balls of radius in needed to cover - (C.7.)
The sequences and (respectively, and ) verify
(7)
- (C.8.)
There exist sequences , and () and constants , such that
and(8)
(9)
(10)
(11)
-
Additional/alternative conditions
- (C.1’.)
For
- (C.1’.1)
We have
with , and is absolutely continuous in a neighborhood of the origin. - (C.1’.2)
We have
where is a non-negative function, , and as satisfying is bounded.
- (C.1’.1)
- (C.2’.)
The kernel function is supported within , and there exist some constants such that for
- (C.1’.)
Comments
In our nonparametric functional regression model, a significant theoretical challenge arises in establishing functional central limit theorems for conditional empirical processes and conditional U-processes under functional absolute regularity. Additionally, we incorporate random (or data-dependent) bandwidths based on the k-Nearest Neighbor (kNN) approach. Traditional statistical methods cannot be directly applied in this functional setting, so many of the conditions we impose are tailored to the characteristics of infinite-dimensional spaces, such as the topological structure of , the probability distribution of X, and the concept of measurability for the classes and . It is worth noting that these conditions draw inspiration from [56,58,65,111,112]. We begin with assumption (C.1.1), adapted from [111], which itself is influenced by [58]. As explained in [111], when , Condition (C.1.1) aligns with the fundamental axioms of probability calculus. Moreover, if is an infinite-dimensional Hilbert space, then can decay exponentially as . Condition (C.1.1) is a standard assumption on small ball probability, which controls the behavior of around zero. This condition allows us to express the small ball probability as the product of two independent functions, and ; for further details, see, for example, ref. [137] for diffusion processes, [138] for Gaussian measures, and [139] for general Gaussian processes. A widely recognized result in the literature expresses small ball probability in the form , where , with and . This formulation corresponds to various processes, such as Ornstein–Uhlenbeck and general diffusion processes (with and ) and fractal processes (with and ). For additional examples, refer to [140]. When dealing with functional data, it is crucial to gather information about the variability of the small ball probability to adapt it to the bias of nonparametric estimators. This information is typically obtained by assuming
- (C.1.1”)
It is important to highlight that condition (C.4.2) may be substituted with more general assumptions regarding the moments of , as detailed in [141]. Specifically:
-
(M.1)″ We let be a nonnegative continuous function, increasing on , such that for some , as ,
(12)
For each , we define such that . We assume further that
-
The following choices of are of particular interest:
-
(i). for some ;
-
(ii). for some .
-
In the following section, we comprehensively present the main results concerning the uniform consistency of the regression estimators and the conditional U-statistics. This includes detailed discussions on the theoretical foundations and the conditions under which these estimators achieve uniform consistency. We also explore the implications of these findings for practical applications, highlighting how they enhance the reliability and accuracy of statistical inference in regression analysis and conditional U-statistics.
3. Uniform Consistency
For simplicity reasons, Condition (C.1.1) on the small ball probability is replaced by
- (H.1)
For and
(13)
(14)
which is similar to Condition (C.1) used in [49,142], so we deal with instead of (whenever we encounter a similar situation in the proofs). This approach does not merely serve as a notational convenience but also facilitates the integration of the UIB and UINN results, enhancing the theoretical coherence between the two frameworks.3.1. Uniform Consistency of the kNN Kernel Estimator for Regression
In this section, we investigate the uniform consistency of the functional regression operator in its general form, which is expressed for all as
(15)
where depending on n and In fact, the kNN operator presented in (15) can be regarded as a natural generalization, broadening its applicability and enhancing its flexibility in handling more complex, high-dimensional functional data, of the traditional kernel regression estimator(16)
where the bandwidth depends on n (but does not depend on t).3.1.1. UIB Consistency for Functional Regression
Recall the bandwidths and given in Condition (C.7.). The following theorem plays an instrumental role in the sequel.
Under Assumptions(H.1.), (C.2.1) , (C.3.1) , (C.4.1) , (C.4.3) , (C.6.) and (C.7.) (for ), we have, as ,
(17)
The following result gives uniform consistency when the class of functions is unbounded.
Under Assumptions (H.1.) , (C.2.1) , (C.3.1) , (C.4.2) , (C.4.3) , (C.6.) and (C.7.) (for ), we have
(18)
3.1.2. UINN Consistency for Functional Regression
Now, we can state the main results of this section concerning the kNN functional regression. Recall the bandwidths and given in Condition (C.8.).
Under Assumptions (H.1.) , (C.2.1) , (C.3.1) , (C.4.1) , (C.4.3) , (C.6.) and (C.7.) (for ), if, in addition, Condition (C.8.) is satisfied, we have
The following result gives uniform consistency when the class of functions is unbounded.
Under Assumptions (H.1.) , (C.2.1) , (C.3.1) , (C.4.2) , (C.4.3) , (C.6.) and (C.7.) (for ) and if Condition (C.8.) is satisfied, we have
(19)
3.2. Relative-Error Prediction
Recall that the operator is typically estimated by minimizing the expected squared loss function . While this loss function is commonly used to assess prediction performance, it may not be suitable in all contexts. Least-squares regression assigns equal weight to all variables, potentially making the results highly sensitive to outliers. In this paper, we address the limitations of classical regression by proposing the estimation of the operator m through the minimization of the mean squared relative error (MSRE), which offers a more suited alternative:
(20)
This criterion provides a more meaningful measure of prediction performance than the least-squares error, particularly in cases where the range of predicted values is large. Moreover, the solution to (20) can be explicitly formulated as the ratio of the first two conditional inverse moments of Y given X. To develop a regression estimator that optimally minimizes the MSRE, we assume that the first two conditional inverse moments of Y given X, defined as for , exist and are finite almost surely. As demonstrated in [143,144,145], the optimal mean squared relative error predictor of Y given X is Therefore, we estimate the regression operator , which minimizes the MSRE, using(21)
and(22)
By considering the special cases and in Corollaries 1 and 2, we obtain results that complement the work in [145,146,147].Under Assumptions (H.1) , (C.2.1) , (C.3.1) , (C.4.2) , (C.4.3) , (C.6) , and (C.7) (for ), we have
(23)
This result has not been previously addressed in the literature.
Under Assumptions (H.1.) , (C.2.1) , (C.3.1) , (C.4.2) , (C.4.3) , (C.6.) and (C.7.) (for ), and if Condition (C.8.) is satisfied, we have
(24)
3.3. Uniform Consistency of the kNN Functional Conditional U-Statistics
In addition to the conditions previously established, the following assumptions are crucial for obtaining exponential inequalities for dependent data which are used later in the proofs:
- (A1)
Assume that the sequence is strictly stationary, and there exists an absolute constant such that, for any , the -mixing coefficient corresponding to satisfies .
- (A2)
Assume that, uniformly, for any integer J such that and arbitrary indices , the sequence , conditional on , satisfies, for the corresponding -mixing coefficient,
where represents the conditional probability. In particular, for the -mixing coefficient of the sequence , one obtains
The functional conditional U-statistic , given by
is a classical U-statistic with the U-kernel . The investigation of the uniform consistency of relative to presents notable challenges due to the randomness inherent in the bandwidth vector , which introduces significant technical complexities. To address these difficulties, we begin by analyzing the uniform consistency of , where represents a fixed multivariate bandwidth independent of and k. Our analysis extends to both the uniform consistency and UIB consistency of relative to , in cases where as well as when . Furthermore, we propose a more suitable centering factor than the expectation , defined as follows: The second step involves utilizing a general lemma from [51], suitably adapted to our specific framework, following an approach akin to that in [152] (see Appendix A.1). This allows us to rigorously derive the necessary results for the bandwidth .Uniform Consistency and UIB Consistency for a Multivariate Bandwidth
We now present the Uniform In Bandwidth (UIB) results for all , , and . To begin, we establish the result concerning the uniform deviation of the estimator from under the condition that the class of functions is bounded.
We suppose that Conditions (H.1) , (C.3.1) , (C.4.1) , (C.4.4) , (C.6) , and (C.7) hold. Then, as , we have
(25)
The following theorem establishes the uniform deviation of the estimator from its expectation , specifically for an unbounded class of functions that satisfy general moment conditions.
Suppose that Conditions (H.1) , (C.3.1) , (C.4.2) , (C.4.4) , (C.6) , and (C.7) hold. For all , as , we have
(26)
The following theorem provides the uniform deviation result for the estimate from for both bounded and unbounded classes of functions satisfying general moment conditions.
We suppose that Conditions (H.1) , (C.3.1) , (C.4.1) , (C.4.4) , (C.6) , and (C.7) (or alternatively (H.1) , (C.3.1) , (C.4.2) , (C.4.4) , (C.6) , and (C.7) ) hold. For all , as , we have
(27)
We suppose that Conditions (H.1) , (C.2.1) , (C.3.1) , and (C.6) hold. For all , with as , we have
(28)
Under the assumptions of Theorems 5 and 6, it follows that, as ,
(29)
Following the methodology outlined in [110,112,153], we decompose the U-statistics into distinct components. Certain components can be approximated by U-statistics derived from independent blocks, while others, when conditioned on a particular block, behave as empirical processes of independent blocks. To demonstrate the insignificance of the nonlinear terms, we utilize symmetrization techniques and apply maximal inequalities, as thoroughly discussed in [10,154].
3.4. Uniform Consistency and UINN Consistency of Functional Conditional U-Statistics
Let be constants and let be a sequence chosen such that
and The following theorem addresses the uniform deviation of the estimator from its expected value , under the assumption that the class of functions is bounded.Suppose that Conditions (H.1.) , (C.3.1) , (C.4.1) , (C.4.4) , (C.6.) , and (C.7.) are satisfied. Additionally, if assumption (C.8.) holds, then, as ,
(30)
The following theorem addresses the uniform deviation of the estimator from its expected value when the class of functions is unbounded but adheres to general moment conditions.
Suppose that Conditions (H.1.) , (C.3.1) , (C.4.2) , (C.4.4) , (C.6.) , and (C.7.) hold. Additionally, if Assumption (C.8.) is satisfied, then, as ,
(31)
The subsequent results are centered on establishing the uniform consistency of the estimator for both bounded and unbounded classes of functions, ensuring that the consistency holds across varying function spaces.
Suppose that Conditions (H.1.) , (C.3.1) , (C.4.1) , (C.4.4) , (C.6.) , and (C.7.) (or alternatively, (H.1.) , (C.3.1) , (C.4.2) , (C.4.4) , (C.6.) , and (C.7.) ) hold. If, in addition, Assumption (C.8.) is satisfied, then, as ,
(32)
Suppose that Conditions (H.1.) , (C.2.1) , (C.3.1) , and (C.6.) hold. If, in addition, Assumption (C.8.) is satisfied, then, as ,
(33)
Under the assumptions of Theorems 9 and 10, it follows that, as ,
(34)
The selection of parameters and , as defined analogously in Condition (C.8.) , is crucial in determining the convergence rate of the kNN estimator. These parameters are chosen based on the small ball probability function , offering flexibility in adjusting the estimator to enhance performance, depending on the specific characteristics of the data.
This study builds upon the foundational work in [49,51,155], introducing several important advancements. One of the primary distinctions of our approach is the minimal restrictions placed on the choice of the kernel function, with only mild conditions required, as detailed later in the paper. However, determining the optimal bandwidth or the number of neighbors presents a more complex challenge. The selection of bandwidth is pivotal, as it directly impacts the estimator’s consistency rate and bias. Our goal is to identify bandwidth and neighbor parameters that strike an optimal balance between bias and variance. Unlike traditional approaches, we propose a more flexible framework in which bandwidth and the number of neighbors are determined based on specific criteria, data characteristics, and local considerations. For a deeper exploration of this topic, readers are referred to works such as [156] and related literature. In this section, we address these challenges within the context of Functional Data Analysis (FDA), particularly under the assumption of dependent data. Specifically, we extend the analysis to single-index models and, for the first time, explore the application of conditional U-statistics in this setting. Our methodology involves decomposing U-statistics into several components: some of these components are approximated by U-statistics derived from independent blocks, while others, when conditioned on a single block, behave as empirical processes of independent blocks. While this decomposition is a powerful tool, it complicates the proof process—a necessary trade-off for extending the methodology to dependent structures. By utilizing this decomposition, we are able to adapt techniques traditionally applied to independent variables, drawing significantly from the results in [51,155]. Furthermore, this paper introduces a novel exponential inequality specifically designed for the dependent data setting, as presented in [150], representing a notable contribution to the existing literature.
In the following section, we provide a comprehensive presentation of the main results concerning uniform central limit theorems for both regression processes and conditional U-processes. This includes detailed statements of the theorems as well as thorough and meticulous explanations of the methodologies employed in proving these results.
4. Uniform Central Limit Theorems
4.1. kNN Conditional Empirical Process
We define the functional conditional empirical process for univariate bandwidth by
(35)
where designates (3) when , and refers to Regression Function (1), with If for , where is the probability measure and for each then is a random element with values in consisting of all functional on such that Then, it is important to investigate the following weak convergence: It is well established that weak convergence to a Gaussian limit, with uniformly bounded and uniformly continuous paths (with respect to the norm), is equivalent to the combination of finite-dimensional convergence and the existence of a pseudo-metric on . This pseudo-metric ensures that forms a totally bounded pseudo-metric space:(36)
Below, we denote to indicate that the random vector Z follows a normal distribution with mean vector and covariance matrix , and use to denote convergence in distribution. The following theorem, adapted from [111] for the context of kNN estimators, aims to investigate central limit theorems for the functional conditional empirical process defined as(37)
Let us consider the class of functions and suppose that Conditions (C.1’.) , (C.1.2) , (C.2.1) , (C.2.2) , (C.3’.) , (C.5.) , and (C.8.) hold. If the smoothing parameter k satisfies, for all ,
we obtain, for
where is the covariance matrix with
We suppose that Conditions (C.3.1) , (C.4.2)–(C.5.1) , and (C.8.) hold, and for each
Then, we have
The two previous theorems can be summarized as follows:
Under Conditions (C.1’.) , (C.1.2) , (C.2.) , (C.3.1) , (C.3’.) , (C.4.4)(C.5.1) , (C.5.2) , and (C.8.) , the process as ,
converges in law to a Gaussian process that admits a version with uniformly bounded and uniformly continuous paths with respect to the norm.We note that additional applications can be derived from Theorem 13, including the conditional distribution, conditional density, and conditional hazard function. However, these and other relevant applications are not addressed here due to space constraints.
Paper [157] introduces a comprehensive semiparametric functional regression model designed for cases where the predictor sets include a combination of functional and multivariate elements within the i.i.d. framework. This model integrates single-index techniques to handle functional predictors and partial–linear methods to address multivariate predictors. Similarly, ref. [90] conducts an in-depth exploration of the functional semiparametric model (FSIM), providing general asymptotic results for the kNN procedure with a focus on uniformity across all model parameters. In contrast, our study distinguishes itself from the aforementioned works in several key ways. First, while prior studies operate within an independent data setting, our research addresses a mixing data framework. Additionally, we extend the analysis by incorporating uniformity considerations for both classes of functions and the parameter “t”, as presented in Corollaries 4 and 1. Allowing function φ to vary within a class of functions enables results for the conditional distribution to emerge as a special case, among others. Another significant difference lies in our inclusion of weak convergence, as detailed in Theorem 13, which hinges on the intricate result of tightness outlined in Theorem 12, accompanied by rigorous and lengthy proofs. Furthermore, while earlier studies primarily focus on the case where , extending the results to the nonlinear domain where requires utilizing a range of techniques from empirical process theory and U-processes in the conditional setting poses a more complex and fascinating challenge compared to unconditional processes. This paper presents all results for arbitrary values of m. Finally, we explore the domain of censored data, a subject of independent interest that merits further investigation.
4.2. kNN Conditional U-Processes
In this section, we focus on investigating the weak convergence of conditional U-processes under the assumption of absolute regular observations. The class of functions under consideration is , as defined in Section 2. The conditional U-process, indexed by , is given by
(38)
The U-empirical process is defined by It is important to note that in order to establish the weak convergence of (38), it is first necessary to address the convergence of (40) below. In fact, we elaborate on certain key details that are utilized later in the analysis. Since Condition (C.6.) holds, for each , we have(39)
We can write the U-statistic as follows:(40)
We refer to the first term on the right-hand side of (40) as the truncated part, denoted by , and the second term as the remainder part, denoted by . Our initial focus is on . By applying Hoeffding’s decomposition, we obtain(41)
where is a sequence of i.i.d. r.v. with for each i, and and are defined, respectively, as and . In view of (41), we have(42)
the stationarity assumption and some algebras show that Therefore,(43)
Because of the fact that is -canonical, we have to show that
So, to establish the weak convergence of the U-process , it is enough to show where is a Gaussian process indexed by , and for , We have to prove, after, that the remaining part is negligible, in the sense that Nevertheless, when dealing with finite-dimensional convergence, the truncation becomes inconsequential. This implies that establishing the finite-dimensional convergence of is equivalent to establishing the convergence of .(a) Under Conditions (C.1’.) , (C.1.2) , (C.2.) , (C.3’.) , (C.5.1) , (C.5.2) and if is continuous at then, as ,
(44)
where(45)
(b) If, in addition, the smoothing parameter k satisfies Condition (C.8.) , then we have, as ,
(46)
where is defined as in (45), withUnder Conditions (C.1’.) , (C.1.2) , (C.2.) , (C.3’.) , (C.5.) and if as , then we infer that
(47)
Under Conditions (C.1’.) , (C.1.2) , (C.2.) , (C.3’.) , (C.5.1) , (C.5.2) , (C.8.) and as , we let be a measurable VC-subgraph class of functions from such that Condition (C.4.2) is satisfied and, if the β-coefficients of the mixing stationary sequence fulfill
(48)
for some , then converges in law to a Gaussian process which has a version with uniformly bounded and uniformly continuous paths with respect to the norm.The distinctive characteristics of kNN-based estimators introduce inherent complexities in analyzing their asymptotic properties. Specifically, the random variable , which depends on , adds technical intricacies to the proofs. A crucial observation is that the random elements involved in (2) for cannot be simply decomposed into sums of independent variables, as is typically the case with kernel-based estimators. This challenge requires more advanced probabilistic techniques, extending beyond the scope of standard limit theorems for sums of i.i.d. variables. Moreover, the direct application of Hoeffding’s decomposition—a fundamental tool for studying U-statistics—to (2) is not feasible in this context. The first step in proving Theorem 15 involves extending the work in [110,112] to address the multivariate bandwidth problem. It is also important to note that this paper extends the framework of [155] to single-index models. Additionally, we explore new applications, including the set-indexed conditional U-statistic, Kendall’s rank correlation coefficient, and time series prediction based on a continuous set of past values. Another layer of complexity arises from the fact that certain maximal inequalities and symmetrization techniques, as presented in [10,154], cannot be directly applied within this framework. This limitation necessitates a lengthier proof, particularly in establishing the equicontinuity of the empirical processes, a critical component in our analysis.
It is straightforward to adapt the proofs of our results to demonstrate that they remain valid when the entropy condition is replaced by the bracketing condition. Specifically, for some and ,
For the definition of , refer to p. 270 of [3].
We can also consider the scenario where , with satisfying the condition , where . For each , we assume , where converges to zero, as discussed in [90].
The set of functional directions, , is constructed using an approach similar to that of [81,90], as outlined below:
(i) Each direction is derived from a -dimensional space spanned by -spline basis functions, denoted by . Thus, we represent directions as
(49)
(ii) The set of coefficient vectors in (49), denoted by , is generated through the following steps:
- Step 1
For each , where represents a set of J ’seed-coefficients’, we construct the initial functional direction as
- Step 2
For each from Step 1 that satisfies , where is a fixed value in the domain of , we compute and normalize to obtain .
- Step 3
We define as the collection of vectors obtained in Step 2 . Consequently, the final set of admissible functional directions is given by
5. Some Potential Applications
Although only six examples are presented here, they serve as archetypes for a wide range of problems that can be explored using similar methods.
5.1. Set Indexed Conditional U-Statistics
Our objective is to analyze the relationships between X and Y by estimating functional operators associated with the conditional distribution of Y given X. This includes estimating operators such as the regression operator for within a class of sets . Specifically, for , the conditional distribution is defined as
We define the metric entropy of the class of sets by considering the covering number for each given by The logarithm of this covering number is referred to as the metric entropy of the class with respect to . Estimates for such covering numbers are available for various classes of sets (see, e.g., [158]). Often, we assume that the behavior of the covering number is governed by powers of . Specifically, Condition holds if where for some constants A and . As shown in [159], Condition with is satisfied for intervals, rectangles, balls, ellipsoids, and classes formed through finite applications of set operations like unions, intersections, and complements. In the case of convex sets in for , the condition holds with . Further examples of sets satisfying for are found in [158]. An important application of this theory is the estimation of using the following empirical estimator: Using Corollary 6, we conclude thatAn alternative perspective involves considering a compact set , where
The estimator for is given by
As , Corollary 6 implies that
5.2. Kendall Rank Correlation Coefficient
To assess the independence between the one-dimensional random variables and , Kendall [160] proposed a method based on a U-statistic, , with the kernel function:
The rejection region for the independence test is defined as . In this example, we extend Kendall’s method to the multivariate setting. Specifically, to test the conditional independence of and (where ) given X, we introduce a method based on the conditional U-statistic: where and is Kendall’s kernel. We assume that and are - and -dimensional random vectors, respectively, with . Suppose we have observations of , and we wish to test the following hypothesis: Let be a unit vector, where , , and . Let and be the distribution functions of and , respectively. Assume that and are continuous for any unit vector , where and . Here, denotes the transpose of the vector . For the case where , let and , where and for . Define By applying Corollary 6, conclude that as ,5.3. Discrimination Problems
We now apply the results to the discrimination problem outlined in Section 3 of [161], with reference to [162]. We adopt similar notation and framework. We let be a function that takes a finite number of values, say . The sets
partition the feature space. Predicting the value of is equivalent to predicting the partition set to which belongs. For any discrimination rule g, we have the following inequality: where This inequality becomes an equality if where is the Bayes rule. The associated probability of error, known as the Bayes risk, is given by Each of the unknown functions can be consistently estimated using the methods discussed earlier. For , the estimator is defined as We define the estimated rule as and introduce the estimated probability of error: The discrimination rule is asymptotically Bayes risk consistent, meaning that as , This result follows from Corollary 6 and the fact that5.4. Generalized U-Statistics
The extension to the case of multiple samples is straightforward. Consider k independent collections of independent observations:
Let, for , be defined as where is assumed, without loss of generality, to be symmetric within each of the k blocks of its arguments. The conditional U-statistic for estimating , corresponding to the kernel and assuming , is given by where and is a vector of positive random variables depending on the sets Here, denotes a set of distinct elements from the set for , and represents summation over all such combinations. The extension of Hoeffding’s treatment of one-sample U-statistics to the multi-sample case is due to [163] and [164]. Using Corollary 6, it is possible to infer that5.5. Conditional U-Statistics for Censored Data
We consider a triple of random variables defined on , where Y is the variable of interest, C is a censoring variable, and X is a concomitant variable. We work with a sample consisting of independent and identically distributed replications of , with . In the right-censorship model, the pairs , are not directly observed. Instead, the available information is provided by and for . Therefore, the observed sample is
Such censoring occurs frequently in practical settings, for example, in survival data from clinical trials or failure time data in reliability studies. Often, even under well-controlled conditions, samples are incomplete. For instance, clinical data tracking survival from many diseases are typically censored due to competing risks, like death from other causes. In the following, we impose several assumptions on the distribution of . For , we let denote the right-continuous distribution functions of Y, C, and Z, respectively. For any right-continuous distribution function defined on , we let be the upper point of the corresponding distribution. Now, we consider a pointwise measurable class of real-valued measurable functions defined on and assume that is of VC type. We recall the regression function of evaluated at , for and , given by in the case where Y is right-censored. To estimate , we use the Inverse Probability of Censoring Weighted (I.P.C.W.) estimators, which have gained popularity in the censored data literature (see [165,166]). The key concept behind I.P.C.W. estimators involves introducing the real-valued function defined on as(50)
This function incorporates both the observed data and the censoring mechanism to provide consistent estimates in the presence of censoring. Assuming that function is known, we observe that is directly observable for every . Furthermore, under Assumption () outlined below,- ()
C and are independent.
We have
(51)
Therefore, any estimate of , which is constructed based on fully observed data, also serves as an estimate for . This key property allows most statistical procedures used to estimate the regression function in the uncensored case to be naturally extended to the censored case. For example, kernel-type estimates can be readily constructed. For , , and , we define(52)
Based on Equations (50)–(52), when is known, a kernel estimator for is given by(53)
Function is generally unknown and has to be estimated. We denote by the Kaplan–Meier estimator of function [167]. Namely, adopting conventions and and setting we have Given this notation, we investigate the following estimator of(54)
referring to [165,168]. Adopting the convention that , the given quantity is well-defined, since only if and , where represents the kth order statistic from the sample for , and corresponds to such that . A right-censored version of an unconditional U-statistic, with a kernel of degree , can be introduced via the principle of a mean-preserving reweighting scheme, as outlined in [169]. The almost sure convergence of multi-sample U-statistics under random censorship was established by the authors of [170] who also explored the consistency of a new class of tests designed to evaluate equality in distribution. To mitigate potential biases from right-censoring and confounding covariates, ref. [171] proposed modifications to classical U-statistics. Moreover, ref. [172] introduced an alternative estimation method for the U-statistic by employing a substitution estimator for the conditional kernel, derived from observed data. For additional references, see [42]. To the best of our knowledge, the problem of estimating conditional U-statistics in the presence of censoring—particularly when using variable bandwidths—remains an unresolved issue. This provides the primary motivation for the present study. A natural extension of the function defined in (50) is given by(55)
From this, we have an analogous relation to (51) given by(56)
An analog estimator to (3) in the censored case is given by(57)
where, for ,(58)
The estimator that we investigate is given by(59)
The main result of this section is given in the following corollary.Under the assumptions of Theorems 9 and 10, it follows that, as ,
(60)
This last result is a direct consequence of Corollary 6, and the law of iterated logarithm for established in [173] ensures that
For more details, refer to [131].5.6. Conditional U-Statistics for Left Truncated and Right Censored Data
Building on the notation from the previous section, we now introduce a truncation variable, denoted by L, and assume that is independent of Y. We consider a scenario where we have random vectors , with . In this section, our goal is to define conditional U-statistics for data that are left truncated and right censored (LTRC), drawing on ideas from [38], which dealt with the unconditional setting. To this end, we propose the following extension of Function for LTRC data:
According to , we determine that An analog estimator to (3) for LTRC data can be expressed as follows:(61)
where is defined as in (58). Since is unknown, it needs to be estimated. To achieve this, we introduce and , representing the counting processes for the variable of interest and the censoring variable, respectively. Additionally, we define and . We introduce the risk indicators as and , where represents the risk set at time t, consisting of subjects who entered the study before time t and are still under observation at that time. Importantly, is a local sub-martingale with respect to the appropriate filtration . The martingale associated with the censoring counting process under filtration is given by where represents the hazard function associated with the censoring variable C under left truncation. The cumulative hazard function for the censoring variable is defined as . We let . Next, we define the sub-distribution function of corresponding to and as follows: We let where . Also, we denote Then, an estimate for the survival function of the censoring variable C under left truncation, denoted as —see [174]—can be formulated as follows:(62)
Similar to the Nelson–Aalen estimator— for instance, see [175]—the estimator for the cumulative hazard function of the censoring variable C under left truncation is represented as(63)
In both the definitions presented in (62) and (63), we make the assumption that is non-zero with probability one. The interrelation between and can be expressed as We let and denote the left and right endpoints of the support. For LTRC data, as in [176], is identifiable if and . By Corollary 2.2. [176], for , we readily infer that(64)
From the above, Estimator (61) can be rewritten directly as follows:(65)
The last estimator is the conditional version of that studied in [38]. Following the same reasoning of Corollary 8, one can infer that as
(66)
Remark 11.
-
In the study by the authors of [177], the practical utility of U-statistics is exemplified through an analysis of data from a case–control study among African Americans. The researchers investigated whether the similarity of haplotypes between a case and a control participant differs from the similarity between two control participants. By employing U-statistics, they effectively measured and compared haplotype similarities, highlighting the method’s applicability in genetic association studies and its capacity to handle complex genetic data.
-
Expanding on the traditional use of U-statistics, the authors of [178] introduced a new family that includes the well-known Wilcoxon signed-rank statistic and extends to other statistics with substantially higher power in sensitivity analyses of observational studies. These enhanced statistics were applied in three diverse examples drawn from epidemiology, clinical medicine, and genetic toxicology. The results demonstrated not only improved performance but also the flexibility of these new U-statistics in addressing various research questions across different fields.
-
In another significant contribution, the authors of [179] combined the properties of U-statistics with large-dimensional random matrix theory to identify group structures among numerous variables. By analyzing the eigenvalues and eigenvectors of the scaled sample matrix, they uncovered underlying patterns and groupings within high-dimensional data sets. This approach showcases the power of U-statistics in dimensionality reduction and the exploration of complex data structures, which are common challenges in modern data analysis.
-
Furthermore, the authors of [180] proposed the heterogeneity-weighted U (HWU)-method for association analyses that consider genetic heterogeneity. This innovative method leverages U-statistics to provide a computationally efficient solution suitable for high-dimensional genetic data and various types of phenotypes. The HWU method addresses the significant challenge of genetic heterogeneity in association studies, offering researchers a powerful tool to detect associations that might be overlooked by traditional methods.
5.7. Examples of Classes of Functions
The set of all indicator functions representing intervals in satisfies the following inequality:
for any probability measure and for . We observe that
For further details, refer to Example 2.5.4 in [181] and [182] (p. 157). The covering numbers for the class of intervals in higher dimensions also satisfy a similar bound, though with a higher power of ; see Theorem 9.19 in [182].
(Classes of functions that are Lipschitz in a parameter, Section 2.7.4 in [181]). We let be the class of functions that are Lipschitz continuous with respect to the parameter . We suppose
for some metric d on the index set T and a function defined on the sample space , for all x. According to Theorem 2.7.11 in [181] and Lemma 9.18 in [182], it follows that for any norm on Thus, if satisfies the result holds for .We consider the class of functions that are smooth up to order α, as discussed in Section 2.7.1 of [181] and Section 2 of [183]. For , we let denote the greatest integer smaller than α. For any vector of d integers, we define the differential operator:
where . For a function , we define the norm
where the suprema are taken over all in the interior of . We let denote the set of continuous functions such that . For , this class consists of bounded functions satisfying a Lipschitz condition. According to [135], the entropy of class for the uniform norm was computed, and [183] showed that there exists a constant K, depending on α, d, and the diameter of , such that for every measure γ and every ,
where denotes the bracketing number (see Definition 2.1.6 in [181]). A variant of this inequality is given in Theorem 2.7.1 of [181]. Lemma 9.18 of [182] also implies
5.8. Examples of U-Kernels
In this section, we present some classical U-kernels.
Ref. [1] introduced the parameter
where and is the joint distribution function of and . Parameter Δ has the property that if and only if and are independent. As shown in [11], an alternative expression for Δ can be developed by introducing the following functions:
and
This leads to expression
(Hoeffding’s D). The symmetric kernel
recovers Hoeffding’s D statistic, a rank-based U-statistic of Order 5, which leads to Hoeffding’s D correlation measure .(Blum–Kiefer–Rosenblatt’s R). The symmetric kernel
yields Blum–Kiefer–Rosenblatt’s R statistic [184,185,186,187].(Bergsma–Dassios–Yanagimoto’s ). In [188], a rank correlation statistic was introduced as a U-statistic of Order 4 with the symmetric kernel:
Here, denotes the indicator that both and .(The Wilcoxon Statistic). We suppose is symmetric around zero. As an estimator of the quantity
we consider the Wilcoxon statistic, which is useful for testing whether the mean μ is located at zero.(The Takens Estimator). We let denote the Euclidean norm on . In [189], the following estimator of the correlation integral,
is given by When a scaling law holds, i.e., for some , the U-statistic provides the Takens estimator of the correlation dimension α.We let denote the oriented angle between and , where T is the unit circle in . We define
Ref. [190] used this kernel to propose a U-process for testing uniformity on the circle.
For , we let
We have
and the corresponding conditional U-statistic can be considered a conditional analog of the Hollander–Proschan test statistic [191], which tests whether the conditional distribution of given is exponential or of the New-Better-than-Used type.
(The Gini Mean Difference). The Gini index, another popular measure of dispersion, corresponds to the case where and . The Gini statistic is given by
In the next section, we provide more details on how several neighborhood selection methodologies from the literature can be combined with our results.
6. The Bandwidth Selection Criterion
Numerous methodologies have been proposed for constructing asymptotically optimal bandwidth selection rules, particularly in the context of the Nadaraya–Watson regression estimator. Prominent references in this area include [42,51,52,192,193,194]. The accurate bandwidth selection, whether in a finite- or infinite-dimensional setting, is critical to achieving optimal performance in practical applications. However, to the best of our knowledge, no studies have yet addressed the general functional conditional U-statistic. Unlike the real-valued case, where the selection of the number k has been thoroughly explored in [195], we propose an extension of the leave-one-out cross-validation procedure. For any fixed , we define
(67)
where . This expression defines the leave-one-out- estimator for functional regression and serves as a predictor for . To minimize the quadratic loss function, we introduce the following criterion, where represents a known non-negative weight function:(68)
where . Similarly, we define(69)
Following the ideas in [81,90,194], an intuitive approach to bandwidth selection is to minimize this criterion. We let and We let minimize over , and let minimize over , We now state the following corollary:
-
Under the assumptions of Corollary 5, as ,
(70)
-
Under the assumptions of Corollary 6, as ,
(71)
A key advantage of our results lies in deriving asymptotics for data-driven parameters. We let be a density function in and let denote the number of neighborhoods associated with the values. The conditional density can be estimated as follows:
The leave-one-out estimator is given by Although cross-validation procedures aim to approximate the quadratic estimation errors, alternative methods for selecting smoothing parameters can be introduced to optimize predictive power. The criterion is given byAs emphasized in [90], the range of values for parameter h, denoted as , is typically chosen by minimizing a criterion function, such as the cross-validation (CV) function, over a broad interval. This ensures that any reasonable values of h satisfying the technical conditions are included. While the automatic determination of the interval remains unresolved in one-dimensional nonparametric statistics, it is often of lesser concern since the criterion function typically exhibits flat behavior around its minimum. Early works, such as [193,196], suggest selecting an interval allowing bandwidths to capture up to 95% of the sample. As discussed in [90], this remains effective in the functional framework. Similarly, for parameter k, the set of values is determined by similar principles, with global kernel estimate considerations applying equally to local bandwidth estimates (see [197] for early developments). The same recommendation applies to the selection of .
In the following section, we provide a comprehensive summary of our main findings, highlighting the key results. Additionally, we outline potential directions for future work, suggesting areas where further investigation could be beneficial.
7. Concluding Remarks
This paper is centered on the k-Nearest Neighbor (kNN) kernel-type estimator for single-index conditional U-statistics, with the single-index Nadaraya–Watson estimator serving as a special case within a functional framework applied to regular datasets. Our results are predicated on specific regularity conditions related to conditional U-statistics and conditional moments, along with decay rates governing the probability distribution of variables within diminishing open balls and favorable decreasing rates for mixing coefficients. The conditional moment assumption plays a critical role, as it allows us to work with unbounded function classes. The proof of weak convergence adheres to a classical methodology, which entails establishing finite-dimensional convergence and rigorously controlling the equicontinuity of conditional U-processes. Finite-dimensional convergence is achieved through a block decomposition technique that ensures independence, after which a central limit theorem for independent variables is proven. However, the challenge of controlling equicontinuity necessitates a more sophisticated approach, particularly given the generality and complexity of the framework we consider. Full details of these proofs are provided in the subsequent sections. It is important to note that while we assume mixing—a form of asymptotic independence—primarily for simplicity, this assumption may not fully capture scenarios involving strong data dependence. As highlighted in [198], -mixing is regarded as the minimal assumption that allows for a “complete” empirical process theory, incorporating maximal inequalities and uniform central limit theorems. Explicit upper bounds for -mixing coefficients are available in the context of Markov chains (cf. [199]) and for V-geometric mixing coefficients (cf. [200]). Various stationary time series models, including linear processes (see [201] for -mixing), ARMA models (cf. [202]), and nonlinear autoregressive models (cf. [203]), have been extensively studied under similar conditions. A common assumption across these works is the continuity of either the observed process or the innovations associated with it. The application of nonparametric functional methods to general dependence structures remains a relatively unexplored domain; see [204,205,206,207,208,209,209]. Notably, our study employs an ergodic framework that circumvents the need for strong mixing conditions and their variations, thus avoiding the intricate probabilistic computations these conditions typically entail. While extending our results to encompass functional ergodic data would undoubtedly be a valuable direction for future research, it would necessitate significant mathematical advancements and falls beyond the scope of this paper. The primary challenge lies in the development of new probabilistic tools analogous to those employed in this study, but specifically adapted for -mixing samples. Finally, it would be of great interest to investigate the potential connection between our results and change-point detection, a method widely utilized for identifying abrupt shifts in stochastic systems caused by external perturbations. This technique has found applications across diverse scientific disciplines (see [210]), and exploring how our findings may relate to change-point problems presents a promising avenue for future inquiry. It will be of interest to find some links of the present with papers [211,212].
8. Mathematical Developments
This section is dedicated to the detailed presentation of the proofs underlying our results, maintaining consistency with the notation established earlier. Our proof techniques are inspired by those developed in [155], which we adapt and extend to the single-index framework. Moreover, we incorporate several of the more intricate arguments from [109], as used in prior works such as [110,112].
8.1. Proofs of Uniform Consistency Results
8.1.1. Proof of Theorem 1
To derive the convergence rates, it is essential to introduce the following notation. For every , we define
This allows us to write Now, let us consider the following decomposition:Therefore, the proof of (A2) is based on Lemmas A2, A3 and A9.
8.1.2. Proof of Theorem 2
In a manner analogous to [51], verifying the conditions of Lemma A1 for the case of is crucial for establishing Theorem 2. To achieve this, we first introduce the following designations: , , and . Subsequently, we define
We then define and such that(72)
(73)
We denote and define for any increasing sequence such that . It is important to observe that for all , , and , we haveUsing Condition (8), it follows that the bandwidths and both belong to the interval
The remainder of the proof follows the same steps as those in [155] and is therefore omitted. □
8.2. Preliminaries of the Proofs
This section is primarily devoted to the analysis of functional conditional U-statistics. As in the case of , where is covered by
for some radius , it follows that for each , there exists , where , such that Thus, for each , the closest center is , and the corresponding ball with the closest center is The consistency proofs for the uniform integrated bandwidth (UIB) concerning the multivariate bandwidths follow the same approach as the UIB consistency proofs for the univariate smoothing parameter in [49,51]. Furthermore, analogous to the proof of Theorem 1, we partition the sequence into alternating blocks, where the block sizes and differ, satisfying(74)
and for , we set8.2.1. Proof of Theorem 3
In this section, we consider a bandwidth . To establish Theorem 3, we represent the U-statistic for each as follows:
(75)
(76)
Let us begin with the term . We have By applying the Telescoping binomial, we obtain(77)
(78)
From Condition (C.4.1), we could claim that Similarly, we have So, (77) satisfies where Therefore, we infer that(79)
(80)
The transition from Equations (79) to (80) is enabled by the Lipschitz continuity of the kernel function . Uniformly for all and , we derive the following: by (14), where and , with component by component. The idea is to apply Lemma A21 on function which satisfies, for all and , It is important to observe that Constant appearing on the final right-hand side of the preceding inequality is derived from Condition (7). At this stage, we can invoke Lemma A21 with , yielding(81)
(82)
such that . By rigorously performing the calculations while adhering to the specified conditions, particularly (C.6) and (C.7), we derive the following result:(83)
The study of term is deduced from the previous one. In fact,(84)
(85)
To facilitate the transition from (84) to (85), we apply Jensen’s inequality, leveraging specific properties of the absolute value function. By proceeding along the same steps as previously outlined, we obtain where Notice that we have That implies This offers(86)
Continuing, now with , Assuming symmetry of the kernel function , it becomes necessary to decompose the U-statistic using the [1] decomposition, leading to the following expression:(87)
We define new classes of functions for , , and These classes are VC-type classes of functions with the same characteristics and the envelope function satisfying Let us start with the linear term of (87), which is From Hoeffding’s projection, we have One can see that is an empirical process based on a VC-type class of functions contained in with the same characteristics, and the elements are defined by Hence, the proof of this part is similar to that of the Lemma A2, and then Now, let us proceed to the nonlinear terms. The objective is to demonstrate that, for(88)
To achieve that, we need to decompose the interval into smaller intervals. First, let us consider the intervals for all . We note where and , and we set and We can observe that and(89)
Now, we set the following new classes for , , and , , and Thus, to prove (88), we need to prove that for , and , Notice that for each , and , we have At this stage, we focus on analyzing the aforementioned equation for the case of to simplify the proof (the same reasoning applies for ). Thus, we obtain and(90)
We begin by analyzing the term . Assuming that the sequence of independent blocks has a size of , an application of [119] demonstrates that We retain the choice of and such that , which implies that as . Consequently, the primary term under consideration is the second summand. The central idea is to employ Lemma A19. It is clear that the class of functions is uniformly bounded, i.e., Furthermore, by applying Proposition 2.6 from [10], we obtain the following for every and for Rademacher variables :(91)
where and We see that Given that is a VC-type class of functions that satisfies (C.4.4), it follows that the class is also a VC-type class of functions with similar characteristics to . Therefore,(92)
All the conditions of Lemma A19 are satisfied; thus, a direct application yields, for each ,(93)
where Next, let us study the same blocks . We have Following the same argument as in blocks , we obtain(94)
where andAgain, using Lemma A19, we readily obtain
(95)
(96)
The results for the remaining blocks can be derived by following the same approach as described above. Consequently, we obtain where With this, the proof of the theorem is concluded. □8.2.2. Proof of Theorem 5
We observe that
Given the imposed assumptions and the results derived earlier, and for certain constants and , we obtain that Hence, we can now apply Theorem 3 to handle and Theorems 3 and 4 to handle depending on whether class satisfies (C.4.1) or (C.4.2). As a result, we obtain, for some , with probability 1,Thus, the proof is concluded. □
8.2.3. Proof of Theorem 6
Given Conditions (C.3.1), we have
Considering Hypotheses (H.1), (C.2.1), (C.3.1), and (C.6.), we have and where With this, the proof of the theorem is concluded. □8.2.4. Proof of Corollary 6
In this section, we establish Corollary 6 by applying Lemma A1. Following the same reasoning as in the case of functional regression, we adopt the following notation: , , ,
We choose and such that for all where are increasing sequences that belong to and We denote where We can easily see that, for all(97)
(98)
Using Condition one deetrmines that, for all there exist constants such that We put , and thus, and belong to the interval We denote and ; therefore, We also note, for all and Ultimately, we can choose constants and a sequence while satisfying Condition (C.8) such that , , and It is evident that is fulfilled due to Condition (C.3.1), and from (97) and (98), we can readily verify that the construction of and satisfies Condition . □8.3. Proofs of Weak Convergence Results
8.3.1. Preliminaries of the Proofs
As previously noted, a direct approach proves ineffective when dealing with random bandwidths. Consequently, we often rely on general lemmas (see, for instance, [51]) to facilitate the application of results established for non-random bandwidths. In this section, we introduce key results from [111,112], which were obtained for a positive bandwidth . These results are instrumental in the proofs that follow. We denote the bias term and the centered variate as the following quantities:
(99)
Decomposition (99) plays a pivotal role in our proof. Following the methodology employed in [111], we show that convergence in the quadratic mean to one is achieved and that the bias satisfiesProof of Theorem 11
By applying the Cramér–Wold device, it is sufficient to demonstrate the convergence of the one-dimensional distribution to establish Theorem 11. In fact, leveraging the linearity of is enough to show that
for all of the form Hence, we solely illustrate the convergence in a single dimension. Recall that we are dealing with(100)
We set To obtain the desired result, we write where(101)
and(102)
To derive the desired results, we adopt the strategy of [213].Proof of Theorem 12
In this section, we utilize the same methodology as in [112] and previously in [109]. This approach involves applying the blocking method, which partitions a strictly stationary sequence into equal-sized blocks, each of length . It is important to refer to the notation outlined in the proof of Lemma A2. The primary objective is to establish the asymptotic equicontinuity of the conditional empirical process
Let us introduce, for any and(103)
(104)
Then, we have(105)
where for , we have We investigate the asymptotic equicontinuity of each of the preceding terms. For a given class of functions , we let denote an empirical process based on , indexed by : and for a measurable function and , , we set That implies Again, keeping in mind that when , we establish the asymptotic equi-continuity of which means, for every , that where This approach entails working with the independent block sequence in place of the dependent sequence, facilitated by the results in [119]. Accordingly, we obtain(106)
where is defined by(107)
We choose Note that in our context, is given by the following expression: Utilizing Condition (C.5.1), we have as . Consequently, we focus on the right-hand side term of (106). We begin by assuming that the blocks are independent. Symmetrization is performed using a sequence of i.i.d. Rademacher variables, i.e., random variables satisfying . It is important to note that the sequence is independent of the sequence . Thus, we need to establish the following for all : Again, using the fact that that when and [155], it suffices to show that Since the -conditional moment satisfies (C.4.2), we can truncate and obtain, for each , as ,(108)
(109)
Hence, Then, it suffices to show We have the following: This is achieved by using the chaining method. In [109], , was provided, where(110)
and we let class of measurable functions of There is a map that takes each to its closest function in such that We apply the chaining method(111)
We let be in such a way that(112)
We let r be chosen so small in such a way that Therefore, from (111), we readily infer that By the fact that the terms composing are bounded by , and by applying the Bernstein inequality, we obtain By using (110), we have that means(113)
In view of (112), we assume that . In a similar way, we have Finally, by (110), it suffices to prove, for each , Making use of the square root trick (Lemma 5.2 [214])—see also [215] in a similar way to that in [109]—we obtain(114)
Let us introduce the semi-norm and the covering number defined for any class of functions by By utilizing the latter approach, we can bound (the detailed calculations can be found in [110]). Similarly, as in [110] and earlier in [109], leveraging the independence between the blocks and Condition (C.4.3), we apply Lemma 5.2 from [214] to obtain Consequently, the theorem is established. □8.3.2. Proof of Theorem 14
As previously mentioned, the examination of the weak convergence of the conditional U-process is grounded in the analysis of two components: the truncated part and the remainder part.
Let be a uniformly bounded class of measurable canonical functions from , where . Assume that there exist finite constants and such that the covering number of satisfies
(115)
for every and every probability measure Q. If the mixing coefficient β of the stationary sequence fulfills(116)
for some , then
8.3.3. Proof of Theorem 15
It is well established that the weak convergence of an empirical process can be derived from its finite-dimensional convergence and its asymptotic equicontinuity, provided certain conditions are satisfied. Theorem 14 establishes the finite-dimensional convergence of the conditional U-process . Thus, the remaining task is to demonstrate its asymptotic equicontinuity. We decompose the U-process into two components: the truncated part and the remainder part.
Following the same reasoning as in [155], we also have(117)
Therefore, it suffices to establish the weak convergence of and rather than directly studying . The steps of the proof closely follow those in [112], with the distinction that we now consider a multivariate bandwidth instead of a univariate bandwidth . In this section, we present the proof for , with the same approach applicable to . As shown earlier, the truncated part is decomposed following Hoeffding’s decomposition: We first investigate the linear term . Notice that We can write We need to introduce a new function Hence, The linear term of the process is given by Thus, the linear term of the U-process is an empirical process indexed by the class of functions , defined as follows: Hence, its weak convergence can be established in a manner similar to the proof of Theorem 14. It is evident that . Now, turning to the nonlinear part, we must demonstrate that This is a consequence of Lemma 1. It is important to note that the selection of the number and size of the blocks must be performed in such a way that the terms converge to 0. We need to prove thatAgain, we confine our discussion to the case for clarity. We have
We employ blocking arguments to handle the resulting terms. We begin by examining the first term . We have Notice that (48) readily implies that and recall that for all ,By the symmetry of Function , it holds that
(118)
Employing Chebyshev’s inequality and Hoeffding’s trick while maintaining order, we obtain
(119)
Under (C.6.) we have, for each , which tends to 0 as . Terms , and are handled in the same manner as the first, with the exception that for , we do not need to apply Hoeffding’s trick. This is because our variables (or for ) are in the same blocks. For term , we deduce its study from those of and . Let us consider term . As for the truncated part, we have(120)
We also have Since Equation (118) is still satisfied, the problem is reduced to and we follow the same procedure as in (119). The remaining terms have been shown to be asymptotically negligible. Consequently, process converges in distribution to a Gaussian process, which admits a version with uniformly bounded and uniformly continuous paths with respect to the -norm. By applying the same reasoning, the same result holds for process . Thus, according to (117), it follows that process also converges in distribution to a Gaussian process, which has a version with uniformly bounded and uniformly continuous paths with respect to the -norm. Similarly, process is treated in the same manner, and for , , and , the analysis follows the same steps as in the proof of Theorem 14. □No new data were created or analyzed in this study.
The author extends his sincere gratitude to the Editor-in-Chief, the Associate Editor, and the three referees for their invaluable feedback and for pointing out a number of oversights in the version initially submitted. Their insightful comments have greatly refined and focused the original work, resulting in a markedly improved presentation.
The author declares that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
1. Hoeffding, W. A class of statistics with asymptotically normal distribution. Ann. Math. Stat.; 1948; 19, pp. 293-325. [DOI: https://dx.doi.org/10.1214/aoms/1177730196]
2. Halmos, P.R. The theory of unbiased estimation. Ann. Math. Statist.; 1946; 17, pp. 34-43. [DOI: https://dx.doi.org/10.1214/aoms/1177731020]
3. van der Vaart, A. Asymptotic Statistics; Cambridge Series in Statistical and Probabilistic Mathematics Cambridge University Press: Cambridge, UK, 1998; Volume 3, xvi+443. [DOI: https://dx.doi.org/10.1017/CBO9780511802256]
4. Rubin, H.; Vitale, R.A. Asymptotic distribution of symmetric statistics. Ann. Statist.; 1980; 8, pp. 165-170. [DOI: https://dx.doi.org/10.1214/aos/1176344898]
5. Serfling, R.J. Approximation Theorems of Mathematical Statistics; Wiley Series in Probability and Mathematical Statistics John Wiley Sons, Inc.: New York, NY, USA, 1980; xiv+371.
6. Dynkin, E.B.; Mandelbaum, A. Symmetric statistics, Poisson point processes, and multiple Wiener integrals. Ann. Statist.; 1983; 11, pp. 739-745. [DOI: https://dx.doi.org/10.1214/aos/1176346241]
7. Bretagnolle, J. Lois limites du bootstrap de certaines fonctionnelles. Ann. Inst. H. Poincaré Sect. B; 1983; 19, pp. 281-296.
8. von Mises, R. On the asymptotic distribution of differentiable statistical functions. Ann. Math. Stat.; 1947; 18, pp. 309-348. [DOI: https://dx.doi.org/10.1214/aoms/1177730385]
9. Filippova, A.A. Mises theorem on the limit behaviour of functionals derived from empirical distribution functions. Dokl. Akad. Nauk SSSR; 1959; 129, pp. 44-47.
10. Arcones, M.A.; Giné, E. Limit theorems for U-processes. Ann. Probab.; 1993; 21, pp. 1494-1542. [DOI: https://dx.doi.org/10.1214/aop/1176989128]
11. Lee, A.J. U-Statistics; Statistics: Textbooks and Monographs Marcel Dekker, Inc.: New York, NY, USA, 1990; Volume 110, xii+302.
12. Borovskikh, Y.V. U-Statistics in Banach Spaces; VSP: Utrecht, The Netherlands, 1996; xii+420.
13. Ghosal, S.; Sen, A.; van der Vaart, A.W. Testing monotonicity of regression. Ann. Statist.; 2000; 28, pp. 1054-1082. [DOI: https://dx.doi.org/10.1214/aos/1015956707]
14. Abrevaya, J.; Jiang, W. A nonparametric approach to measuring and testing curvature. J. Bus. Econom. Statist.; 2005; 23, pp. 1-19. [DOI: https://dx.doi.org/10.1198/073500104000000316]
15. Nolan, D.; Pollard, D. U-processes: Rates of convergence. Ann. Statist.; 1987; 15, pp. 780-799. [DOI: https://dx.doi.org/10.1214/aos/1176350374]
16. Sherman, R.P. Maximal inequalities for degenerate U-processes with applications to optimization estimators. Ann. Statist.; 1994; 22, pp. 439-459. [DOI: https://dx.doi.org/10.1214/aos/1176325377]
17. de la Peña, V.H.; Giné, E. From dependence to independence, Randomly stopped processes. U-statistics and processes. Martingales and beyond. Decoupling; Probability and its Applications (New York) Springer: New York, NY, USA, 1999; xvi+392. [DOI: https://dx.doi.org/10.1007/978-1-4612-0537-1]
18. Arcones, M.A.; Wang, Y. Some new tests for normality based on U-processes. Statist. Probab. Lett.; 2006; 76, pp. 69-82. [DOI: https://dx.doi.org/10.1016/j.spl.2005.07.003]
19. Schick, A.; Wang, Y.; Wefelmeyer, W. Tests for normality based on density estimators of convolutions. Statist. Probab. Lett.; 2011; 81, pp. 337-343. [DOI: https://dx.doi.org/10.1016/j.spl.2010.10.022]
20. Nikitin, Y.Y.; Ahsanullah, M. New U-empirical test of symmetry based on extremal order statistics, and their efficiencies. Mathematical Statistics and Limit Theorems; Springer: Cham, Switzerland, 2015; pp. 231-248.
21. Yeo, I.K.; Johnson, R.A. A uniform strong law of large numbers for U-statistics with application to transforming to near symmetry. Statist. Probab. Lett.; 2001; 51, pp. 63-69. [DOI: https://dx.doi.org/10.1016/S0167-7152(00)00143-7]
22. Joly, E.; Lugosi, G. Robust estimation of U-statistics. Stoch. Process. Appl.; 2016; 126, pp. 3760-3773. [DOI: https://dx.doi.org/10.1016/j.spa.2016.04.021]
23. Janson, S. A functional limit theorem for random graphs with applications to subgraph count statistics. Random Struct. Algorithms; 1990; 1, pp. 15-37. [DOI: https://dx.doi.org/10.1002/rsa.3240010103]
24. Frees, E.W. Infinite order U-statistics. Scand. J. Statist.; 1989; 16, pp. 29-45.
25. Heilig, C.; Nolan, D. Limit theorems for the infinite-degree U-process. Statist. Sin.; 2001; 11, pp. 289-302.
26. Song, Y.; Chen, X.; Kato, K. Approximating high-dimensional infinite-order U-statistics: Statistical and computational guarantees. Electron. J. Stat.; 2019; 13, pp. 4794-4848. [DOI: https://dx.doi.org/10.1214/19-EJS1643]
27. Soukarieh, I.; Bouzebda, S. Renewal type bootstrap for increasing degree U-process of a Markov chain. J. Multivar. Anal.; 2023; 195, 105143. [DOI: https://dx.doi.org/10.1016/j.jmva.2022.105143]
28. Peng, W.; Coleman, T.; Mentch, L. Rates of convergence for random forests via generalized U-statistics. Electron. J. Stat.; 2022; 16, pp. 232-292. [DOI: https://dx.doi.org/10.1214/21-EJS1958]
29. Faivishevsky, L.; Goldberger, J. ICA based on a Smooth Estimation of the Differential Entropy. Proceedings of the Advances in Neural Information Processing Systems; Koller, D.; Schuurmans, D.; Bengio, Y.; Bottou, L. Curran Associates, Inc.: New York, NY, USA, 2008; Volume 21.
30. Liu, Q.; Lee, J.; Jordan, M. A Kernelized Stein Discrepancy for Goodness-of-fit Tests. Proceedings of the 33rd International Conference on Machine Learning; New York, NY, USA, 20–22 June 2016; Balcan, M.F.; Weinberger, K.Q. Volume 48, pp. 276-284.
31. Cybis, G.B.; Valk, M.; Lopes, S.R.C. Clustering and classification problems in genetics through U-statistics. J. Stat. Comput. Simul.; 2018; 88, pp. 1882-1902. [DOI: https://dx.doi.org/10.1080/00949655.2017.1374387]
32. Lim, F.; Stojanovic, V.M. On U-statistics and compressed sensing I: Non-asymptotic average-case analysis. IEEE Trans. Signal Process.; 2013; 61, pp. 2473-2485. [DOI: https://dx.doi.org/10.1109/TSP.2013.2247598]
33. Jadhav, S.; Ma, S. An association test for functional data based on Kendall’s tau. J. Multivar. Anal.; 2021; 184, 104740. [DOI: https://dx.doi.org/10.1016/j.jmva.2021.104740]
34. Bello, D.Z.; Valk, M.; Cybis, G.B. Towards U-statistics clustering inference for multiple groups. J. Stat. Comput. Simul.; 2024; 94, pp. 204-222. [DOI: https://dx.doi.org/10.1080/00949655.2023.2239978]
35. Kim, I.; Ramdas, A. Dimension-agnostic inference using cross U-statistics. Bernoulli; 2024; 30, pp. 683-711. [DOI: https://dx.doi.org/10.3150/23-BEJ1613]
36. Chen, L.; Wan, A.T.K.; Zhang, S.; Zhou, Y. Distributed algorithms for U-statistics-based empirical risk minimization. J. Mach. Learn. Res.; 2023; 24, pp. 1-43.
37. Janson, S. Asymptotic normality for m-dependent and constrained U-statistics, with applications to pattern matching in random strings and permutations. Adv. Appl. Probab.; 2023; 55, pp. 841-894. [DOI: https://dx.doi.org/10.1017/apr.2022.51]
38. Sudheesh, K.K.; Anjana, S.; Xie, M. U-statistics for left truncated and right censored data. Statistics; 2023; 57, pp. 900-917. [DOI: https://dx.doi.org/10.1080/02331888.2023.2217314]
39. Le Minh, T. U-statistics on bipartite exchangeable networks. ESAIM Probab. Stat.; 2023; 27, pp. 576-620. [DOI: https://dx.doi.org/10.1051/ps/2023010]
40. Nadaraja, E.A. On a regression estimate. Teor. Verojatnost. I Primenen.; 1964; 9, pp. 157-159.
41. Watson, G.S. Smooth regression analysis. Sankhyā Ser. A; 1964; 26, pp. 359-372.
42. Bouzebda, S.; Taachouche, N. Rates of the Strong Uniform Consistency for the Kernel-Type Regression Function Estimators with General Kernels on Manifolds. Math. Methods Statist.; 2023; 32, pp. 27-80. [DOI: https://dx.doi.org/10.3103/S1066530723010027]
43. Bouzebda, S.; Taachouche, N. Rates of the strong uniform consistency with rates for conditional U-statistics estimators with general kernels on manifolds. Math. Methods Statist.; 2024; 33, pp. 95-153. [DOI: https://dx.doi.org/10.3103/S1066530724700066]
44. Sen, A. Uniform strong consistency rates for conditional U-statistics. Sankhyā Ser. A; 1994; 56, pp. 179-194.
45. Prakasa Rao, B.L.S.; Sen, A. Limit distributions of conditional U-statistics. J. Theoret. Probab.; 1995; 8, pp. 261-301. [DOI: https://dx.doi.org/10.1007/BF02212880]
46. Harel, M.; Puri, M.L. Conditional U-statistics for dependent random variables. J. Multivar. Anal.; 1996; 57, pp. 84-100. [DOI: https://dx.doi.org/10.1006/jmva.1996.0023]
47. Stute, W. Symmetrized NN-conditional U-statistics. Research Developments in Probability and Statistics; VSP: Utrecht, The Netherlands, 1996; pp. 231-237.
48. Fu, K.A. An application of U-statistics to nonparametric functional data analysis. Comm. Statist. Theory Methods; 2012; 41, pp. 1532-1542. [DOI: https://dx.doi.org/10.1080/03610926.2010.526747]
49. Bouzebda, S.; Nemouchi, B. Uniform consistency and uniform in bandwidth consistency for nonparametric regression estimates and conditional U-statistics involving functional data. J. Nonparametr. Stat.; 2020; 32, pp. 452-509. [DOI: https://dx.doi.org/10.1080/10485252.2020.1759597]
50. Bouzebda, S.; Elhattab, I.; Nemouchi, B. On the uniform-in-bandwidth consistency of the general conditional U-statistics based on the copula representation. J. Nonparametr. Stat.; 2021; 33, pp. 321-358. [DOI: https://dx.doi.org/10.1080/10485252.2021.1937621]
51. Bouzebda, S.; Nezzal, A. Uniform consistency and uniform in number of neighbors consistency for nonparametric regression estimates and conditional U-statistics involving functional data. Jpn. J. Stat. Data Sci.; 2022; 5, pp. 431-533. [DOI: https://dx.doi.org/10.1007/s42081-022-00161-3]
52. Bouzebda, S.; Nezzal, A. Asymptotic properties of conditional U-statistics using delta sequences. Comm. Statist. Theory Methods; 2024; 53, pp. 4602-4657. [DOI: https://dx.doi.org/10.1080/03610926.2023.2179887]
53. Bouzebda, S.; Nezzal, A.; Zari, T. Uniform Consistency for Functional Conditional U-Statistics Using Delta-Sequences. Mathematics; 2023; 11, 161. [DOI: https://dx.doi.org/10.3390/math11010161]
54. Ramsay, J.O.; Silverman, B.W. Functional Data Analysis; 2nd ed. Springer Series in Statistics Springer: New York, NY, USA, 2005; xx+426.
55. Ramsay, J.O.; Silverman, B.W. Methods and case studies. Applied Functional Data Analysis; Springer Series in Statistic Springer: New York, NY, USA, 2002; x+190. [DOI: https://dx.doi.org/10.1007/b98886]
56. Ferraty, F.; Vieu, P. Theory and practice. Nonparametric Functional Data Analysis; Springer Series in Statistics Springer: New York, NY, USA, 2006; xx+258.
57. Araujo, A.; Giné, E. The Central Limit Theorem for Real and Banach Valued Random Variables; Wiley Series in Probability and Mathematical Statistics John Wiley & Sons: New York, NY, USA, Chichester, UK, Brisbane, Australia, 1980; xiv+233.
58. Gasser, T.; Hall, P.; Presnell, B. Nonparametric estimation of the mode of a distribution of random curves. J. R. Stat. Soc. Ser. B Stat. Methodol.; 1998; 60, pp. 681-691. [DOI: https://dx.doi.org/10.1111/1467-9868.00148]
59. Bosq, D. Lecture Notes in Statistics. Linear Processes in Function Spaces: Theory and Applications; Springer: New York, NY, USA, 2000; Volume 149, xiv+283. [DOI: https://dx.doi.org/10.1007/978-1-4612-1154-9]
60. Horváth, L.; Kokoszka, P. Inference for Functional Data with Applications; Springer Series in Statistics Springer: New York, NY, USA, 2012; xiv+422. [DOI: https://dx.doi.org/10.1007/978-1-4614-3655-3]
61. Kokoszka, P.; Reimherr, M. Introduction to Functional Data Analysis; Texts in Statistical Science Series CRC Press: Boca Raton, FL, USA, 2017; xvi+290.
62. Ferraty, F.; Laksaci, A.; Tadj, A.; Vieu, P. Rate of uniform consistency for nonparametric estimates with functional variables. J. Statist. Plann. Inference; 2010; 140, pp. 335-352. [DOI: https://dx.doi.org/10.1016/j.jspi.2009.07.019]
63. Kara-Zaitri, L.; Laksaci, A.; Rachdi, M.; Vieu, P. Uniform in bandwidth consistency for various kernel estimators involving functional data. J. Nonparametr. Stat.; 2017; 29, pp. 85-107. [DOI: https://dx.doi.org/10.1080/10485252.2016.1254780]
64. Attouch, M.; Laksaci, A.; Rafaa, F. On the local linear estimate for functional regression: Uniform in bandwidth consistency. Comm. Statist. Theory Methods; 2019; 48, pp. 1836-1853. [DOI: https://dx.doi.org/10.1080/03610926.2018.1440308]
65. Ling, N.; Meng, S.; Vieu, P. Uniform consistency rate of kNN regression estimation for functional time series data. J. Nonparametr. Stat.; 2019; 31, pp. 451-468. [DOI: https://dx.doi.org/10.1080/10485252.2019.1583338]
66. Mohammedi, M.; Bouzebda, S.; Laksaci, A. The consistency and asymptotic normality of the kernel type expectile regression estimator for functional data. J. Multivar. Anal.; 2021; 181, 104673. [DOI: https://dx.doi.org/10.1016/j.jmva.2020.104673]
67. Almanjahie, I.M.; Bouzebda, S.; Kaid, Z.; Laksaci, A. Nonparametric estimation of expectile regression in functional dependent data. J. Nonparametr. Stat.; 2022; 34, pp. 250-281. [DOI: https://dx.doi.org/10.1080/10485252.2022.2027412]
68. Almanjahie, I.M.; Bouzebda, S.; Chikr Elmezouar, Z.; Laksaci, A. The functional kNN estimator of the conditional expectile: Uniform consistency in number of neighbors. Stat. Risk Model.; 2022; 38, pp. 47-63. [DOI: https://dx.doi.org/10.1515/strm-2019-0029]
69. Bouzebda, S.; Laksaci, A.; Mohammedi, M. Single index regression model for functional quasi-associated time series data. Revstat; 2022; 20, pp. 605-631.
70. Bouzebda, S.; Laksaci, A.; Mohammedi, M. The k-nearest neighbors method in single index regression model for functional quasi-associated time series data. Rev. Mat. Complut.; 2023; 36, pp. 361-391. [DOI: https://dx.doi.org/10.1007/s13163-022-00436-z]
71. Almanjahie, I.M.; Bouzebda, S.; Kaid, Z.; Laksaci, A. The local linear functional kNN estimator of the conditional expectile: Uniform consistency in number of neighbors. Metrika; 2024; 87, pp. 1007-1035. [DOI: https://dx.doi.org/10.1007/s00184-023-00942-0]
72. Härdle, W.; Hall, P.; Ichimura, H. Optimal smoothing in single-index models. Ann. Statist.; 1993; 21, pp. 157-178. [DOI: https://dx.doi.org/10.1214/aos/1176349020]
73. Bhattacharjee, S.; Müller, H.G. Single index Fréchet regression. Ann. Statist.; 2023; 51, pp. 1770-1798. [DOI: https://dx.doi.org/10.1214/23-AOS2307]
74. Liang, H.; Liu, X.; Li, R.; Tsai, C.L. Estimation and testing for partially linear single-index models. Ann. Statist.; 2010; 38, pp. 3811-3836. [DOI: https://dx.doi.org/10.1214/10-AOS835]
75. Stute, W.; Zhu, L.X. Nonparametric checks for single-index models. Ann. Statist.; 2005; 33, pp. 1048-1083. [DOI: https://dx.doi.org/10.1214/009053605000000020]
76. Gu, L.; Yang, L. Oracally efficient estimation for single-index link function with simultaneous confidence band. Electron. J. Stat.; 2015; 9, pp. 1540-1561. [DOI: https://dx.doi.org/10.1214/15-EJS1051]
77. Cuevas, A. A partial overview of the theory of statistics with functional data. J. Statist. Plann. Inference; 2014; 147, pp. 1-23. [DOI: https://dx.doi.org/10.1016/j.jspi.2013.04.002]
78. Goia, A.; Vieu, P. An introduction to recent advances in high/infinite dimensional statistics [Editorial]. J. Multivar. Anal.; 2016; 146, pp. 1-6. [DOI: https://dx.doi.org/10.1016/j.jmva.2015.12.001]
79. Ling, N.; Vieu, P. Nonparametric modelling for functional data: Selected survey and tracks for future. Statistics; 2018; 52, pp. 934-949. [DOI: https://dx.doi.org/10.1080/02331888.2018.1487120]
80. Ferraty, F.; Peuch, A.; Vieu, P. Modèle à indice fonctionnel simple. C. R. Math. Acad. Sci. Paris; 2003; 336, pp. 1025-1028. [DOI: https://dx.doi.org/10.1016/S1631-073X(03)00239-5]
81. Ait-Saïdi, A.; Ferraty, F.; Kassa, R.; Vieu, P. Cross-validated estimations in the single-functional index model. Statistics; 2008; 42, pp. 475-494. [DOI: https://dx.doi.org/10.1080/02331880801980377]
82. Attaoui, S.; Bentat, B.; Bouzebda, S.; Laksaci, A. The strong consistency and asymptotic normality of the kernel estimator type in functional single index model in presence of censored data. AIMS Math.; 2024; 9, pp. 7340-7371. [DOI: https://dx.doi.org/10.3934/math.2024356]
83. Jiang, Z.; Huang, Z.; Zhang, J. Functional single-index composite quantile regression. Metrika; 2023; 86, pp. 595-603. [DOI: https://dx.doi.org/10.1007/s00184-022-00887-w]
84. Nie, Y.; Wang, L.; Cao, J. Estimating functional single index models with compact support. Environmetrics; 2023; 34, e2784. [DOI: https://dx.doi.org/10.1002/env.2784]
85. Zhu, H.; Zhang, R.; Liu, Y.; Ding, H. Robust estimation for a general functional single index model via quantile regression. J. Korean Statist. Soc.; 2022; 51, pp. 1041-1070. [DOI: https://dx.doi.org/10.1007/s42952-022-00174-4]
86. Tang, Q.; Kong, L.; Rupper, D.; Karunamuni, R.J. Partial functional partially linear single-index models. Statist. Sin.; 2021; 31, pp. 107-133. [DOI: https://dx.doi.org/10.5705/ss.202018.0316]
87. Ling, N.; Cheng, L.; Vieu, P.; Ding, H. Missing responses at random in functional single index model for time series data. Statist. Pap.; 2022; 63, pp. 665-692. [DOI: https://dx.doi.org/10.1007/s00362-021-01251-2]
88. Ling, N.; Cheng, L.; Vieu, P. Single functional index model under responses MAR and dependent observations. Functional and High-Dimensional Statistics and Related Fields; Contrib. Stat. Springer: Cham, Switzerland, 2020; pp. 161-168. [DOI: https://dx.doi.org/10.1007/978-3-030-47756-1_22]
89. Feng, S.; Tian, P.; Hu, Y.; Li, G. Estimation in functional single-index varying coefficient model. J. Statist. Plann. Inference; 2021; 214, pp. 62-75. [DOI: https://dx.doi.org/10.1016/j.jspi.2021.01.003]
90. Novo, S.; Aneiros, G.; Vieu, P. Automatic and location-adaptive estimation in functional single-index regression. J. Nonparametr. Stat.; 2019; 31, pp. 364-392. [DOI: https://dx.doi.org/10.1080/10485252.2019.1567726]
91. Li, J.; Huang, C.; Zhu, H. A functional varying-coefficient single-index model for functional response data. J. Amer. Statist. Assoc.; 2017; 112, pp. 1169-1181. [DOI: https://dx.doi.org/10.1080/01621459.2016.1195742] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/29200540]
92. Attaoui, S.; Ling, N. Asymptotic results of a nonparametric conditional cumulative distribution estimator in the single functional index modeling for time series data with applications. Metrika; 2016; 79, pp. 485-511. [DOI: https://dx.doi.org/10.1007/s00184-015-0564-6]
93. Chen, D.; Hall, P.; Müller, H.G. Single and multiple index functional regression models with nonparametric link. Ann. Statist.; 2011; 39, pp. 1720-1747. [DOI: https://dx.doi.org/10.1214/11-AOS882]
94. Fix, E.; Hodges, J.L.J. Discriminatory Analysis—Nonparametric Discrimination: Consistency Properties; Technical Report Project 21-49-004, Report 4 USAF School of Aviation Medicine, Randolph Field: San Antonio, TX, USA, 1951.
95. Fix, E.; Hodges, J.L.J. Discriminatory analysis–nonparametric discrimination: Consistency properties. Int. Stat. Rev.; 1989; 57, pp. 238-247. [DOI: https://dx.doi.org/10.2307/1403797]
96. Loftsgaarden, D.O.; Quesenberry, C.P. A nonparametric estimate of a multivariate density function. Ann. Math. Statist.; 1965; 36, pp. 1049-1051. [DOI: https://dx.doi.org/10.1214/aoms/1177700079]
97. Biau, G.; Devroye, L. Lectures on the Nearest Neighbor Method; Springer Series in the Data Sciences Springer: Cham, Switzerland, 2015; ix+290. [DOI: https://dx.doi.org/10.1007/978-3-319-25388-6]
98. Collomb, G. Estimation de la régression par la méthode des k points les plus proches avec noyau: Quelques propriétés de convergence ponctuelle. Proceedings of the Nonparametric Asymptotic Statistics (Proc. Conf., Rouen, 1979) (French); Lecture Notes in Math Springer: Berlin/Heidelberg, Germany, 1980; Volume 821, pp. 159-175.
99. Mack, Y.P. Local Properties of k-NN Regression Estimates. SIAM J. Algebr. Discret. Methods; 1981; 2, pp. 311-323. [DOI: https://dx.doi.org/10.1137/0602035]
100. Györfi, L. The rate of convergence of k-NN regression estimation and classification. IEEE Trans. Inform. Theory.; 1981; 27, pp. 500-509. [DOI: https://dx.doi.org/10.1109/TIT.1981.1056344]
101. Bhattacharya, P.K.; Mack, Y.P. Weak convergence of k-NN density and regression estimators with varying k and applications. Ann. Statist.; 1987; 15, pp. 976-994. [DOI: https://dx.doi.org/10.1214/aos/1176350487]
102. Devroye, L.; Györfi, L.; Krzyzak, A.; Lugosi, G. On the Strong Universal Consistency of Nearest Neighbor Regression Function Estimates. Ann. Statist.; 1994; 22, pp. 1371-1385. [DOI: https://dx.doi.org/10.1214/aos/1176325633]
103. Laloë, T. A k-nearest neighbor approach for functional regression. Statist. Probab. Lett.; 2008; 78, pp. 1189-1193. [DOI: https://dx.doi.org/10.1016/j.spl.2007.11.014]
104. Chikr-Elmezouar, Z.; Almanjahie, I.M.; Laksaci, A.; Rachdi, M. FDA: Strong consistency of the kNN local linear estimation of the functional conditional density and mode. J. Nonparametr. Stat.; 2019; 31, pp. 175-195. [DOI: https://dx.doi.org/10.1080/10485252.2018.1538450]
105. Ling, N.; Aneiros, G.; Vieu, P. kNN estimation in functional partial linear modeling. Statist. Pap.; 2020; 61, pp. 423-444. [DOI: https://dx.doi.org/10.1007/s00362-017-0946-0]
106. Einmahl, U.; Mason, D.M. Uniform in bandwidth consistency of kernel-type function estimators. Ann. Statist.; 2005; 33, pp. 1380-1403. [DOI: https://dx.doi.org/10.1214/009053605000000129]
107. Dony, J.; Einmahl, U. Uniform in bandwidth consistency of kernel regression estimators at a fixed point. High Dimensional Probability V: The Luminy Volume; Institute of Mathematical Statistics: Beachwood, OH, USA, 2009; Volume 5, pp. 308-325. [DOI: https://dx.doi.org/10.1214/09-IMSCOLL520]
108. Kara, L.Z.; Laksaci, A.; Rachdi, M.; Vieu, P. Data-driven kNN estimation in nonparametric functional data analysis. J. Multivar. Anal.; 2017; 153, pp. 176-188. [DOI: https://dx.doi.org/10.1016/j.jmva.2016.09.016]
109. Arcones, M.A.; Yu, B. Central limit theorems for empirical and U-processes of stationary mixing sequences. J. Theoret. Probab.; 1994; 7, pp. 47-71. [DOI: https://dx.doi.org/10.1007/BF02213360]
110. Bouzebda, S.; Nemouchi, B. Central limit theorems for conditional empirical and conditional U-processes of stationary mixing sequences. Math. Methods Statist.; 2019; 28, pp. 169-207. [DOI: https://dx.doi.org/10.3103/S1066530719030013]
111. Masry, E. Nonparametric regression estimation for dependent functional data: Asymptotic normality. Stoch. Process. Appl.; 2005; 115, pp. 155-177. [DOI: https://dx.doi.org/10.1016/j.spa.2004.07.006]
112. Bouzebda, S.; Nemouchi, B. Weak-convergence of empirical conditional processes and conditional U-processes involving functional mixing data. Stat. Inference Stoch. Process.; 2023; 26, pp. 33-88. [DOI: https://dx.doi.org/10.1007/s11203-022-09276-6]
113. Hristache, M.; Juditsky, A.; Spokoiny, V. Direct estimation of the index coefficient in a single-index model. Ann. Statist.; 2001; 29, pp. 595-623. [DOI: https://dx.doi.org/10.1214/aos/1009210682]
114. Ferraty, F.; Vieu, P. The functional nonparametric model and application to spectrometric data. Comput. Statist.; 2002; 17, pp. 545-564. [DOI: https://dx.doi.org/10.1007/s001800200126]
115. Ferraty, F.; Park, J.; Vieu, P. Estimation of a functional single index model. Recent Advances in Functional Data Analysis and Related Topics; Contrib. Statist. Physica-Verlag/Springer: Berlin/Heidelberg, Germany, 2011; pp. 111-116. [DOI: https://dx.doi.org/10.1007/978-3-7908-2736-1_17]
116. Geenens, G. Curse of dimensionality and related issues in nonparametric functional regression. Stat. Surv.; 2011; 5, pp. 30-43. [DOI: https://dx.doi.org/10.1214/09-SS049]
117. Attouch, M.; Laksaci, A.; Rafaa, F. Estimation locale linéaire de la régression non paramétrique fonctionnelle par la méthode des k plus proches voisins. C. R. Math. Acad. Sci. Paris; 2017; 355, pp. 824-829. [DOI: https://dx.doi.org/10.1016/j.crma.2017.05.007]
118. Burba, F.; Ferraty, F.; Vieu, P. k-nearest neighbour method in functional nonparametric regression. J. Nonparametr. Stat.; 2009; 21, pp. 453-469. [DOI: https://dx.doi.org/10.1080/10485250802668909]
119. Eberlein, E. Weak convergence of partial sums of absolutely regular sequences. Statist. Probab. Lett.; 1984; 2, pp. 291-293. [DOI: https://dx.doi.org/10.1016/0167-7152(84)90067-1]
120. Volkonski, V.; Rozanov, Y. Some limit theorems for random functions, Part I. Teor. Veroyatn. Primen. 4 186–207. Engl. Transl. Theory Probab. Appl.; 1959; 4, pp. 178-197. [DOI: https://dx.doi.org/10.1137/1104015]
121. Rosenblatt, M. A central limit theorem and a strong mixing condition. Proc. Natl. Acad. Sci. USA; 1956; 42, pp. 43-47. [DOI: https://dx.doi.org/10.1073/pnas.42.1.43]
122. Davydov, J.A. Mixing conditions for Markov chains. Teor. Verojatnost. I Primenen.; 1973; 18, pp. 321-338. [DOI: https://dx.doi.org/10.1137/1118033]
123. Bolthausen, E. The Berry-Esseen theorem for functionals of discrete Markov chains. Z. Wahrsch. Verw. Geb.; 1980; 54, pp. 59-73. [DOI: https://dx.doi.org/10.1007/BF00535354]
124. Bouzebda, S.; Soukarieh, I. Renewal type bootstrap for U-process Markov chains. Markov Process. Relat. Fields; 2022; 28, pp. 673-735.
125. Akaike, H. An approximation to the density function. Ann. Inst. Statist. Math. Tokyo; 1954; 6, pp. 127-132. [DOI: https://dx.doi.org/10.1007/BF02900741]
126. Parzen, E. On estimation of a probability density function and mode. Ann. Math. Statist.; 1962; 33, pp. 1065-1076. [DOI: https://dx.doi.org/10.1214/aoms/1177704472]
127. Rosenblatt, M. Remarks on some nonparametric estimates of a density function. Ann. Math. Statist.; 1956; 27, pp. 832-837. [DOI: https://dx.doi.org/10.1214/aoms/1177728190]
128. Devroye, L. A Course in Density Estimation; Progress in Probability and Statistics Birkhäuser Boston Inc.: Boston, MA, USA, 1987; Volume 14, xx+183.
129. Silverman, B.W. Density Estimation for Statistics and Data Analysis; Monographs on Statistics and Applied Probability Chapman & Hall: London, UK, 1986; x+175. [DOI: https://dx.doi.org/10.1007/978-1-4899-3324-9]
130. Bouzebda, S. General tests of conditional independence based on empirical processes indexed by functions. Jpn. J. Stat. Data Sci.; 2023; 6, pp. 115-177. [DOI: https://dx.doi.org/10.1007/s42081-023-00193-3]
131. Bouzebda, S. On the weak convergence and the uniform-in-bandwidth consistency of the general conditional U-processes based on the copula representation: Multivariate setting. Hacet. J. Math. Stat.; 2023; 52, pp. 1303-1348. [DOI: https://dx.doi.org/10.15672/hujms.1134334]
132. Bouzebda, T.; Taachouche, N. Oracle inequalities and upper bounds for kernel conditional U-statistics estimators on manifolds and more general metric spaces associated with operators. Stochastics; 2024; pp. 1-64. [DOI: https://dx.doi.org/10.1080/17442508.2024.2391898]
133. Bouzebda, S. Limit Theorems in the Nonparametric Conditional Single-Index U-Processes for Locally Stationary Functional Random Fields under Stochastic Sampling Design. Mathematics; 2024; 12, 1996. [DOI: https://dx.doi.org/10.3390/math12131996]
134. Vapnik, V.N.; Chervonenkis, A.J. The uniform convergence of frequencies of the appearance of events to their probabilities. Teor. Verojatnost. I Primenen.; 1971; 16, pp. 264-279.
135. Kolmogorov, A.N.; Tikhomirov, V.M. ε-entropy and ε-capacity of sets in function spaces. Uspekhi Mat. Nauk; 1959; 14, pp. 3-86.
136. Dudley, R.M. The sizes of compact subsets of Hilbert space and continuity of Gaussian processes. J. Funct. Anal.; 1967; 1, pp. 290-330. [DOI: https://dx.doi.org/10.1016/0022-1236(67)90017-1]
137. Mayer-Wolf, E.; Zeitouni, O. The probability of small Gaussian ellipsoids and associated conditional moments. Ann. Probab.; 1993; 21, pp. 14-24. [DOI: https://dx.doi.org/10.1214/aop/1176989391]
138. Bogachev, V.I. Gaussian Measures; Mathematical Surveys and Monographs American Mathematical Society: Providence, RI, USA, 1998; Volume 62, xii+433. [DOI: https://dx.doi.org/10.1090/surv/062]
139. Li, W.V.; Shao, Q.M. Gaussian processes: Inequalities, small ball probabilities and applications. Stochastic Processes: Theory and Methods; North-Holland: Amsterdam, The Netherlands, 2001; Volume 19, pp. 533-597. [DOI: https://dx.doi.org/10.1016/S0169-7161(01)19019-X]
140. Ferraty, F.; Mas, A.; Vieu, P. Nonparametric regression on functional data: Inference and practical aspects. Aust. N. Z. J. Stat.; 2007; 49, pp. 267-286. [DOI: https://dx.doi.org/10.1111/j.1467-842X.2007.00480.x]
141. Deheuvels, P. One bootstrap suffices to generate sharp uniform bounds in functional estimation. Kybernetika; 2011; 47, pp. 855-865.
142. Soukarieh, I.; Bouzebda, S. Weak convergence of the conditional U-statistics for locally stationary functional time series. Stat. Inference Stoch. Process.; 2024; 27, pp. 227-304. [DOI: https://dx.doi.org/10.1007/s11203-023-09305-y]
143. Park, H.; Stefanski, L.A. Relative-error prediction. Statist. Probab. Lett.; 1998; 40, pp. 227-236. [DOI: https://dx.doi.org/10.1016/S0167-7152(98)00088-1]
144. Jones, M.C.; Park, H.; Shin, K.I.; Vines, S.K.; Jeong, S.O. Relative error prediction via kernel regression smoothers. J. Statist. Plann. Inference; 2008; 138, pp. 2887-2898. [DOI: https://dx.doi.org/10.1016/j.jspi.2007.11.001]
145. Demongeot, J.; Hamie, A.; Laksaci, A.; Rachdi, M. Relative-error prediction in nonparametric functional statistics: Theory and practice. J. Multivar. Anal.; 2016; 146, pp. 261-268. [DOI: https://dx.doi.org/10.1016/j.jmva.2015.09.019]
146. Bouhadjera, F.; Lemdani, M.; Ould Saïd, E. Strong uniform consistency of the local linear relative error regression estimator under left truncation. Statist. Pap.; 2023; 64, pp. 421-447. [DOI: https://dx.doi.org/10.1007/s00362-022-01325-9]
147. Bouhadjera, F.; Ould Saïd, E. Strong consistency of the local linear relative regression estimator for censored data. Opusc. Math.; 2022; 42, pp. 805-832. [DOI: https://dx.doi.org/10.7494/OpMath.2022.42.6.805]
148. Dehling, H.; Wendler, M. Central limit theorem and the bootstrap for U-statistics of strongly mixing data. J. Multivar. Anal.; 2010; 101, pp. 126-137. [DOI: https://dx.doi.org/10.1016/j.jmva.2009.06.002]
149. Yoshihara, K.I. Limiting behavior of U-statistics for stationary, absolutely regular processes. Z. Wahrscheinlichkeitstheorie Und Verw. Geb.; 1976; 35, pp. 237-252. [DOI: https://dx.doi.org/10.1007/BF00532676]
150. Han, F. An exponential inequality for U-statistics under mixing conditions. J. Theoret. Probab.; 2018; 31, pp. 556-578. [DOI: https://dx.doi.org/10.1007/s10959-016-0722-4]
151. Merlevède, F.; Peligrad, M.; Rio, E. A Bernstein type inequality and moderate deviations for weakly dependent sequences. Probab. Theory Relat. Fields; 2011; 151, pp. 435-474. [DOI: https://dx.doi.org/10.1007/s00440-010-0304-9]
152. Kudraszow, N.L.; Vieu, P. Uniform consistency of kNN regressors for functional variables. Statist. Probab. Lett.; 2013; 83, pp. 1863-1870. [DOI: https://dx.doi.org/10.1016/j.spl.2013.04.017]
153. Arcones, M.A. A Bernstein-type inequality for U-statistics and U-processes. Statist. Probab. Lett.; 1995; 22, pp. 239-247. [DOI: https://dx.doi.org/10.1016/0167-7152(94)00072-G]
154. de la Peña, V.H. Decoupling and Khintchine’s inequalities for U-statistics. Ann. Probab.; 1992; 20, pp. 1877-1892. [DOI: https://dx.doi.org/10.1214/aop/1176989533]
155. Bouzebda, S.; Nezzal, A. Uniform in number of neighbors consistency and weak convergence of kNN empirical conditional processes and kNN conditional U-processes involving functional mixing data. AIMS Math.; 2024; 9, pp. 4427-4550. [DOI: https://dx.doi.org/10.3934/math.2024218]
156. Mason, D.M. Proving consistency of non-standard kernel estimators. Stat. Inference Stoch. Process.; 2012; 15, pp. 151-176. [DOI: https://dx.doi.org/10.1007/s11203-012-9068-4]
157. Novo, S.; Aneiros, G.; Vieu, P. A kNN procedure in semiparametric functional data analysis. Statist. Probab. Lett.; 2021; 171, 109028. [DOI: https://dx.doi.org/10.1016/j.spl.2020.109028]
158. Dudley, R.M. A course on empirical processes. École D’été de Probabilités de Saint-Flour, XII—1982; Springer: Berlin/Heidelberg, Germany, 1984; Volume 1097, pp. 1-142. [DOI: https://dx.doi.org/10.1007/BFb0099432]
159. Polonik, W.; Yao, Q. Set-indexed conditional empirical and quantile processes based on dependent data. J. Multivar. Anal.; 2002; 80, pp. 234-255. [DOI: https://dx.doi.org/10.1006/jmva.2001.1988]
160. Kendall, M.G. A New Measure of Rank Correlation. Biometrika; 1938; 30, pp. 81-93. [DOI: https://dx.doi.org/10.1093/biomet/30.1-2.81]
161. Stute, W. Universally consistent conditional U-statistics. Ann. Statist.; 1994; 22, pp. 460-473. [DOI: https://dx.doi.org/10.1214/aos/1176325378]
162. Stute, W. Lp-convergence of conditional U-statistics. J. Multivar. Anal.; 1994; 51, pp. 71-82. [DOI: https://dx.doi.org/10.1006/jmva.1994.1050]
163. Lehmann, E.L. A general concept of unbiasedness. Ann. Math. Stat.; 1951; 22, pp. 587-592. [DOI: https://dx.doi.org/10.1214/aoms/1177729549]
164. Dwass, M. The large-sample power of rank order tests in the two-sample problem. Ann. Math. Statist.; 1956; 27, pp. 352-374. [DOI: https://dx.doi.org/10.1214/aoms/1177728263]
165. Kohler, M.; Máthé, K.; Pintér, M. Prediction from randomly right censored data. J. Multivar. Anal.; 2002; 80, pp. 73-100. [DOI: https://dx.doi.org/10.1006/jmva.2000.1973]
166. Carbonez, A.; Györfi, L.; van der Meulen, E.C. Partitioning-estimates of a regression function under random censoring. Statist. Decis.; 1995; 13, pp. 21-37. [DOI: https://dx.doi.org/10.1524/strm.1995.13.1.21]
167. Kaplan, E.L.; Meier, P. Nonparametric estimation from incomplete observations. J. Am. Statist. Assoc.; 1958; 53, pp. 457-481. [DOI: https://dx.doi.org/10.1080/01621459.1958.10501452]
168. Maillot, B.; Viallon, V. Uniform limit laws of the logarithm for nonparametric estimators of the regression function in presence of censored data. Math. Methods Statist.; 2009; 18, pp. 159-184. [DOI: https://dx.doi.org/10.3103/S1066530709020045]
169. Datta, S.; Bandyopadhyay, D.; Satten, G.A. Inverse probability of censoring weighted U-statistics for right-censored data with an application to testing hypotheses. Scand. J. Stat.; 2010; 37, pp. 680-700. [DOI: https://dx.doi.org/10.1111/j.1467-9469.2010.00697.x]
170. Stute, W.; Wang, J.L. Multi-sample U-statistics for censored data. Scand. J. Statist.; 1993; 20, pp. 369-374.
171. Chen, Y.; Datta, S. Adjustments of multi-sample U-statistics to right censored data and confounding covariates. Comput. Statist. Data Anal.; 2019; 135, pp. 1-14. [DOI: https://dx.doi.org/10.1016/j.csda.2019.01.012]
172. Yuan, A.; Giurcanu, M.; Luta, G.; Tan, M.T. U-statistics with conditional kernels for incomplete data models. Ann. Inst. Statist. Math.; 2017; 69, pp. 271-302. [DOI: https://dx.doi.org/10.1007/s10463-015-0537-6]
173. Földes, A.; Rejto, L. A LIL type result for the product limit estimator. Z. Wahrsch. Verw. Geb.; 1981; 56, pp. 75-86. [DOI: https://dx.doi.org/10.1007/BF00531975]
174. Tsai, W.Y.; Jewell, N.P.; Wang, M.C. A note on the product-limit estimator under right censoring and left truncation. Biometrika; 1987; 74, pp. 883-886. [DOI: https://dx.doi.org/10.1093/biomet/74.4.883]
175. Andersen, P.K.; Borgan, O.r.; Gill, R.D.; Keiding, N. Statistical Models Based on Counting Processes; Springer Series in Statistics Springer: New York, NY, USA, 1993; xii+767. [DOI: https://dx.doi.org/10.1007/978-1-4612-4348-9]
176. Zhou, Y.; Yip, P.S.F. A strong representation of the product-limit estimator for left truncated and right censored data. J. Multivar. Anal.; 1999; 69, pp. 261-280. [DOI: https://dx.doi.org/10.1006/jmva.1998.1806]
177. Satten, G.A.; Kong, M.; Datta, S. Multisample adjusted U-statistics that account for confounding covariates. Stat. Med.; 2018; 37, pp. 3357-3372. [DOI: https://dx.doi.org/10.1002/sim.7825]
178. Rosenbaum, P.R. A new u-statistic with superior design sensitivity in matched observational studies. Biometrics; 2011; 67, pp. 1017-1027. [DOI: https://dx.doi.org/10.1111/j.1541-0420.2010.01535.x]
179. Zhang, W.; Jin, B.; Bai, Z. Learning block structures in U-statistic-based matrices. Biometrika; 2021; 108, pp. 933-946. [DOI: https://dx.doi.org/10.1093/biomet/asaa099]
180. Wei, C.; Elston, R.C.; Lu, Q. A weighted U statistic for association analyses considering genetic heterogeneity. Stat. Med.; 2016; 35, pp. 2802-2814. [DOI: https://dx.doi.org/10.1002/sim.6877] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/26833871]
181. van der Vaart, A.W.; Wellner, J.A. Weak Convergence and Empirical Processes; Springer Series in Statistics Springer: New York, NY, USA, 1996; xvi+508. [DOI: https://dx.doi.org/10.1007/978-1-4757-2545-2]
182. Kosorok, M.R. Introduction to Empirical Processes and Semiparametric Inference; Springer Series in Statistics Springer: New York, NY, USA, 2008; xiv+483. [DOI: https://dx.doi.org/10.1007/978-0-387-74978-5]
183. van der Vaart, A. New Donsker classes. Ann. Probab.; 1996; 24, pp. 2128-2140. [DOI: https://dx.doi.org/10.1214/aop/1041903221]
184. Blum, J.R.; Kiefer, J.; Rosenblatt, M. Distribution free tests of independence based on the sample distribution function. Ann. Math. Statist.; 1961; 32, pp. 485-498. [DOI: https://dx.doi.org/10.1214/aoms/1177705055]
185. Bouzebda, S.; Keziou, A.; Zari, T. K-sample problem using strong approximations of empirical copula processes. Math. Methods Statist.; 2011; 20, pp. 14-29. [DOI: https://dx.doi.org/10.3103/S1066530711010029]
186. Bouzebda, S.; Keziou, A. New estimates and tests of independence in semiparametric copula models. Kybernetika; 2010; 46, pp. 178-201.
187. Bouzebda, S.; Keziou, A. A new test procedure of independence in copula models via χ2-divergence. Comm. Statist. Theory Methods; 2010; 39, pp. 1-20. [DOI: https://dx.doi.org/10.1080/03610920802645379]
188. Bergsma, W.; Dassios, A. A consistent test of independence based on a sign covariance related to Kendall’s tau. Bernoulli; 2014; 20, pp. 1006-1028. [DOI: https://dx.doi.org/10.3150/13-BEJ514]
189. Borovkova, S.; Burton, R.; Dehling, H. Consistency of the Takens estimator for the correlation dimension. Ann. Appl. Probab.; 1999; 9, pp. 376-390. [DOI: https://dx.doi.org/10.1214/aoap/1029962747]
190. Silverman, B.W. Distances on circles, toruses and spheres. J. Appl. Probab.; 1978; 15, pp. 136-143. [DOI: https://dx.doi.org/10.2307/3213243]
191. Hollander, M.; Proschan, F. Testing whether new is better than used. Ann. Math. Statist.; 1972; 43, pp. 1136-1146. [DOI: https://dx.doi.org/10.1214/aoms/1177692466]
192. Hall, P. Asymptotic properties of integrated square error and cross-validation for kernel estimation of a regression function. Z. Wahrsch. Verw. Geb.; 1984; 67, pp. 175-196. [DOI: https://dx.doi.org/10.1007/BF00535267]
193. Härdle, W.; Marron, J.S. Optimal bandwidth selection in nonparametric regression function estimation. Ann. Statist.; 1985; 13, pp. 1465-1481. [DOI: https://dx.doi.org/10.1214/aos/1176349748]
194. Rachdi, M.; Vieu, P. Nonparametric regression for functional data: Automatic smoothing parameter selection. J. Statist. Plann. Inference; 2007; 137, pp. 2784-2801. [DOI: https://dx.doi.org/10.1016/j.jspi.2006.10.001]
195. Dony, J.; Mason, D.M. Uniform in bandwidth consistency of conditional U-statistics. Bernoulli; 2008; 14, pp. 1108-1133. [DOI: https://dx.doi.org/10.3150/08-BEJ136]
196. Marron, J.S. An asymptotically efficient solution to the bandwidth problem of kernel density estimation. Ann. Statist.; 1985; 13, pp. 1011-1023. [DOI: https://dx.doi.org/10.1214/aos/1176349653]
197. Vieu, P. Nonparametric regression: Optimal local bandwidth choice. J. Roy. Statist. Soc. Ser. B; 1991; 53, pp. 453-464. [DOI: https://dx.doi.org/10.1111/j.2517-6161.1991.tb01837.x]
198. Dedecker, J.; Louhichi, S. Maximal inequalities and empirical central limit theorems. Empirical Process Techniques for Dependent Data; Birkhäuser Boston: Boston, MA, USA, 2002; pp. 137-159.
199. Heinrich, L. Bounds for the absolute regularity coefficient of a stationary renewal process. Yokohama Math. J.; 1992; 40, pp. 25-33.
200. Robinson, P.M. Large-sample inference for nonparametric regression with dependent errors. Ann. Statist.; 1997; 25, pp. 2054-2083. [DOI: https://dx.doi.org/10.1214/aos/1069362387]
201. Yajima, Y. On estimation of a regression model with long-memory stationary errors. Ann. Statist.; 1988; 16, pp. 791-807. [DOI: https://dx.doi.org/10.1214/aos/1176350837]
202. Toussoun, O. Mémoire sur L’histoire du Nil; Mémoires de l’Institut d’Egypte, Institut d’Egypte: Cairo, Egypt, 1925; Volume 9.
203. Karlsen, H.A.; Tjøstheim, D. Nonparametric estimation in null recurrent time series. Ann. Statist.; 2001; 29, pp. 372-416. [DOI: https://dx.doi.org/10.1214/aos/1009210546]
204. Bouzebda, S.; Didi, S. Additive regression model for stationary and ergodic continuous time processes. Comm. Statist. Theory Methods; 2017; 46, pp. 2454-2493. [DOI: https://dx.doi.org/10.1080/03610926.2015.1048882]
205. Bouzebda, S.; Didi, S. Multivariate wavelet density and regression estimators for stationary and ergodic discrete time processes: Asymptotic results. Comm. Statist. Theory Methods; 2017; 46, pp. 1367-1406. [DOI: https://dx.doi.org/10.1080/03610926.2015.1019144]
206. Bouanani, O.; Bouzebda, S. Limit theorems for local polynomial estimation of regression for functional dependent data. AIMS Math.; 2024; 9, pp. 23651-23691. [DOI: https://dx.doi.org/10.3934/math.20241150]
207. Lasota, A. Dynamical Systems on Measures; [Układy dynamiczne na mi- arach] Wydawnictwa Uniwersytetu Ślaskiego: Katowice, Poland, 2008; (In Polish)
208. Mitkowski, P.J. Mathematical Structures of Ergodicity and Chaos in Population Dynamics; Studies in Systems, Decision and Control Springer: Cham, Switzerland, 2021; Volume 312, xii+97. [DOI: https://dx.doi.org/10.1007/978-3-030-57678-3]
209. Cornfeld, I.P.; Fomin, S.V.; Sinai, Y.G. Ergodic Theory. Transl. From the Russian by A. B. Sossinskii; Grundlehren Math. Wiss. Springer: Cham, Switzerland, 1982; Volume 245.
210. Bouzebda, S.; Limnios, N. On general bootstrap of empirical estimator of a semi-Markov kernel with applications. J. Multivar. Anal.; 2013; 116, pp. 52-62. [DOI: https://dx.doi.org/10.1016/j.jmva.2012.11.008]
211. Cai, G.; Shehu, Y.; Iyiola, O.S. Inertial Tseng’s extragradient method for solving variational inequality problems of pseudo-monotone and non-Lipschitz operators. J. Ind. Manag. Optim.; 2022; 18, pp. 2873-2902. [DOI: https://dx.doi.org/10.3934/jimo.2021095]
212. Song, Y.; Ullah, M.Z. On a sparse and stable solver on graded meshes for solving high-dimensional parabolic pricing PDEs. Comput. Math. Appl.; 2023; 143, pp. 224-233. [DOI: https://dx.doi.org/10.1016/j.camwa.2023.05.008]
213. Mohammedi, M.; Bouzebda, S.; Laksaci, A.; Bouanani, O. Asymptotic normality of the k-NN single index regression estimator for functional weak dependence data. Comm. Statist. Theory Methods; 2024; 53, pp. 3143-3168. [DOI: https://dx.doi.org/10.1080/03610926.2022.2150823]
214. Giné, E.; Zinn, J. Some limit theorems for empirical processes. Ann. Probab.; 1984; 12, pp. 929-998. [DOI: https://dx.doi.org/10.1214/aop/1176993138]
215. LeCam, L. A remark on empirical measures. A Festschrift for Erich Lehmann in Honor of His Sixty-Fifth Birthday; Wadsworth Statist./Probab. Ser. Wadsworth: Belmont, CA, USA, 1983; pp. 305-327.
216. Davydov, J.A. Convergence of distributions generated by stationary stochastic processes. Theory Probab. Appl.; 1968; 13, pp. 691-696. [DOI: https://dx.doi.org/10.1137/1113086]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2024 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
U-statistics are fundamental in modeling statistical measures that involve responses from multiple subjects. They generalize the concept of the empirical mean of a random variable X to include summations over each m-tuple of distinct observations of X. W. Stute introduced conditional U-statistics, extending the Nadaraya–Watson estimates for regression functions. Stute demonstrated their strong pointwise consistency with the conditional expectation
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer