Jump to content

User:Poorbita/sandbox

From Wikipedia, the free encyclopedia

Functional nonlinear regression models

[edit]

Direct nonlinear extensions of the classical functional linear regression models (FLMs) still involve a linear predictor, but combine it with a nonlinear link function, analogous to the idea of generalized linear model from the conventional linear model. Developments towards fully nonparametric regression models for functional data encounter problems such as curse of dimensionality. In order to bypass the “curse” and the metric selection problem, we are motivated to consider nonlinear functional regression models, which are subject to some structural constraints but do not overly infringe flexibility. One desires models that retain polynomial rates of convergence, while being more flexible than, say, functional linear models. Such models are particularly useful when diagnostics for the functional linear model indicate lack of fit, which is often encountered in real life situations. Functional nonlinear regression models such as generalized functional linear models and extensions to single and multiple index models provide not only enhanced flexibility but also structural stability while model fits converge at polynomial rates. In particular, functional polynomial models, functional single and multiple index models and functional additive models are three special cases of functional nonlinear regression models.

Functional polynomial regression models

[edit]

Functional polynomial regression models may be viewed as a natural extension of the Functional Linear Models (FLMs) with scalar responses, analogous to extending linear regression model to polynomial regression model. For a scalar response and a functional covariate with domain and the corresponding centered predictor processes , the simplest and the most prominent member in the family of functional polynomial regression models is the quadratic functional regression[1] given as follows,

where is the centered functional covariate, is a scalar coefficient, and are coefficient functions with domains and , respectively. In addition to the parameter function β that the above functional quadratic regression model shares with the functional linear model, it also features a parameter surface γ. By analogy to FLMs with scalar responses, estimation of functional polynomial models can be obtained through expanding both the centered covariate and the coefficient functions and in an orthonormal basis.[1]

The extension to higher order polynomials is pretty obvious. These models can be equivalently represented as polynomials in the corresponding Functional Principal Components (FPCs) . Now a natural question is whether the linear model is sufficient or needs to be extended to a model that includes a quadratic term. There is a corresponding test to answer this question[2].

Functional single and multiple index models

[edit]

A functional multiple index model is given as below, with symbols having their usual meanings as formerly described,

Here g represents an (unknown) general smooth function defined on a p-dimensional domain. The case yields a functional single index model while multiple index models correspond to the case . However, for , this model is problematic due to curse of dimensionality. With and relatively small sample sizes, the estimator given by this model often has large variance.[3][4]

In order to deal with this curse of dimensionality in multiple index models with -dimensional smoother, we resort to an alternative -component functional multiple index model expressed as,

Functional additive models (FAMs)

[edit]

Let denote an expansion of a functional covariate with domain in an orthonormal basis of the function space.


A functional linear model with scalar responses [as shown in model (2) of FLM] can be thus be written as follows,

One form of FAMs is obtained by replacing the linear function of in the above expression ( i.e., ) by a general smooth function , analogous to the extension of multiple linear regression models to additive models.

The resulting form of functional additive model is thus given by, where satisfies for .[5] This constraint on the general smooth functions ensures identifiability in the sense that the estimates of these additive component functions do not interfere with that of the intercept term .

Another form of FAMs comprises a sequence of time-additive models on increasingly dense finite grids of size p and under the assumption that functions are smoothly varying in t. Such models may be described as, where is a dense grid on with increasing size , and for a smooth bivariate function g., [6]. In the limit p tends to infinity, the above FAM model becomes the continuously additive model.

Extensions

[edit]

An obvious and direct extension of FLMs with scalar responses [shown in model (2)] is to add a link function so as to create a generalized functional linear model (GFLM)[7] by analogy to extending linear model to generalized linear model (GLM), of which the three components are:

  1. Linear predictor ; [systematic component]
  2. Variance function , where is the conditional mean; [random component]
  3. Link function connecting the conditional mean and the linear predictor through . [systematic component]

Classification of functional data

[edit]

Functional clustering (discussed in the last section) aims at finding clusters by either (i) minimizing an objective function [such as (8) in functional clustering section above], or more generally, (ii) by maximizing the conditional cluster membership probability given predictor [denoted by ]. Functional classification on the other hand assigns a group membership to a new data object on the basis of a discriminant function or a classifier. In particular classification for functional data is either based on functional regression or functional discriminant analysis. Most of the popular approaches for functional data classification are based on functional regression models that feature class labels as responses and the observed functional data and other covariates as predictors. This leads to regression based functional data classification methods, such as, functional generalized linear regression models and functional multiclass logit models. Similar to functional data clustering, most functional data classification methods apply a dimension reduction technique using a truncated expansion in a pre-specified function basis or in the data-adaptive eigen basis.

Functional regression for classification

[edit]

As already stated, functional data classification methods based on functional regression models use class levels as responses and the observed functional data and other covariates as predictors. For regression based functional classification models, functional generalized linear models or more specifically, functional binary regression, such as functional logistic regression for binary response, are commonly used classification approaches. We have a functional multiclass logit model to deal with multiclass classification problems as follows.

Let us consider a random sample {}, where represents the class label for the unit under study with label ∈ {1, 2, . . . , L} for L classes under consideration and represents the associated functional observation for the same individual.

A classification model for an observation based on functional multiclass logit regression is thus given by,

,

where k = 1, 2,..., L-1. Clearly here is an intercept term and denotes the coefficient function corresponding to the kth class label. Also as is quite apparent, here we use the class label L as the baseline category and thus take,

.

This is thus a functional extension of the baseline odds model in multinomial regression.

Given a new observation , the model based Bayes classification rule is to choose the class label with the maximal posterior probability among . More generally, the generalized functional linear regression model based on the FPCA approach is used[8]. When the logit link is used in the model, it becomes the functional logistic regression model, several variants of which have been studied in literature.

Functional discriminant analysis for classification

[edit]

Functional Linear Discriminant Analysis (FLDA) has also been considered as a classification method for functional data. Developments for functional data along these lines include, Functional linear discriminant analysis approach to classify curves[9], Functional data-analytic approach to signal discrimination, using the FPCA method for dimension reduction[10] and Kernel functional classification rules for nonparametric curve discrimination.[11][12][13] Another very important development in the domain of functional data classification involves the use of density ratios of projections and is described below.[14]

We must first take a note that Bayes classifiers for functional data pose several challenges. A major difficulty is that probability density functions do not exist for functional data, so that the classical Bayes classifier using density quotients needs to be modified. The proposed classifiers can in fact be viewed as an extension to functional data of some of the very earliest nonparametric Bayes classifiers that were based on simple density ratios in the one-dimensional case. By means of the factorization of the density quotients, the curse of dimensionality that would otherwise severely affect Bayes classifiers for functional data, can be avoided.[14]

A study of the asymptotic behaviour of the proposed classifiers in the large sample limit shows that under certain conditions the misclassification rate converges to zero, a phenomenon that has been referred to as "perfect classification". The proposed classifiers also perform favourably in finite sample applications, as demonstrated through comparisons with other functional classifiers in simulations and various data applications, including spectral data, functional magnetic resonance imaging data for attention deficit hyperactivity disorder patients, and yeast gene expression data[14]. Theoretical support and a similar notion of “perfect classification” standing for asymptotically vanishing misclassification probabilities has also been discussed separately for linear[15] and quadratic functional classification[16] in some other research works cited in literature.

References

[edit]
  1. ^ a b Yao, F; Müller, HG. (2010). "Functional quadratic regression". Biometrika. 97 (1):49–64.
  2. ^ Horváth, L; Reeder, R. (2013). "A test of significance in functional quadratic regression". Bernoulli. 19 (5A): 2120–2151.
  3. ^ Chen, D; Hall, P; Müller HG. (2011). "Single and multiple index functional regression models with nonparametric link". The Annals of Statistics. 39 (3):1720–1747.
  4. ^ Jiang, CR; Wang JL. (2011). "Functional single index models for longitudinal data". he Annals of Statistics. 39 (1):362–388.
  5. ^ Müller, HG; Yao, F. (2008). "Functional additive models". Journal of the American Statistical Association. 103 (484):1534–1544.
  6. ^ Fan, Y; James, GM; Radchenko, P. (2015). "Functional additive regression". The Annals of Statistics. 43 (5):2296–2325.
  7. ^ Müller HG; Stadmüller, U. (2005). "Generalized Functional Linear Models". The Annals of Statistics. 33 (2): 774–805.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  8. ^ Leng, X; Müller, HG. (2006). "Classification using functional data analysis for temporal gene expression data". Bioinformatics. 22 (1): 68–76.
  9. ^ James, GM; Hastie, TJ. (2001). "Functional linear discriminant analysis for irregularly sampled curves". Journal of the Royal Statistical Society: Series B (Statistical Methodology). 63 (3): 533–550.
  10. ^ Hall, P; Poskitt, DS; Presnell, B. (2001). "A Functional Data—Analytic Approach to Signal Discrimination". Technometrics. 43 (1): 1–9.
  11. ^ Ferraty, F; Vieu, P. (2003). "Curves discrimination: a nonparametric functional approach". Computational Statistics & Data Analysis. 44 (1–2): 161–173.
  12. ^ Chang, C; Chen, Y; Ogden, RT. (2014). "Functional data classification: a wavelet approach". Computational Statistics. 29 (6): 1497–1513.
  13. ^ Zhu, H; Brown, PJ; Morris, JS. (2012). "Robust Classification of Functional and Quantitative Image Data Using Functional Mixed Models". Biometrics. 68 (4): 1260–1268.
  14. ^ a b c Dai, X; Müller, HG; Yao, F. (2017). "Optimal Bayes classifiers for functional data and density ratios". Biometrika. 104 (3): 545–560.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  15. ^ Delaigle, A; Hall, P. (2011). "Achieving near perfect classification for functional data". Journal of the Royal Statistical Society. 74 (2): 267–286.
  16. ^ Delaigle, A; Hall, P. (2013). "Classification Using Censored Functional Data". Journal of the American Statistical Association. 108 (504): 1269–1283.

Category:Regression analysis