Jump to content

Generalized functional linear model

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Dragon warrior rises (talk | contribs) at 03:25, 6 March 2015. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

This sandbox is in the article namespace. Either move this page into your userspace, or remove the {{User sandbox}} template. The Generalized Functional Linear Model(GFLM) is an extension of the Generalized linear model(GLM) that allows one to regress univariate responses of various types (continuous or discrete) on functional predictors, which are mostly random trajectories generated by a square-integrable stochastic processes. Similarly to GLM, a link function relates the expected value of the response variable to a linear predictor, which in case of GFLM is obtained by forming the scalar product of the random predictor function with a smooth parameter function . Functional Linear Regression, Functional Poisson Regression and Functional Binomial Regression, with the important Functional Logistic Regression included, are special cases of GFLM. Applications of GFLM include classification and discrimination of stochastic processes and functional data.

Overview

A key aspect of GFLM is estimation and inference for the smooth parameter function which is usually obtained by dimension reduction of the infinite dimensional functional predictor. A common method is to expand the predictor function in an orthonormal basis of L2 space, the Hilbert space of square integrable functions with the simultaneous expansion of the parameter function in the same basis. This is then applied to reduce the contribution of the of the parameter function in the linear predictor to a finite number of regression coefficients.Functional principal component analysis (FPCA) , that employs the Karhunen-Loève expansion is a common and parsimonious approach to accomplish this. However other orthogonal expansions, like Fourier expansions and B-spline expansions may also be employed for the dimension reduction step. Akaike Information Criteria(AIC) can be used for selecting the number of included components but minimization of cross-validation prediction errors is another criteria often used in classification applications. Once the dimension of the predictor process is reduced, the simplified linear predictor allows the usage of GLM and quasi-likelihood estimation techniques to obtain estimates of the finite dimensional regression coefficients which in turn provide an estimate of the parameter function in the GFLM.

Model Components

Linear Predictor

The predictor function typically are square integrable stochastic processes on a real interval and the unknown smooth parameter function is assumed to be square integrable on . Given a real measure on , the linear predictor is given by . Inclusion of the intercept allows us to require for all in .

Response Variable and Variance Function

The outcome is typically a real valued random variable which may be either continuous or discrete. Often the conditional distribution of given the predictor process is specified within the exponential family. However it is also sufficient to consider the functional quasi-likelihood set up, where instead of the distribution of the response one specifies the conditional variance function, as a function of the conditional mean, .

The link function is a smooth invertible function, that relates the conditional mean of the response with the linear predictor . The relationship is given by .

Formulation

In order to implement the necessary dimension reduction, the centered predictor function and the parameter function are expanded as,

and ,

where is an orthonormal basis of the function space such that .
The random variables are given by and the coefficients as for .
Note that and and denoting , we also have .
From the orthonormality of the basis functions , it follows immediately that .
Therefore ,where is a constant term and can be absorbed into the intercept for the required analysis.
The key step is then approximating by for a suitably chosen .
FPCA gives the most parsimonious approximation of the linear predictor as given a fixed number of basis functions, the eigenfunction basis explains more of the variation than any other set of basis functions.


For a differentiable link function with bounded first derivative, the approximation error of the -truncated model i.e. the linear predictor trunctated to the summation of the first components, is a constant multiple of .
A heuristic motivation for the truncation strategy derives from the fact that which is a consequence of the Cauchy-Schwarz inequality and by noting that the of the last inequality converges to 0 as since both and are finite.


For the special case of the eigenfunction basis, the sequence corresponds to the sequence of the eigenvalues of the covariance kernel .
For a data with i.i.d observations, setting , and , the approximated linear predictors can be represented as which are related to the means through .

Estimation

The main aim is to estimate the parameter function .
Once has been fixed, standard GLM and quasi-likelihood methods can be used for the -truncated model to estimate by solving the estimating equation or the score equation
The vector valued score function turns out to be which depends on through and .
Just like in GLM, the equation is solved using iterative methods like Newton Raphson(NR) or Fisher Scoring(FS) or Iterated Weighted Least Squares(IWLS) to get the estimate of the regression coefficients ,
leading to the estimate of the parameter function in the GFLM as . When using the canonical link function these methods are equivalent.
Several analysis is available in literature of -truncated models as which provide asymptotic inference for the deviation of the estimated parametric function from the true parametric function and also asymptotic tests for regression effects and asymptotic confidence regions.