𝔖 Bobbio Scriptorium
✦   LIBER   ✦

‘Model selection for generalized linear models with factor-augmented predictors’

✍ Scribed by Hansheng Wang; Chih-Ling Tsai


Publisher
John Wiley and Sons
Year
2009
Tongue
English
Weight
37 KB
Volume
25
Category
Article
ISSN
1524-1904

No coin nor oath required. For personal study only.

✦ Synopsis


Discussion

'Model selection for generalized linear models with factor-augmented predictors'

Professors Ando and Tsay should be congratulated for such nice work, which provides an effective statistical method to handle high-dimensional data sets with generalized linear models. In this discussion, we supply and review additional references from the literature of both sufficient dimension reduction and variable selection. Thus, the authors' contribution could be further appreciated under an even larger context, both theoretically and practically.

The fast development of information technology enables us to collect an enormous amount of information at very low cost. Statistically, this means that we might face data sets with ultrahigh-dimensional predictors. As a result, analyzing high dimensional data effectively has become a fundamental and important subject in the last decade; see Fan and Li [1] for excellent discussions. To effectively conduct dimension reduction in regression, one could construct 'factors' so that they attain the optimal predictability. As noted by the authors, the traditional principal component analysis fails to be the mostly effective approach because the role of the response is completely ignored, and the resulting principal components might not be sufficiently predictive to the response. In fact, the same problem was also noted by Li [2], which leads to the seminal theory of effective dimension reduction and an innovative method of sliced inverse regression (SIR). Later, such a theory was further developed by Cook [3], which resulted in the theory of sufficient dimension reduction. Subsequently, various inverse regression methods have been proposed. For a comprehensive review, we refer to Cook [4] and Luo et al. [5]. A close look at the computational details of most inverse regression methods (e.g. SIR) reveals that they are indeed a supervised version of dimension reduction, where the effect of the response has been carefully taken into consideration. Similar philosophy was also shared in the partial least square (PLS) literature [6][7][8][9][10][11].

While inverse regression is a useful method for dimension reduction, Bair et al.

[12] considered the supervised principal component analysis, which can be considered as a two-stage procedure. In the first stage, a set of relevant variables are selected by simple univariate regressions, which leads to a reduced (or supervised) data set. In the second stage, principal components are extracted from the reduced data set and are used for predictions. Recently, such a method was further improved by the method of pre-conditioning [13]. In addition to the inverse regression and supervised principal component approaches, Fan and Lv [14] proposed an innovative and stimulating method to analyze ultra-high-dimensional data. It is namely Sure Independence Screening (SIS), which is based on correlation learning to reduce dimensionality. The theoretical and empirical performances of the above methods have been well documented in the literature.

In this article, Professors Ando and Tsay obtain an information selection criterion for generalized linear models by employing factor analysis to cope with the high-dimensional variable problem. We believe exploring its relationships with aforementioned dimension reduction methods would not only enrich the authors' results but also broaden the applicability of those methods to generalized linear, dynamic, and time series models. Finally, we want to congratulate the authors again for such an interesting work. We look forward to seeing its important applications in the foreseeable future.


📜 SIMILAR VOLUMES


‘Model selection for generalized linear
✍ T. Ando; R. S. Tsay 📂 Article 📅 2009 🏛 John Wiley and Sons 🌐 English ⚖ 56 KB

We would like to thank all the discussants for their wide-ranging discussions and constructive suggestions. They provide many useful references and directions for further research. We organize our reply in sections and hope that they can attract more discussions and encourage deeper research and dev

Model selection for generalized linear m
✍ Tomohiro Ando; Ruey S. Tsay 📂 Article 📅 2009 🏛 John Wiley and Sons 🌐 English ⚖ 236 KB

## Abstract This paper considers generalized linear models in a data‐rich environment in which a large number of potentially useful explanatory variables are available. In particular, it deals with the case that the sample size and the number of explanatory variables are of similar sizes. We adopt

‘Model selection for generalized linear
✍ W. K. Li; Guodong Li 📂 Article 📅 2009 🏛 John Wiley and Sons 🌐 English ⚖ 48 KB

We congratulate the authors for contributing such an innovative paper in terms of both modeling methodology and subject matter significance. This paper enriches the application of generalized linear models in a data-rich environment. The principal component method was first applied to reduce the num

Selection of Log-Linear Models with Orth
✍ Fred Böker 📂 Article 📅 2007 🏛 John Wiley and Sons 🌐 English ⚖ 371 KB

A model aelection criterion for log-linear models with orthonormal basis for contingency tablea is developed ueing the Gauee discrepancy between the logarithms of'the freqnenciea. The wntribution of each parameter to the criterion may be determined eeprately. A teet for the hypothesis that the w e o

Generalized Broadwell models for the dis
✍ Mitsuru Yamazaki 📂 Article 📅 2000 🏛 John Wiley and Sons 🌐 English ⚖ 97 KB

A generalization of the Broadwell models for the discrete Boltzmann equation with linear and quadratic terms is investigated. We prove that there exists a time-global solution to this model in one space-dimension for locally bounded initial data, using a maximum principle of solutions. The boundedne

On estimating equations for parameters i
✍ Roy F. Bartlett; Brajendra C. Sutradhar 📂 Article 📅 1999 🏛 John Wiley and Sons 🌐 English ⚖ 168 KB 👁 2 views

In view of the cumbersome and often intractable numerical integrations required for a full likelihood analysis, several suggestions have been made recently for approximate inference in generalized linear mixed models and other nonlinear variance component models. For example, we refer to the penaliz