We would like to thank all the discussants for their wide-ranging discussions and constructive suggestions. They provide many useful references and directions for further research. We organize our reply in sections and hope that they can attract more discussions and encourage deeper research and dev
‘Model selection for generalized linear models with factor-augmented predictors’
✍ Scribed by Hansheng Wang; Chih-Ling Tsai
- Publisher
- John Wiley and Sons
- Year
- 2009
- Tongue
- English
- Weight
- 37 KB
- Volume
- 25
- Category
- Article
- ISSN
- 1524-1904
- DOI
- 10.1002/asmb.782
No coin nor oath required. For personal study only.
✦ Synopsis
Discussion
'Model selection for generalized linear models with factor-augmented predictors'
Professors Ando and Tsay should be congratulated for such nice work, which provides an effective statistical method to handle high-dimensional data sets with generalized linear models. In this discussion, we supply and review additional references from the literature of both sufficient dimension reduction and variable selection. Thus, the authors' contribution could be further appreciated under an even larger context, both theoretically and practically.
The fast development of information technology enables us to collect an enormous amount of information at very low cost. Statistically, this means that we might face data sets with ultrahigh-dimensional predictors. As a result, analyzing high dimensional data effectively has become a fundamental and important subject in the last decade; see Fan and Li [1] for excellent discussions. To effectively conduct dimension reduction in regression, one could construct 'factors' so that they attain the optimal predictability. As noted by the authors, the traditional principal component analysis fails to be the mostly effective approach because the role of the response is completely ignored, and the resulting principal components might not be sufficiently predictive to the response. In fact, the same problem was also noted by Li [2], which leads to the seminal theory of effective dimension reduction and an innovative method of sliced inverse regression (SIR). Later, such a theory was further developed by Cook [3], which resulted in the theory of sufficient dimension reduction. Subsequently, various inverse regression methods have been proposed. For a comprehensive review, we refer to Cook [4] and Luo et al. [5]. A close look at the computational details of most inverse regression methods (e.g. SIR) reveals that they are indeed a supervised version of dimension reduction, where the effect of the response has been carefully taken into consideration. Similar philosophy was also shared in the partial least square (PLS) literature [6][7][8][9][10][11].
While inverse regression is a useful method for dimension reduction, Bair et al.
[12] considered the supervised principal component analysis, which can be considered as a two-stage procedure. In the first stage, a set of relevant variables are selected by simple univariate regressions, which leads to a reduced (or supervised) data set. In the second stage, principal components are extracted from the reduced data set and are used for predictions. Recently, such a method was further improved by the method of pre-conditioning [13]. In addition to the inverse regression and supervised principal component approaches, Fan and Lv [14] proposed an innovative and stimulating method to analyze ultra-high-dimensional data. It is namely Sure Independence Screening (SIS), which is based on correlation learning to reduce dimensionality. The theoretical and empirical performances of the above methods have been well documented in the literature.
In this article, Professors Ando and Tsay obtain an information selection criterion for generalized linear models by employing factor analysis to cope with the high-dimensional variable problem. We believe exploring its relationships with aforementioned dimension reduction methods would not only enrich the authors' results but also broaden the applicability of those methods to generalized linear, dynamic, and time series models. Finally, we want to congratulate the authors again for such an interesting work. We look forward to seeing its important applications in the foreseeable future.
📜 SIMILAR VOLUMES
## Abstract This paper considers generalized linear models in a data‐rich environment in which a large number of potentially useful explanatory variables are available. In particular, it deals with the case that the sample size and the number of explanatory variables are of similar sizes. We adopt
We congratulate the authors for contributing such an innovative paper in terms of both modeling methodology and subject matter significance. This paper enriches the application of generalized linear models in a data-rich environment. The principal component method was first applied to reduce the num
A model aelection criterion for log-linear models with orthonormal basis for contingency tablea is developed ueing the Gauee discrepancy between the logarithms of'the freqnenciea. The wntribution of each parameter to the criterion may be determined eeprately. A teet for the hypothesis that the w e o
A generalization of the Broadwell models for the discrete Boltzmann equation with linear and quadratic terms is investigated. We prove that there exists a time-global solution to this model in one space-dimension for locally bounded initial data, using a maximum principle of solutions. The boundedne
In view of the cumbersome and often intractable numerical integrations required for a full likelihood analysis, several suggestions have been made recently for approximate inference in generalized linear mixed models and other nonlinear variance component models. For example, we refer to the penaliz