TY - JOUR
T1 - A Slicing-Free Perspective to Sufficient Dimension Reduction
T2 - Selective Review and Recent Developments
AU - Li, Lu
AU - Shao, Xiaofeng
AU - Yu, Zhou
N1 - Xiaofeng Shao was funded by University of Illinois at Urbana\u2010Champaign, Grant/Award Number NSF\u2010DMS 1607489 and 2210002. Zhou Yu was funded by East China Normal University, Grant/Award Number: Basic Research Project of Shanghai Science and Technology Commission 22JC1400800 and National Natural Science Foundation of China (12371289).
Xiaofeng Shao was funded by University of Illinois at Urbana-Champaign, Grant/Award Number NSF-DMS 1607489 and 2210002. Zhou Yu was funded by East China Normal University, Grant/Award Number: Basic Research Project of Shanghai Science and Technology Commission 22JC1400800 and National Natural Science Foundation of China (12371289).
PY - 2024/12
Y1 - 2024/12
N2 - Since the pioneering work of sliced inverse regression, sufficient dimension reduction has been growing into a mature field in statistics and it has broad applications to regression diagnostics, data visualisation, image processing and machine learning. In this paper, we provide a review of several popular inverse regression methods, including sliced inverse regression (SIR) method and principal hessian directions (PHD) method. In addition, we adopt a conditional characteristic function approach and develop a new class of slicing-free methods, which are parallel to the classical SIR and PHD, and are named weighted inverse regression ensemble (WIRE) and weighted PHD (WPHD), respectively. Relationship with recently developed martingale difference divergence matrix is also revealed. Numerical studies and a real data example show that the proposed slicing-free alternatives have superior performance than SIR and PHD.
AB - Since the pioneering work of sliced inverse regression, sufficient dimension reduction has been growing into a mature field in statistics and it has broad applications to regression diagnostics, data visualisation, image processing and machine learning. In this paper, we provide a review of several popular inverse regression methods, including sliced inverse regression (SIR) method and principal hessian directions (PHD) method. In addition, we adopt a conditional characteristic function approach and develop a new class of slicing-free methods, which are parallel to the classical SIR and PHD, and are named weighted inverse regression ensemble (WIRE) and weighted PHD (WPHD), respectively. Relationship with recently developed martingale difference divergence matrix is also revealed. Numerical studies and a real data example show that the proposed slicing-free alternatives have superior performance than SIR and PHD.
KW - Martingale difference divergence
KW - principal hessian directions
KW - sliced inverse regression
KW - sufficient dimension reduction
UR - http://www.scopus.com/inward/record.url?scp=85187113971&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85187113971&partnerID=8YFLogxK
U2 - 10.1111/insr.12565
DO - 10.1111/insr.12565
M3 - Article
AN - SCOPUS:85187113971
SN - 0306-7734
VL - 92
SP - 355
EP - 382
JO - International Statistical Review
JF - International Statistical Review
IS - 3
ER -