TY - GEN
T1 - Beyond PCA
T2 - 60th Annual Allerton Conference on Communication, Control, and Computing, Allerton 2024
AU - Yaghooti, Bahram
AU - Raviv, Netanel
AU - Sinopoli, Bruno
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Linear feature extraction at the presence of nonlinear dependencies among the data is a fundamental challenge in unsupervised learning. We propose using a Gram-Schmidt (GS) type orthogonalization process over function spaces in order to detect and remove redundant dimensions. Specifically, by applying the GS process over a family of functions which presumably captures the nonlinear dependencies in the data, we construct a series of covariance matrices that can either be used to identify new large-variance directions, or to remove those dependencies from the principal components. In the former case, we provide information-theoretic guarantees in terms of entropy reduction. In the latter, we prove that under certain assumptions the resulting algorithms detect and remove nonlinear dependencies whenever those dependencies lie in the linear span of the chosen function family. Both proposed methods extract linear features from the data while removing nonlinear redundancies. We provide simulation results on synthetic and real-world datasets which show improved performance over state-of-the-art feature extraction algorithms.
AB - Linear feature extraction at the presence of nonlinear dependencies among the data is a fundamental challenge in unsupervised learning. We propose using a Gram-Schmidt (GS) type orthogonalization process over function spaces in order to detect and remove redundant dimensions. Specifically, by applying the GS process over a family of functions which presumably captures the nonlinear dependencies in the data, we construct a series of covariance matrices that can either be used to identify new large-variance directions, or to remove those dependencies from the principal components. In the former case, we provide information-theoretic guarantees in terms of entropy reduction. In the latter, we prove that under certain assumptions the resulting algorithms detect and remove nonlinear dependencies whenever those dependencies lie in the linear span of the chosen function family. Both proposed methods extract linear features from the data while removing nonlinear redundancies. We provide simulation results on synthetic and real-world datasets which show improved performance over state-of-the-art feature extraction algorithms.
UR - http://www.scopus.com/inward/record.url?scp=85211190540&partnerID=8YFLogxK
U2 - 10.1109/Allerton63246.2024.10735325
DO - 10.1109/Allerton63246.2024.10735325
M3 - Conference contribution
AN - SCOPUS:85211190540
T3 - 2024 60th Annual Allerton Conference on Communication, Control, and Computing, Allerton 2024
BT - 2024 60th Annual Allerton Conference on Communication, Control, and Computing, Allerton 2024
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 24 September 2024 through 27 September 2024
ER -