Abstract: This paper presents a new data-dependent random projection method D$^2$RP for least square regressions, which maps data into the row space of a randomly mapped training data matrix. Our theoretical analysis suggests D$^2$RP may not preserve pairwise data distance as well as its data-independent ancestors, but preserves enough information for reconstructing the training data. Our further analysis shows least square regression in the D$^2$RP projected space has an $O(e^{-k/n})$ empirical excess risk that decays exponentially faster as $k$ increases, partly suggesting its high dimension efficiency. On the practical side, we apply D$^2$RP to speed up least square regression, kernel ridge regression and ensemble regression. Experimental results on real-world data sets show it achieves the best tradeoff between
computation efficiency and dimension efficiency compared to multiple baselines methods.
Loading