Tim Cannings
Tue 26 Mar 2019, 11:00 - 12:00
IF 4.31/4.33

If you have a question about this talk, please contact: Gareth Beedham (gbeedham)

Random-projection ensemble classification.

We introduce a very general method for high-dimensional classification, based on careful combination of the results of applying an arbitrary base classifier to random projections of the feature vectors into a lower-dimensional space. In one special case presented here, the random projections are divided into non-overlapping blocks, and within each block we select the projection yielding the smallest estimate of the test error. Our random projection ensemble classifier then aggregates the results of applying the base classifier on the selected projections, with a data-driven voting threshold to determine the final assignment.  Our theoretical results elucidate the effect on performance of increasing the number of projections.  Moreover, under a boundary condition implied by the sufficient dimension reduction assumption, we control the test excess risk of the random projection ensemble classifier.  A simulation comparison with several other popular high-dimensional classifiers reveals its excellent finite-sample performance.  This is joint work with Richard Samworth (University of Cambridge).