Learning with group invariant features: A Kernel perspective

We analyze in this paper a random feature map based on a theory of invariance (I-theory) introduced in [1]. More specifically, a group invariant signal signature is obtained through cumulative distributions of group-transformed random projections. Our analysis bridges invariant feature learning with...

Full description

Bibliographic Details
Main Authors: Mroueh, Youssef (Author), Poggio, Tomaso A (Contributor), Voinea, Stephen Constantin (Contributor)
Other Authors: Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences (Contributor)
Format: Article
Language:English
Published: Association for Computing Machinery, 2017-11-28T19:15:40Z.
Subjects:
Online Access:Get fulltext
LEADER 01748 am a22001933u 4500
001 112309
042 |a dc 
100 1 0 |a Mroueh, Youssef  |e author 
100 1 0 |a Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences  |e contributor 
100 1 0 |a Poggio, Tomaso A  |e contributor 
100 1 0 |a Voinea, Stephen Constantin  |e contributor 
700 1 0 |a Poggio, Tomaso A  |e author 
700 1 0 |a Voinea, Stephen Constantin  |e author 
245 0 0 |a Learning with group invariant features: A Kernel perspective 
260 |b Association for Computing Machinery,   |c 2017-11-28T19:15:40Z. 
856 |z Get fulltext  |u http://hdl.handle.net/1721.1/112309 
520 |a We analyze in this paper a random feature map based on a theory of invariance (I-theory) introduced in [1]. More specifically, a group invariant signal signature is obtained through cumulative distributions of group-transformed random projections. Our analysis bridges invariant feature learning with kernel methods, as we show that this feature map defines an expected Haar-integration kernel that is invariant to the specified group action. We show how this non-linear random feature map approximates this group invariant kernel uniformly on a set of N points. Moreover, we show that it defines a function space that is dense in the equivalent Invariant Reproducing Kernel Hilbert Space. Finally, we quantify error rates of the convergence of the empirical risk minimization, as well as the reduction in the sample complexity of a learning algorithm using such an invariant representation for signal classification, in a classical supervised learning setting. 
655 7 |a Article 
773 |t Proceedings of the 28th International Conference on Neural Information Processing Systems (NIPS '15)