Learning Infinite Layer Networks Without the Kernel Trick

Roi Livni, Daniel Carmon, Amir Globerson
Proceedings of the 34th International Conference on Machine Learning, PMLR 70:2198-2207, 2017.

Abstract

Infinite Layer Networks (ILN) have been proposed as an architecture that mimics neural networks while enjoying some of the advantages of kernel methods. ILN are networks that integrate over infinitely many nodes within a single hidden layer. It has been demonstrated by several authors that the problem of learning ILN can be reduced to the kernel trick, implying that whenever a certain integral can be computed analytically they are efficiently learnable. In this work we give an online algorithm for ILN, which avoids the kernel trick assumption. More generally and of independent interest, we show that kernel methods in general can be exploited even when the kernel cannot be efficiently computed but can only be estimated via sampling. We provide a regret analysis for our algorithm, showing that it matches the sample complexity of methods which have access to kernel values. Thus, our method is the first to demonstrate that the kernel trick is not necessary, as such, and random features suffice to obtain comparable performance.

Cite this Paper


BibTeX
@InProceedings{pmlr-v70-livni17a, title = {Learning Infinite Layer Networks Without the Kernel Trick}, author = {Roi Livni and Daniel Carmon and Amir Globerson}, booktitle = {Proceedings of the 34th International Conference on Machine Learning}, pages = {2198--2207}, year = {2017}, editor = {Precup, Doina and Teh, Yee Whye}, volume = {70}, series = {Proceedings of Machine Learning Research}, month = {06--11 Aug}, publisher = {PMLR}, pdf = {http://proceedings.mlr.press/v70/livni17a/livni17a.pdf}, url = {https://proceedings.mlr.press/v70/livni17a.html}, abstract = {Infinite Layer Networks (ILN) have been proposed as an architecture that mimics neural networks while enjoying some of the advantages of kernel methods. ILN are networks that integrate over infinitely many nodes within a single hidden layer. It has been demonstrated by several authors that the problem of learning ILN can be reduced to the kernel trick, implying that whenever a certain integral can be computed analytically they are efficiently learnable. In this work we give an online algorithm for ILN, which avoids the kernel trick assumption. More generally and of independent interest, we show that kernel methods in general can be exploited even when the kernel cannot be efficiently computed but can only be estimated via sampling. We provide a regret analysis for our algorithm, showing that it matches the sample complexity of methods which have access to kernel values. Thus, our method is the first to demonstrate that the kernel trick is not necessary, as such, and random features suffice to obtain comparable performance.} }
Endnote
%0 Conference Paper %T Learning Infinite Layer Networks Without the Kernel Trick %A Roi Livni %A Daniel Carmon %A Amir Globerson %B Proceedings of the 34th International Conference on Machine Learning %C Proceedings of Machine Learning Research %D 2017 %E Doina Precup %E Yee Whye Teh %F pmlr-v70-livni17a %I PMLR %P 2198--2207 %U https://proceedings.mlr.press/v70/livni17a.html %V 70 %X Infinite Layer Networks (ILN) have been proposed as an architecture that mimics neural networks while enjoying some of the advantages of kernel methods. ILN are networks that integrate over infinitely many nodes within a single hidden layer. It has been demonstrated by several authors that the problem of learning ILN can be reduced to the kernel trick, implying that whenever a certain integral can be computed analytically they are efficiently learnable. In this work we give an online algorithm for ILN, which avoids the kernel trick assumption. More generally and of independent interest, we show that kernel methods in general can be exploited even when the kernel cannot be efficiently computed but can only be estimated via sampling. We provide a regret analysis for our algorithm, showing that it matches the sample complexity of methods which have access to kernel values. Thus, our method is the first to demonstrate that the kernel trick is not necessary, as such, and random features suffice to obtain comparable performance.
APA
Livni, R., Carmon, D. & Globerson, A.. (2017). Learning Infinite Layer Networks Without the Kernel Trick. Proceedings of the 34th International Conference on Machine Learning, in Proceedings of Machine Learning Research 70:2198-2207 Available from https://proceedings.mlr.press/v70/livni17a.html.

Related Material