Leveraging Kernels for Unsupervised Learning
Permanent link
https://hdl.handle.net/10037/19911Date
2020-12-11Type
Doctoral thesisDoktorgradsavhandling
Author
Løkse, SigurdAbstract
Kernel methods have been a central part of the machine learning arsenal for several decades. Within this framework, unsupervised learning has been a particularly challenging area. This is due to the inherent nature of unsupervised learning tasks, where important information about the structure of the data is unknown to the user, and as such it is difficult to design a kernel or system to solve the problem at hand. This thesis aims to bridge this knowledge gap on a multitude of challenges within the field.
Firstly, we address an important challenge within kernel methods for unsupervised learning, namely that of kernel parameter sensitivity. The process of finding the best parameter for the problem at hand usually depends on information which is unavailable for unsupervised tasks.Inspired by ideas from ensemble learning, we design a kernel for vectorial data with missing elements that automatically adapts to the inherent structures of the data, in order to decouple the parameter choice from the problem at hand. We perform experiments on spectral clustering and unsupervised ranking tasks with promising results.
Next, we develop a kernelized approximation method for unsupervised ranking using the Personalized PageRank (PPR). This method is based on novel insights on the PPR, which through a PPR-specific low-rank representation akin to Kernel PCA naturally leads to an out-of-sample approximation. The method is based on the spectrum of a specific matrix. We provide error bounds for the approximation which is used to order the eigenvectors/eigenvalues to minimize the approximation error. We perform a range of experiments to support our novel insights and show that our method may even outperform the PPR.
The final part of this thesis synergetically combines kernel methods and neural networks in various unsupervised tasks. Firstly, we design a kernelized autoencoder that incorporates similarities between datapoints through a kernel function in order to learn meaningful representations in code space. Secondly, we propose a novel deep learning approach to clustering, utilizing kernel-based information theoretic losses, with promising experimental results when compared to state--of--the art methods on challenging problems. Finally, we incorporate an unsupervised dimensionality reduction method (e.g. Kernel PCA) in-between the reservoir and readout layer of an Echo State Network in order to capture the dynamics of the time series while reducing the dimensionality of the trainable parameters and improving accuracy.
Has part(s)
Paper I: Løkse, S., Bianchi, F.M., Salberg, A.-B. & Jenssen, R. Unsupervised learning using PCKID - A Probabilistic Cluster Kernel for Incomplete Data. (Submitted manuscript).
Paper II: Løkse, S. & Jenssen, R. Kernel Personalized PageRank. (Submitted manuscript).
Paper III: Kampffmeyer, M., Løkse, S., Bianchi, F.M., Jenssen, R. & Livi, R. (2018). The deep kernelized autoencoder. Applied Soft Computing, 71, 816-825. Also available at https://doi.org/10.1016/j.asoc.2018.07.029. Accepted manuscript version available in Munin at https://hdl.handle.net/10037/16437.
Paper IV: Kampffmeyer, M., Løkse, S., Bianchi, F.M., Livi, L., Salberg, A.-B. & Jenssen, R. (2019). Deep divergence-based approach to clustering. Neural Networks, 113, 91-101. Also available in Munin at https://hdl.handle.net/10037/17759.
Paper V: Løkse, S., Bianchi, F.M. & Jenssen, R. (2017). Training Echo State Networks with Regularization through Dimensionality Reduction. Cognitive Computation, 9, 364–378. Also available at https://doi.org/10.1007/s12559-017-9450-z. Submitted manuscript version available in Munin at https://hdl.handle.net/10037/13086.
Publisher
UiT Norges arktiske universitetUiT The Arctic University of Norway
Metadata
Show full item recordCollections
The following license file are associated with this item: