Vis enkel innførsel

dc.contributor.advisorJenssen, Robert
dc.contributor.authorLøkse, Sigurd
dc.date.accessioned2020-11-24T10:46:19Z
dc.date.available2020-11-24T10:46:19Z
dc.date.issued2020-12-11
dc.description.abstract<p>Kernel methods have been a central part of the machine learning arsenal for several decades. Within this framework, unsupervised learning has been a particularly challenging area. This is due to the inherent nature of unsupervised learning tasks, where important information about the structure of the data is unknown to the user, and as such it is difficult to design a kernel or system to solve the problem at hand. This thesis aims to bridge this knowledge gap on a multitude of challenges within the field. <p>Firstly, we address an important challenge within kernel methods for unsupervised learning, namely that of kernel parameter sensitivity. The process of finding the best parameter for the problem at hand usually depends on information which is unavailable for unsupervised tasks.Inspired by ideas from ensemble learning, we design a kernel for vectorial data with missing elements that automatically adapts to the inherent structures of the data, in order to decouple the parameter choice from the problem at hand. We perform experiments on spectral clustering and unsupervised ranking tasks with promising results. <p>Next, we develop a kernelized approximation method for unsupervised ranking using the Personalized PageRank (PPR). This method is based on novel insights on the PPR, which through a PPR-specific low-rank representation akin to Kernel PCA naturally leads to an out-of-sample approximation. The method is based on the spectrum of a specific matrix. We provide error bounds for the approximation which is used to order the eigenvectors/eigenvalues to minimize the approximation error. We perform a range of experiments to support our novel insights and show that our method may even outperform the PPR. <p>The final part of this thesis synergetically combines kernel methods and neural networks in various unsupervised tasks. Firstly, we design a kernelized autoencoder that incorporates similarities between datapoints through a kernel function in order to learn meaningful representations in code space. Secondly, we propose a novel deep learning approach to clustering, utilizing kernel-based information theoretic losses, with promising experimental results when compared to state--of--the art methods on challenging problems. Finally, we incorporate an unsupervised dimensionality reduction method (e.g. Kernel PCA) in-between the reservoir and readout layer of an Echo State Network in order to capture the dynamics of the time series while reducing the dimensionality of the trainable parameters and improving accuracy.en_US
dc.description.doctoraltypeph.d.en_US
dc.description.popularabstractWithin the field of machine learning, unsupervised learning refers to learning with no (or minimal) support from labels or prior knowledge. Due to large amounts of unlabeled data being available, and manually labelling data being resource intensive, unsupervised learning will be increasingly important in the future for machine learning, with applications in e.g. medical data, marketing, logistics and identifying fraudulent activity. To this end, we address several challenges within unsupervised learning, and develop novel methodology by leveraging the so-called kernels and kernel methods. The contributions of this thesis are two-fold. Firstly, we focus on pure kernel methods and develop a kernel for data with missing elements which automatically adapts to the inherent structures in the data. In addition we develop new kernel-based ranking methodology. Secondly, we propose several novel deep learning methods which are fused with kernels in order to solve unsupervised learning tasks.en_US
dc.identifier.isbn978-82-8236-414-0 (trykt), 978-82-8236-415-7 (PDF)
dc.identifier.urihttps://hdl.handle.net/10037/19911
dc.language.isoengen_US
dc.publisherUiT Norges arktiske universiteten_US
dc.publisherUiT The Arctic University of Norwayen_US
dc.relation.haspart<p>Paper I: Løkse, S., Bianchi, F.M., Salberg, A.-B. & Jenssen, R. Unsupervised learning using <i>PCKID</i> - A Probabilistic Cluster Kernel for Incomplete Data. (Submitted manuscript). <p>Paper II: Løkse, S. & Jenssen, R. Kernel Personalized PageRank. (Submitted manuscript). <p>Paper III: Kampffmeyer, M., Løkse, S., Bianchi, F.M., Jenssen, R. & Livi, R. (2018). The deep kernelized autoencoder. <i>Applied Soft Computing, 71</i>, 816-825. Also available at <a href=https://doi.org/10.1016/j.asoc.2018.07.029>https://doi.org/10.1016/j.asoc.2018.07.029</a>. Accepted manuscript version available in Munin at <a href=https://hdl.handle.net/10037/16437>https://hdl.handle.net/10037/16437</a>. <p>Paper IV: Kampffmeyer, M., Løkse, S., Bianchi, F.M., Livi, L., Salberg, A.-B. & Jenssen, R. (2019). Deep divergence-based approach to clustering. <i>Neural Networks, 113</i>, 91-101. Also available in Munin at <a href=https://hdl.handle.net/10037/17759>https://hdl.handle.net/10037/17759</a>. <p>Paper V: Løkse, S., Bianchi, F.M. & Jenssen, R. (2017). Training Echo State Networks with Regularization through Dimensionality Reduction. <i>Cognitive Computation, 9</i>, 364–378. Also available at <a href=https://doi.org/10.1007/s12559-017-9450-z>https://doi.org/10.1007/s12559-017-9450-z</a>. Submitted manuscript version available in Munin at <a href=https://hdl.handle.net/10037/13086>https://hdl.handle.net/10037/13086</a>.en_US
dc.rights.accessRightsopenAccessen_US
dc.rights.holderCopyright 2020 The Author(s)
dc.subject.courseIDDOKTOR-004
dc.subjectMachine Learningen_US
dc.subjectVDP::Technology: 500::Information and communication technology: 550en_US
dc.subjectVDP::Teknologi: 500::Informasjons- og kommunikasjonsteknologi: 550en_US
dc.titleLeveraging Kernels for Unsupervised Learningen_US
dc.typeDoctoral thesisen_US
dc.typeDoktorgradsavhandlingen_US


Tilhørende fil(er)

Thumbnail
Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel