Add abstract
Want to add your dissertation abstract to this database? It only takes a minute!
Search abstract
Search for abstracts by subject, author or institution
Want to add your dissertation abstract to this database? It only takes a minute!
Search for abstracts by subject, author or institution
Unsupervised Learning under Uncertainty
by Michael Mathieu
Institution: | New York University |
---|---|
Year: | 2017 |
Keywords: | Artificial intelligence; Computer science |
Posted: | 02/01/2018 |
Record ID: | 2168687 |
Full text PDF: | http://pqdtopen.proquest.com/#viewpdf?dispub=10261120 |
Deep learning, in particular neural networks, achieved remarkable success in the recent years. However, most of it is based on supervised learning, and relies on ever larger datasets, and immense computing power. One step towards general artificial intelligence is to build a model of the world, with enough knowledge to acquire a kind of ``common sense''. Representations learned by such a model could be reused in a number of other tasks. It would reduce the requirement for labelled samples and possibly acquire a deeper understanding of the problem. The vast quantities of knowledge required to build common sense precludes the use of supervised learning, and suggests to rely on unsupervised learning instead. The concept of <i>uncertainty</i> is central to unsupervised learning. The task is usually to learn a complex, multimodal distribution. Density estimation and generative models aim at representing the whole distribution of the data, while predictive learning consists of predicting the state of the world given the context and, more often than not, the prediction is not unique. That may be because the model lacks the capacity or the computing power to make a certain prediction, or because the future depends on parameters that are not part of the observation. Finally, the world can be chaotic of truly stochastic. Representing complex, multimodal continuous distributions with deep neural networks is still an open problem. In this thesis, we first assess the difficulties of representing probabilities in high dimensional spaces, and review the related work in this domain. We then introduce two methods to address the problem of video prediction, first using a novel form of linearizing auto-encoders and latent variables, and secondly using Generative Adversarial Networks (GANs). We show how GANs can be seen as trainable loss functions to represent uncertainty, then how they can be used to disentangle factors of variation. Finally, we explore a new non-probabilistic framework for GANs.
Want to add your dissertation abstract to this database? It only takes a minute!
Search for abstracts by subject, author or institution
Electric Cooperative Managers' Strategies to Enhan...
|
|
Bullied!
Coping with Workplace Bullying
|
|
The Filipina-South Floridian International Interne...
Agency, Culture, and Paradox
|
|
Solution or Stalemate?
Peace Process in Turkey, 2009-2013
|
|
Performance, Managerial Skill, and Factor Exposure...
|
|
The Deritualization of Death
Toward a Practical Theology of Caregiving for the ...
|
|
Emotional Intelligence and Leadership Styles
Exploring the Relationship between Emotional Intel...
|
|
Commodification of Sexual Labor
Contribution of Internet Communities to Prostituti...
|
|
The Census of Warm Debris Disks in the Solar Neigh...
|
|
Risk Factors and Business Models
Understanding the Five Forces of Entrepreneurial R...
|
|