Composable, Distributed-state Models for High-dimensional Time Series

In this thesis we develop a class of nonlinear generative models for high-dimensional time series. The first key property of these models is their distributed, or "componential" latent state, which is characterized by binary stochastic variables which interact to explain the data. The seco...

Full description

Bibliographic Details
Main Author: Taylor, Graham William
Other Authors: Hinton, Geoffrey
Language:en_ca
Published: 2009
Subjects:
Online Access:http://hdl.handle.net/1807/19238
id ndltd-TORONTO-oai-tspace.library.utoronto.ca-1807-19238
record_format oai_dc
spelling ndltd-TORONTO-oai-tspace.library.utoronto.ca-1807-192382013-01-08T04:03:43ZComposable, Distributed-state Models for High-dimensional Time SeriesTaylor, Graham Williammachine learningtime seriesneural networksunsupervised learningrestricted Boltzmann machineshidden Markov modelsgraphical modelsmotion capturedynamical modelsgenerative modelscomputer visiontracking0984In this thesis we develop a class of nonlinear generative models for high-dimensional time series. The first key property of these models is their distributed, or "componential" latent state, which is characterized by binary stochastic variables which interact to explain the data. The second key property is the use of an undirected graphical model to represent the relationship between latent state (features) and observations. The final key property is composability: the proposed class of models can form the building blocks of deep networks by successively training each model on the features extracted by the previous one. We first propose a model based on the Restricted Boltzmann Machine (RBM) that uses an undirected model with binary latent variables and real-valued "visible" variables. The latent and visible variables at each time step receive directed connections from the visible variables at the last few time-steps. This "conditional" RBM (CRBM) makes on-line inference efficient and allows us to use a simple approximate learning procedure. We demonstrate the power of our approach by synthesizing various motion sequences and by performing on-line filling in of data lost during motion capture. We also explore CRBMs as priors in the context of Bayesian filtering applied to multi-view and monocular 3D person tracking. We extend the CRBM in a way that preserves its most important computational properties and introduces multiplicative three-way interactions that allow the effective interaction weight between two variables to be modulated by the dynamic state of a third variable. We introduce a factoring of the implied three-way weight tensor to permit a more compact parameterization. The resulting model can capture diverse styles of motion with a single set of parameters, and the three-way interactions greatly improve its ability to blend motion styles or to transition smoothly among them. In separate but related work, we revisit Products of Hidden Markov Models (PoHMMs). We show how the partition function can be estimated reliably via Annealed Importance Sampling. This enables us to demonstrate that PoHMMs outperform various flavours of HMMs on a variety of tasks and metrics, including log likelihood.Hinton, GeoffreyRoweis, Sam2009-112010-03-03T15:25:46ZNO_RESTRICTION2010-03-03T15:25:46Z2010-03-03T15:25:46ZThesishttp://hdl.handle.net/1807/19238en_ca
collection NDLTD
language en_ca
sources NDLTD
topic machine learning
time series
neural networks
unsupervised learning
restricted Boltzmann machines
hidden Markov models
graphical models
motion capture
dynamical models
generative models
computer vision
tracking
0984
spellingShingle machine learning
time series
neural networks
unsupervised learning
restricted Boltzmann machines
hidden Markov models
graphical models
motion capture
dynamical models
generative models
computer vision
tracking
0984
Taylor, Graham William
Composable, Distributed-state Models for High-dimensional Time Series
description In this thesis we develop a class of nonlinear generative models for high-dimensional time series. The first key property of these models is their distributed, or "componential" latent state, which is characterized by binary stochastic variables which interact to explain the data. The second key property is the use of an undirected graphical model to represent the relationship between latent state (features) and observations. The final key property is composability: the proposed class of models can form the building blocks of deep networks by successively training each model on the features extracted by the previous one. We first propose a model based on the Restricted Boltzmann Machine (RBM) that uses an undirected model with binary latent variables and real-valued "visible" variables. The latent and visible variables at each time step receive directed connections from the visible variables at the last few time-steps. This "conditional" RBM (CRBM) makes on-line inference efficient and allows us to use a simple approximate learning procedure. We demonstrate the power of our approach by synthesizing various motion sequences and by performing on-line filling in of data lost during motion capture. We also explore CRBMs as priors in the context of Bayesian filtering applied to multi-view and monocular 3D person tracking. We extend the CRBM in a way that preserves its most important computational properties and introduces multiplicative three-way interactions that allow the effective interaction weight between two variables to be modulated by the dynamic state of a third variable. We introduce a factoring of the implied three-way weight tensor to permit a more compact parameterization. The resulting model can capture diverse styles of motion with a single set of parameters, and the three-way interactions greatly improve its ability to blend motion styles or to transition smoothly among them. In separate but related work, we revisit Products of Hidden Markov Models (PoHMMs). We show how the partition function can be estimated reliably via Annealed Importance Sampling. This enables us to demonstrate that PoHMMs outperform various flavours of HMMs on a variety of tasks and metrics, including log likelihood.
author2 Hinton, Geoffrey
author_facet Hinton, Geoffrey
Taylor, Graham William
author Taylor, Graham William
author_sort Taylor, Graham William
title Composable, Distributed-state Models for High-dimensional Time Series
title_short Composable, Distributed-state Models for High-dimensional Time Series
title_full Composable, Distributed-state Models for High-dimensional Time Series
title_fullStr Composable, Distributed-state Models for High-dimensional Time Series
title_full_unstemmed Composable, Distributed-state Models for High-dimensional Time Series
title_sort composable, distributed-state models for high-dimensional time series
publishDate 2009
url http://hdl.handle.net/1807/19238
work_keys_str_mv AT taylorgrahamwilliam composabledistributedstatemodelsforhighdimensionaltimeseries
_version_ 1716502639768764416