Multi-task learning for sequential data via iHMMs and the nested Dirichlet process

Kai Ni, Lawrence Carin, David Dunson

Research output: Chapter in Book/Report/Conference proceedingConference contribution

13 Scopus citations

Abstract

A new hierarchical nonparametric Bayesian model is proposed for the problem of multitask learning (MTL) with sequential data. Sequential data are typically modeled with a hidden Markov model (HMM), for which one often must choose an appropriate model structure (number of states) before learning. Here we model sequential data from each task with an infinite hidden Markov model (iHMM), avoiding the problem of model selection. The MTL for iHMMs is implemented by imposing a nested Dirichlet process (nDP) prior on the base distributions of the iHMMs. The nDP-iHMM MTL method allows us to perform task-level clustering and data-level clustering simultaneously, with which the learning for individual iHMMs is enhanced and between-task similarities are learned. Learning and inference for the nDP-iHMM MTL are based on a Gibbs sampler. The effectiveness of the framework is demonstrated using synthetic data as well as real music data.
Original languageEnglish (US)
Title of host publicationACM International Conference Proceeding Series
Pages689-696
Number of pages8
DOIs
StatePublished - Aug 23 2007
Externally publishedYes

Fingerprint

Dive into the research topics of 'Multi-task learning for sequential data via iHMMs and the nested Dirichlet process'. Together they form a unique fingerprint.

Cite this