Recurrent Hidden Semi-Markov Model

Hanjun Dai, Bo Dai, Yan-Ming Zhang, Shuang Li, Le Song

Nov 04, 2016 (modified: Mar 04, 2017) ICLR 2017 conference submission readers: everyone
  • Abstract: Segmentation and labeling of high dimensional time series data has wide applications in behavior understanding and medical diagnosis. Due to the difficulty in obtaining the label information for high dimensional data, realizing this objective in an unsupervised way is highly desirable. Hidden Semi-Markov Model (HSMM) is a classical tool for this problem. However, existing HSMM and its variants has simple conditional assumptions of observations, thus the ability to capture the nonlinear and complex dynamics within segments is limited. To tackle this limitation, we propose to incorporate the Recurrent Neural Network (RNN) to model the generative process in HSMM, resulting the Recurrent HSMM (R-HSMM). To accelerate the inference while preserving accuracy, we designed a structure encoding function to mimic the exact inference. By generalizing the penalty method to distribution space, we are able to train the model and the encoding function simultaneously. Empirical results show that the proposed R-HSMM achieves the state-of-the-art performances on both synthetic and real-world datasets.
  • TL;DR: We propose to incorporate the RNN to model the generative process in Hidden Semi-Markov Model for unsupervised segmentation and labeling.
  • Keywords: Deep learning, Unsupervised Learning, Structured prediction
  • Conflicts: gatech.edu, nlpr.ia.ac.cn, purdue.edu

Loading