Abstract
How can we efficiently propagate uncertainty in a latent state representation with recurrent neural networks? This paper introduces stochastic recurrent neural networks which glue a deterministic recurrent neural network and a state space model together to form a stochastic and sequential neural generative model. The clear separation of deterministic and stochastic layers allows a structured variational inference network to track the factorization of the model's posterior distribution. By retaining both the nonlinear recursive structure of a recurrent neural network and averaging over the uncertainty in a latent path, like a state space model, we improve the state of the art results on the Blizzard and TIMIT speech modeling data sets by a large margin, while achieving comparable performances to competing methods on polyphonic music modeling.
Original language | English |
---|---|
Title of host publication | Neural Information Processing Systems 2016 |
Editors | D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, R. Garnett |
Number of pages | 9 |
Publisher | Neural Information Processing Systems Foundation |
Publication date | 2016 |
Pages | 2207-2215 |
Publication status | Published - 2016 |
Event | 30th Annual Conference on Neural Information Processing Systems - Barcelona, Spain Duration: 5 Dec 2016 → 10 Dec 2016 Conference number: 30 |
Conference
Conference | 30th Annual Conference on Neural Information Processing Systems |
---|---|
Number | 30 |
Country/Territory | Spain |
City | Barcelona |
Period | 05/12/2016 → 10/12/2016 |
Series | Advances in Neural Information Processing Systems |
---|---|
Volume | 29 |
ISSN | 1049-5258 |