Sequential neural models with stochastic layers
Research output: Chapter in Book/Report/Conference proceeding › Article in proceedings › Research › peer-review
Standard
Sequential neural models with stochastic layers. / Fraccaro, Marco; Sønderby, Søren Kaae; Paquet, Ulrich; Winther, Ole.
Neural Information Processing Systems 2016. ed. / D. D. Lee; M. Sugiyama; U. V. Luxburg; I. Guyon; R. Garnett. Neural Information Processing Systems Foundation, 2016. p. 2207-2215 (Advances in Neural Information Processing Systems, Vol. 29).Research output: Chapter in Book/Report/Conference proceeding › Article in proceedings › Research › peer-review
Harvard
APA
Vancouver
Author
Bibtex
}
RIS
TY - GEN
T1 - Sequential neural models with stochastic layers
AU - Fraccaro, Marco
AU - Sønderby, Søren Kaae
AU - Paquet, Ulrich
AU - Winther, Ole
N1 - Conference code: 30
PY - 2016
Y1 - 2016
N2 - How can we efficiently propagate uncertainty in a latent state representation with recurrent neural networks? This paper introduces stochastic recurrent neural networks which glue a deterministic recurrent neural network and a state space model together to form a stochastic and sequential neural generative model. The clear separation of deterministic and stochastic layers allows a structured variational inference network to track the factorization of the model's posterior distribution. By retaining both the nonlinear recursive structure of a recurrent neural network and averaging over the uncertainty in a latent path, like a state space model, we improve the state of the art results on the Blizzard and TIMIT speech modeling data sets by a large margin, while achieving comparable performances to competing methods on polyphonic music modeling.
AB - How can we efficiently propagate uncertainty in a latent state representation with recurrent neural networks? This paper introduces stochastic recurrent neural networks which glue a deterministic recurrent neural network and a state space model together to form a stochastic and sequential neural generative model. The clear separation of deterministic and stochastic layers allows a structured variational inference network to track the factorization of the model's posterior distribution. By retaining both the nonlinear recursive structure of a recurrent neural network and averaging over the uncertainty in a latent path, like a state space model, we improve the state of the art results on the Blizzard and TIMIT speech modeling data sets by a large margin, while achieving comparable performances to competing methods on polyphonic music modeling.
UR - http://www.scopus.com/inward/record.url?scp=85019203505&partnerID=8YFLogxK
M3 - Article in proceedings
AN - SCOPUS:85019203505
T3 - Advances in Neural Information Processing Systems
SP - 2207
EP - 2215
BT - Neural Information Processing Systems 2016
A2 - Lee, D. D.
A2 - Sugiyama, M.
A2 - Luxburg, U. V.
A2 - Guyon, I.
A2 - Garnett, R.
PB - Neural Information Processing Systems Foundation
Y2 - 5 December 2016 through 10 December 2016
ER -
ID: 179362819