You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
RNN is for consuming deconvolutional decoder to have dependency with previous outputs which is,
instead of
encode every detail of a text fragment instead of high level feature (like semantic)
3.4. Optimization Difficulties
input dropout helped
add auxiliary reconstruction term computed from last deconvolutional layer,
finally,
autoregressive part reuses these features
4. Experiments
4.1. Comparision with LSTM VAE
Historyless decoding
paper model's historyless decoding was better
computationally faster (factor 2)
Decoding with history
Paper checks
is historyless decoding generallizes well?
how model copes latent variable well
paper claims their model does not fail on long texts
4.2. Controlling the KL term
Aux cost weight
using input dropout increases final loss but this is trade off
note that model finds non-trivial latent vectors when a is large enough
Receptive field
goal is to study the relationship between KL term values and expressiveness of decoder
-RNN decoder in LSTM VAE completely ignore information on latent vector
Aux helps as Figure 6
The text was updated successfully, but these errors were encountered:
https://arxiv.org/abs/1702.02390
Abstract
1. Introduction
2. Related Work
3. Model
3.1. Variational Autoencoder
3.2. Deconvolutional Networks
3.3. Hybrid Convolutional-Recurrent VAE (paper model)
instead of
3.4. Optimization Difficulties
4. Experiments
4.1. Comparision with LSTM VAE
Historyless decoding
Decoding with history
4.2. Controlling the KL term
Aux cost weight
Receptive field
-RNN decoder in LSTM VAE completely ignore information on latent vector
The text was updated successfully, but these errors were encountered: