Breaking the Softmax Bottleneck: A High-Rank RNN Language Model
WHY?
This paper first prove that the expresiveness of a language model is restricted by softmax and suggest a way to overcome this limit.
This paper first prove that the expresiveness of a language model is restricted by softmax and suggest a way to overcome this limit.
in Studies on Deep Learning, Deep Learning
Gradient descent methods depend on the first order gradient of a loss function wrt parameters. However, the second order gradient(Hessian) is often neglected.
Recent variational training requires sampling of the variational posterior to estimate gradient. NVIL estimator suggest a method to estimate the gradient of the loss function wrt parameters. Since score function estimator is known to have high variance, baseline is used as variance reduction technique. However, this technique is insufficient to reduce variance in multi-sample setting as in IWAE.
Efficient exploration of agent in reinforcement learning is an important issue. Conventional exploration heuristics includes \epsilon
-greedy for DQN and entropy reward for A3C.
in Studies on Deep Learning, Deep Learning
There had been little study on learning representation that focus on clustering.
in Studies on Deep Learning, Deep Learning
Batch normalization is known as a good method to stablize the optimization of neural network by reducing internal covariate shift. However, batch normalization inheritantly depends on minibatch which impeding the use in recurrent models.
The largest drawback of training Generative Adversarial Network (GAN) is its instability. Especially, the power of discriminator greatly affect the performance of GAN. This paper suggests to weaken the discriminator by restricting the functional space of it to stablize the training.
This paper wanted to disentangle the label related and label unrelated information from data. The model of this paper is simpler and more effective than that of Disentangling Factors of Variation in Deep Representations Using Adversarial Training.