Deterministic Policy Gradient Algorithms
WHY?
Policy gradient usually requires integral over all the possible actions.
Policy gradient usually requires integral over all the possible actions.
in Studies on Deep Learning, Deep Learning
Reparameterization trick is a useful technique for estimating gradient for loss function with stochastic variables. While score function extimators suffer from great variance, RT enable the gradient to be estimated with pathwise derivatives. Even though reparameterization trick can be applied to various kinds of random variables enabling backpropagation, it has not been applicable to discrete random variables.
in Studies on Deep Learning, Deep Learning
While the effect of batch normalization was widely proven empirically, the exact mechanism of it is yet been understood. Commonly known explanation for this was internal covariance shift(ICS) meaning the change in the distribution of layer inputs caused by updates to the preceeding layers.
Instead of instantly responding to incoming stimulus, having a model of environment to make some level of prediction would help perform in reinforcement learning.
Many machine learning problems involves loss function that contains random variables. To perform backpropagation, estimating gradient of the loss function is required.
Former studies on probabilistic reasoning assume that reasoning is memoryless, which means all the inference occur independently without the reuse of previous computation.
평점: 4
Gorila framework separated several actors and learners with a centralized parameter server to parrallelize the learning process. This framework required one GPU per learner.