Soft Actor-Critic is an off-policy maximum entropy deep reinforcement learning algorithm that uses a stochastic actor. It was presented in a 2017 NIPS paper by researchers from OpenAI, UC Berkeley, and DeepMind. Soft Actor-Critic extends the actor-critic framework by incorporating an entropy term into the reward function to encourage exploration. This allows the agent to learn stochastic policies that can operate effectively in environments with complex, sparse rewards. The algorithm was shown to learn robust policies on continuous control tasks using deep neural networks to approximate the policy and action-value functions.
Soft Actor-Critic is an off-policy maximum entropy deep reinforcement learning algorithm that uses a stochastic actor. It was presented in a 2017 NIPS paper by researchers from OpenAI, UC Berkeley, and DeepMind. Soft Actor-Critic extends the actor-critic framework by incorporating an entropy term into the reward function to encourage exploration. This allows the agent to learn stochastic policies that can operate effectively in environments with complex, sparse rewards. The algorithm was shown to learn robust policies on continuous control tasks using deep neural networks to approximate the policy and action-value functions.
9. IMAGENET-TRAINED CNNS ARE BIASED
TOWARDS TEXTURE; INCREASING SHAPE
BIAS IMPROVES ACCURACY AND
ROBUSTNESS.
@mo_takusan9922
20181205
Under review at ICLR 2019 (review scores 8,8,7)
27. 歌深
? Robert Geirhos. IMAGENET-TRAINED CNNS ARE BIASED
TOWARDS TEXTURE; INCREASING SHAPE BIAS IMPROVES
ACCURACY AND ROBUSTNESS. Under review as a
conference paper at ICLR 2019. 2018
? @karpathy. t-SNE visualization of CNN. Results
codeshttps://cs.stanford.edu/people/karpathy/cnnembed/.
(歌孚晩: 2018/12/02)
? Stanford Vision Lab, Stanford University, Princeton
University.ImageNet. http://www.image-net.org/ .2016
Editor's Notes
#15: for VGG-16, BagNet-9/17/33 reach 0.70 / 0.79 / 0.88