This slide introduces the model which is one of the deep Q network. Dueling Network is the successor model of DQN or DDQN. You can easily understand the architecture of Dueling Network.
This slide introduces the model which is one of the deep Q network. Dueling Network is the successor model of DQN or DDQN. You can easily understand the architecture of Dueling Network.
論文紹介:
Pan, Wei-Xing, et al. "Dopamine cells respond to predicted events during classical conditioning: evidence for eligibility traces in the reward-learning network." The Journal of neuroscience 25.26 (2005): 6235-6242.
Sotetsu Koyamada (Presenter), Masanori Koyama, Ken Nakae, Shin Ishii
Graduate School of Informatics, Kyoto University
[Abstract]
We present a novel algorithm (Principal Sensitivity Analysis; PSA) to analyze the knowledge of the classifier obtained from supervised machine learning techniques. In particular, we define principal sensitivity map (PSM) as the direction on the input space to which the trained classifier is most sensitive, and use analogously defined k-th PSM to define a basis for the input space. We train neural networks with artificial data and real data, and apply the algorithm to the obtained supervised classifiers. We then visualize the PSMs to demonstrate the PSA’s ability to decompose the knowledge acquired by the trained classifiers.
[Keywords]
Sensitivity analysis Sensitivity map PCA Dark knowledge Knowledge decomposition
@PAKDD2015
May 20, 2015
Ho Chi Minh City, Viet Namﳟ
http://link.springer.com/chapter/10.1007%2F978-3-319-18038-0_48#page-1
Li, Mu, et al. "Efficient mini-batch training for stochastic optimization." Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, 2014.
http://www.cs.cmu.edu/~muli/file/minibatch_sgd.pdf
KDD2014勉強会関西会場: http://www.ml.ist.i.kyoto-u.ac.jp/kdd2014reading
論文紹介:
Pan, Wei-Xing, et al. "Dopamine cells respond to predicted events during classical conditioning: evidence for eligibility traces in the reward-learning network." The Journal of neuroscience 25.26 (2005): 6235-6242.
Sotetsu Koyamada (Presenter), Masanori Koyama, Ken Nakae, Shin Ishii
Graduate School of Informatics, Kyoto University
[Abstract]
We present a novel algorithm (Principal Sensitivity Analysis; PSA) to analyze the knowledge of the classifier obtained from supervised machine learning techniques. In particular, we define principal sensitivity map (PSM) as the direction on the input space to which the trained classifier is most sensitive, and use analogously defined k-th PSM to define a basis for the input space. We train neural networks with artificial data and real data, and apply the algorithm to the obtained supervised classifiers. We then visualize the PSMs to demonstrate the PSA’s ability to decompose the knowledge acquired by the trained classifiers.
[Keywords]
Sensitivity analysis Sensitivity map PCA Dark knowledge Knowledge decomposition
@PAKDD2015
May 20, 2015
Ho Chi Minh City, Viet Namﳟ
http://link.springer.com/chapter/10.1007%2F978-3-319-18038-0_48#page-1
Li, Mu, et al. "Efficient mini-batch training for stochastic optimization." Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, 2014.
http://www.cs.cmu.edu/~muli/file/minibatch_sgd.pdf
KDD2014勉強会関西会場: http://www.ml.ist.i.kyoto-u.ac.jp/kdd2014reading
4. 背景: Seq2Seq
詳しくはSutskever et al. 2014, Cho et al. 2014, Bahdanau et al. 2015あたりを参照
(Attentionメカニズムについては http://distill.pub/2016/augmented-rnns/ も参考になる)
Sutskever et al. 2014より引用