Slideshare uses cookies to improve functionality and performance, and to provide you with relevant advertising. If you continue browsing the site, you agree to the use of cookies on this website. See our User Agreement and Privacy Policy.

Slideshare uses cookies to improve functionality and performance, and to provide you with relevant advertising. If you continue browsing the site, you agree to the use of cookies on this website. See our Privacy Policy and User Agreement for details.

Successfully reported this slideshow.

Like this presentation? Why not share!

- [264] large scale deep-learning_on_... by NAVER D2 14014 views
- From Conventional Machine Learning ... by Chun-Hao Chang 1438 views
- 20170707 reVisiting Cloud Computing... by m12016changTIIMP 416 views
- DevOps Introduction and the launch ... by Jan-Willem Middel... 507 views
- A Deep Architecture for Content-bas... by Cataldo Musto 693 views
- Initial Coin Offering: A Fundraisin... by Oxana Kunets 2242 views

2,542 views

Published on

Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of big annotated data and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which had been addressed until now with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or text captioning.

Published in:
Data & Analytics

No Downloads

Total views

2,542

On SlideShare

0

From Embeds

0

Number of Embeds

11

Shares

0

Downloads

118

Comments

0

Likes

4

No embeds

No notes for slide

- 1. 1 Xavier Giro-i-Nieto xavier.giro@upc.edu Associate Professor Universitat Politecnica de Catalunya Technical University of Catalonia Deep Learning Architectures #InsightDL2017
- 2. 2 The Full Story F. Van Veen, “The Neural Network Zoo” (2016)
- 3. 3 A Perceptron J. Alammar, “A visual and interactive guide to the Basics of Neural Networks” (2016) F. Van Veen, “The Neural Network Zoo” (2016)
- 4. 4 Neural Network = Multi Layer Perceptron 2 layers of perceptrons F. Van Veen, “The Neural Network Zoo” (2016)
- 5. 5 Deep Neural Network (DNN) F. Van Veen, “The Neural Network Zoo” (2016)
- 6. 6 Recurrent Neural Network (RNN) Alex Graves, “Supervised Sequence Labelling with Recurrent Neural Networks” The hidden layers and the output depend from previous states of the hidden layers
- 7. 7 Alex Graves, “Supervised Sequence Labelling with Recurrent Neural Networks” The hidden layers and the output depend from previous states of the hidden layers Recurrent Neural Network (RNN)
- 8. 8 Recurrent Neural Network (RNN) F. Van Veen, “The Neural Network Zoo” (2016) Further details: D2L8, “Recurrent”
- 9. 9 Recurrent Neural Network (RNN) F. Van Veen, “The Neural Network Zoo” (2016) Further details: D2L8, “Recurrent”
- 10. 10 Recurrent Neural Network (RNN) Deep Learning TV, “Recurrent Neural Networks - Ep. 9”
- 11. 11 Convolutional Neural Network (CNN) F. Van Veen, “The Neural Network Zoo” (2016) Further details: D2L7, “Segmentation”
- 12. 12 Convolutional Neural Network (CNN) LeCun, Y., Bottou, L., Bengio, Y., & Haffner, P. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11), 2278-2324. LeNet-5
- 13. 13 Convolutional Neural Network (CNN) You can also check @boredyannlecun on Twitter... Yann LeCun
- 14. 14 Convolutional Neural Network (CNN) Orange A Krizhevsky, I Sutskever, GE Hinton “Imagenet classification with deep convolutional neural networks” Part of: Advances in Neural Information Processing Systems 25 (NIPS 2012) AlexNet
- 15. 15 Convolutional Neural Network (CNN) Deep Learning TV, “Convolutional Neural Networks - Ep. 8”
- 16. 16 Deconvolutional Neural Network (Deconv) Junting Pan, SalGAN (2017) F. Van Veen, “The Neural Network Zoo” (2016) Further details: D2L7, “Segmentation”
- 17. 17 Deep Residual Learning / Skip connections F. Van Veen, “The Neural Network Zoo” (2016)
- 18. 18 Figure: Kilian Weinberger Deep Residual Learning / Skip connections
- 19. 19Long, Jonathan, Evan Shelhamer, and Trevor Darrell. "Fully convolutional networks for semantic segmentation." CVPR 2015 & PAMI 2016. Deep Residual Learning / Skip connections
- 20. 20 Ronneberger, Olaf, Philipp Fischer, and Thomas Brox. "U-net: Convolutional networks for biomedical image segmentation." In International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 234-241. Springer. U-Net Deep Residual Learning / Skip connections
- 21. 21Manel Baradad, Amaia Salvador, Xavier Giró-i-Nieto, Ferran Marqués (work under progress) Deep Residual Learning / Skip connections
- 22. 22 Milletari, Fausto, Nassir Navab, and Seyed-Ahmad Ahmadi. "V-net: Fully convolutional neural networks for volumetric medical image segmentation." In 3D Vision (3DV), 2016 Fourth International Conference on, pp. 565-571. IEEE, 2016. V-Net Deep Residual Learning / Skip connections
- 23. 23 Dense connections Dense Block of 5-layers with a growth rate of k=4 Connect every layer to every other layer of the same filter size. Huang, Gao, Zhuang Liu, Kilian Q. Weinberger, and Laurens van der Maaten. "Densely connected convolutional networks." arXiv preprint arXiv:1608.06993 (2016). [code] DenseNet
- 24. 24 Dense connections Huang, Gao, Zhuang Liu, Kilian Q. Weinberger, and Laurens van der Maaten. "Densely connected convolutional networks." arXiv preprint arXiv:1608.06993 (2016). [code] DenseNet
- 25. 25 Dense connections Huang, Gao, Zhuang Liu, Kilian Q. Weinberger, and Laurens van der Maaten. "Densely connected convolutional networks." arXiv preprint arXiv:1608.06993 (2016). [code]
- 26. 26 Dense connections Jégou, Simon, Michal Drozdzal, David Vazquez, Adriana Romero, and Yoshua Bengio. "The One Hundred Layers Tiramisu: Fully Convolutional DenseNets for Semantic Segmentation." arXiv preprint arXiv:1611.09326 (2016). [code] [slides]
- 27. 27 Autoencoder (AE) F. Van Veen, “The Neural Network Zoo” (2016)
- 28. 28 Autoencoder (AE) “Deep Learning Tutorial”, Dept. Computer Science, Stanford Autoencoders: ● Predict at the output the same input data. ● Do not need labels: Unsupervised learning
- 29. 29 Autoencoder (AE) “Deep Learning Tutorial”, Dept. Computer Science, Stanford Dimensionality reduction: ● Use hidden layer as a feature extractor of the desired size. Unsupervised learning
- 30. 30 Autoencoder (AE) Deep Learning TV, “Autoencoders - Ep. 10”
- 31. 31 Variational Autoencoder (VAE) F. Van Veen, “The Neural Network Zoo” (2016)
- 32. 32 Variational Autoencoder (VAE) Kevin Frans, “Variational Autoencoders explained” (2016) The latent vector learned in the hidden layer of the basic autoencoder (in green)...
- 33. 33 Variational Autoencoder (VAE) Kevin Frans, “Variational Autoencoders explained” (2016) ...is forced to follow a unit Gaussian distribution in VAEs.
- 34. 34 Variantional Autoencoder (VAE) “Deep Learning Tutorial”, Dept. Computer Science, Stanford Generative model: ● Create new samples by drawing from a Gaussian distribution. Unsupervised learning
- 35. 35 Variantional Autoencoder (VAE) Alec Radford, “Face manifold from conv/deconv variational autoencoder” (2015)
- 36. 36F. Van Veen, “The Neural Network Zoo” (2016) Restricted Boltzmann Machine (RBM)
- 37. 37 Restricted Boltzmann Machine (RBM) Figure: Geoffrey Hinton (2013) Salakhutdinov, Ruslan, Andriy Mnih, and Geoffrey Hinton. "Restricted Boltzmann machines for collaborative filtering." Proceedings of the 24th international conference on Machine learning. ACM, 2007. ● Shallow two-layer net. ● Restricted=No two nodes in a layer share a connection ● Bipartite graph. ● Bidirectional graph ○ Shared weights. ○ Different biases.
- 38. 38 Restricted Boltzmann Machine (RBM) Figure: Geoffrey Hinton (2013) Salakhutdinov, Ruslan, Andriy Mnih, and Geoffrey Hinton. "Restricted Boltzmann machines for collaborative filtering." Proceedings of the 24th international conference on Machine learning. ACM, 2007.
- 39. 39 Restricted Boltzmann Machine (RBM) DeepLearning4j, “A Beginner’s Tutorial for Restricted Boltzmann Machines”. RBMs are a specific type of autoencoder. Unsupervised learning
- 40. 40 Restricted Boltzmann Machine (RBM) Deep Learning TV, “Restricted Boltzmann Machines”.
- 41. 41 Deep Belief Networks (DBN) F. Van Veen, “The Neural Network Zoo” (2016)
- 42. 42 Deep Belief Networks (DBN) Hinton, Geoffrey E., Simon Osindero, and Yee-Whye Teh. "A fast learning algorithm for deep belief nets." Neural computation 18, no. 7 (2006): 1527-1554. ● Architecture like an MLP. ● Training as a stack of RBMs… ● ...so they do not need labels: Unsupervised learning
- 43. 43 Deep Belief Networks (DBN) Hinton, Geoffrey E., Simon Osindero, and Yee-Whye Teh. "A fast learning algorithm for deep belief nets." Neural computation 18, no. 7 (2006): 1527-1554. After the DBN is trained, it can be fine-tuned with a reduced amount of labels to solve a supervised task with superior performance. Supervised learning Softmax
- 44. 44 Deep Belief Networks (DBN) Deep Learning TV, “Deep Belief Nets”
- 45. 45 Deep Belief Networks (DBN) Geoffrey Hinton, "Introduction to Deep Learning & Deep Belief Nets” (2012) Geoorey Hinton, “Tutorial on Deep Belief Networks”. NIPS 2007.
- 46. 46 Deep Belief Networks (DBN) Geoffrey Hinton
- 47. 47 Adversarial Networks Goodfellow, Ian, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. "Generative adversarial nets." NIPS 2014 Goodfellow, Ian. "NIPS 2016 Tutorial: Generative Adversarial Networks." arXiv preprint arXiv:1701.00160 (2016). F. Van Veen, “The Neural Network Zoo” (2016) Further details: D2L3, “Generative”
- 48. 48F. Van Veen, “The Neural Network Zoo” (2016) Graves, Alex, Greg Wayne, and Ivo Danihelka. "Neural turing machines." arXiv preprint arXiv:1410.5401 (2014). [slides] [code] Differentiable Neural Computers (DNC)
- 49. 49 von Neumann architecture (1952) Neural Turing Machine (2014) Differentiable Neural Computers (DNC) Graves, Alex, Greg Wayne, and Ivo Danihelka. "Neural turing machines." arXiv preprint arXiv:1410.5401 (2014). [slides] [code] End-to-end trainable
- 50. 50 Differentiable Neural Computers (DNC) Add a trainable external memory to a neural network. Graves, Alex, Greg Wayne, Malcolm Reynolds, Tim Harley, Ivo Danihelka, Agnieszka Grabska-Barwińska, Sergio Gómez Colmenarejo et al. "Hybrid computing using a neural network with dynamic external memory." Nature 538, no. 7626 (2016): 471-476. [Post by DeepMind]
- 51. 51 Differentiable Neural Computers (DNC) DNC can solve tasks reading information from a trained memory. Graves, Alex, Greg Wayne, Malcolm Reynolds, Tim Harley, Ivo Danihelka, Agnieszka Grabska-Barwińska, Sergio Gómez Colmenarejo et al. "Hybrid computing using a neural network with dynamic external memory." Nature 538, no. 7626 (2016): 471-476. [Post by DeepMind]
- 52. 52 Reinforcement Learning (RL) Mnih, Volodymyr, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. "Playing atari with deep reinforcement learning." arXiv preprint arXiv:1312.5602 (2013).
- 53. 53 Reinforcement Learning (RL) Bernhard Schölkkopf, “Learning to see and act” Nature 2015.
- 54. 54 Reinforcement Learning (RL) https://www.theguardian.com/technology/2014/jan/27/google-acquires-uk-artificial-intelligence-startup-deepmind
- 55. 55 Reinforcement Learning (RL) An agent that is a decision-maker interacts with the environment and learns through trial-and-error Slide credit: UCL Course on RL by David Silver
- 56. 56 Reinforcement Learning (RL) Deep Q-Network (DQN) Mnih, Volodymyr, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, Alex Graves et al. "Human-level control through deep reinforcement learning." Nature 518, no. 7540 (2015): 529-533.
- 57. 57 Reinforcement Learning (RL) Deep Q-Network (DQN) Source: Tambet Matiisen, Demystifying Deep Reinforcement Learning (Nervana) Naive DQN Refined DQN
- 58. 58 Reinforcement Learning (RL) Deep Q-Network (DQN) Andrej Karpathy, “ConvNetJS Deep Q Learning Demo”
- 59. 59 Reinforcement Learning (RL) Slide credit: Míriam Bellver actor state critic ‘q-value’action (5) state action (5) actor performs an action critic assesses how good the action was, and the gradients are used to train the actor and the critic Actor-Critic algorithm Grondman, Ivo, Lucian Busoniu, Gabriel AD Lopes, and Robert Babuska. "A survey of actor-critic reinforcement learning: Standard and natural policy gradients." IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews) 42, no. 6 (2012): 1291-1307.
- 60. 60 Reinforcement Learning (RL) Silver, D., Huang, A., Maddison, C.J., Guez, A., Sifre, L., Van Den Driessche, G., Schrittwieser, J., Antonoglou, I., Panneershelvam, V., Lanctot, M. and Dieleman, S., 2016. Mastering the game of Go with deep neural networks and tree search. Nature, 529(7587), pp.484-489
- 61. 61 Reinforcement Learning (RL) Greg Kohs, “AlphaGo” (2017)
- 62. 62 Reinforcement Learning (RL) Miriam Bellver, Xavier Giro-i-Nieto, Ferran Marques, and Jordi Torres. "Hierarchical Object Detection with Deep Reinforcement Learning." Deep Reinforcement Learning Workshop NIPS 2016.
- 63. 63 Reinforcement Learning (RL) Deep Learning TV, “Reinforcement learning - Ep. 30”
- 64. 64 The Full Story F. Van Veen, “The Neural Network Zoo” (2016)
- 65. 65 More architectures to come... Yann LeCun, “A Path to AI”, Beneficial AI 2017. von Neumann architecture (1952)
- 66. 66 Thanks ! Q&A ? https://imatge.upc.edu/web/people/xavier-giro @DocXavi /ProfessorXavi #InsightDL2017

No public clipboards found for this slide

Be the first to comment