SlideShare a Scribd company logo
1 of 31
CS294-129: Designing, Visualizing and
Understanding Deep Neural Networks
John Canny
Fall 2016
CN: 34652
With considerable help from Andrej
Karpathy, Fei-Fei Li, Justin Johnson, Ian
Goodfellow and several others
Deep Learning: Hype or Hope?
Hype: (n) “extravagant or intensive publicity or promotion”
Hope: (n) “expectation of fulfillment or success”
Deep Learning: Hype or Hope?
LeNet 1989: recognize zip codes, Yann Lecun, Bernhard
Boser and others, ran live in US postal service
Milestones: Digit Recognition
Convolutional NNs: AlexNet (2012): trained on 200 GB of
ImageNet Data
Milestones: Image Classification
Human performance
5.1% error
Recurrent Nets: LSTMs (1997):
Milestones: Speech Recognition
Sequence-to-sequence models with LSTMs and attention:
Source Luong, Cho, Manning ACL Tutorial 2016.
Milestones: Language Translation
Milestones: Deep Reinforcement Learning
In 2013, Deep Mind’s arcade player bests human expert
on six Atari Games. Acquired by Google in 2014,.
In 2016, Deep Mind’s
alphaGo defeats former
world champion Lee Sedol
8
Deep Learning: Is it Hype or Hope?
Yes !
Deep Learning: Is it Hype or Hope?
Other Deep Learning courses at Berkeley:
Fall 2016:
CS294-43: Special Topics in Deep Learning: Trevor Darrell,
Alexei Efros, Marcus Rohrbach, Mondays 10-12 in Newton
room (250 SDH). Emphasis on computer vision research.
Readings, presentations.
CS294-YY: Special Topics in Deep Learning: Trevor
Darrell, Sergey Levine, and Dawn Song, Weds 10-12,
https://people.eecs.berkeley.edu/~dawnsong/cs294-dl.html
Emphasis on security and other emerging applications.
Readings, presentations, project.
11
Other Deep Learning courses at Berkeley:
Spring XX:
CS294-YY: Deep Reinforcement Learning, John Shulman
and Sergey Levine, http://rll.berkeley.edu/deeprlcourse/.
Should happen in Spring 2017. Broad/introductory
coverage, assignments, project.
Stat 212B: Topics in Deep Learning: Joan Bruna
(Statistics), last offered Spring 2016. Next ??
Paper reviews and course project.
12
Learning about Deep Neural Networks
Yann Lecun quote: DNNs require: “an interplay between
intuitive insights, theoretical modeling, practical
implementations, empirical studies, and scientific analyses”
i.e. there isn’t a framework or core set of principles to
explain everything (c.f. graphical models for machine
learning).
We try to cover the ground in Lecun’s quote.
13
This Course (please interrupt with questions)
Goals:
• Introduce deep learning to a broad audience.
• Review principles and techniques (incl. visualization) for
understanding deep networks.
• Develop skill at designing networks for applications.
Materials:
• Book(s)
• Notes
• Lectures
• Visualizations
14
The role of Animation
From A. Karpathy’s cs231n notes.
15
This Course
Work:
• Class Participation: 10%
• Midterm: 20%
• Final Project (in groups): 40%
• Assignments : 30%
Audience: EECS grads and undergrads.
In the long run will probably be “normalized” to a 100/200-
course numbered “mezzanine” offering.
16
Prerequisites
• Knowledge of calculus and linear algebra, Math 53/54 or
equivalent.
• Probability and Statistics, CS70 or Stat 134. CS70 is bare
minimum preparation, a stat course is better.
• Machine Learning, CS189.
• Programming, CS61B or equivalent. Assignments will
mostly use Python.
17
Logistics
• Course Number: CS 294-129 Fall 2016, UC Berkeley
• On bCourses, publicly readable.
• CCN: 34652
• Instructor: John Canny lastname@berkeley.edu
• Time: MW 1pm - 2:30pm
• Location: 306 Soda Hall
• Discussion: Join Piazza for announcements and to ask
questions about the course
• Office hours:
• John Canny - M 2:30-3:30, in 637 Soda
18
• More info later
• Can be combined with other course projects
• Encourage “open-source” projects that can be archived
somewhere.
Course Project
Outline – Basics/Computer Vision
Outline – Applications/Advanced Topics
• Reading: the Deep Learning Book, Introduction
Some History
• 1940s-1960s: Cybernetics: Brain like electronic systems,
morphed into modern control theory and signal processing.
• 1960s-1980s: Digital computers, automata theory,
computational complexity theory: simple shallow circuits are
very limited…
• 1980s-1990s: Connectionism: complex, non-linear networks,
back-propagation.
• 1990s-2010s: Computational learning theory, graphical
models: Learning is computationally hard, simple shallow
circuits are very limited…
• 2006: Deep learning: End-to-end training, large datasets,
explosion in applications.
Phases of Neural Network Research
• Recall the LeNet was a modern visual classification network
that recognized digits for zip codes. Its citations look like this:
• The 2000s were a golden age for machine learning, and
marked the ascent of graphical models. But not so for neural
networks.
Citations of the “LeNet” paper
Second phase Third phase
Deep Learning “Winter”
• From both complexity and learning theory perspectives,
simple networks are very limited.
• Can’t compute parity with a small network.
• NP-Hard to learn “simple” functions like 3SAT formulae,
and i.e. training a DNN is NP-hard.
Why the success of DNNs is surprising
• The most successful DNN training algorithm is a version of
gradient descent which will only find local optima. In other
words, it’s a greedy algorithm. Backprop:
loss = f(g(h(y)))
d loss/dy = f’(g) x g’(h) x h’(y)
• Greedy algorithms are even more limited in what they can
represent and how well they learn.
• If a problem has a greedy solution, its regarded as an “easy”
problem.
Why the success of DNNs is surprising
• In graphical models, values in a network represent random
variables, and have a clear meaning. The network structure
encodes dependency information, i.e. you can represent rich
models.
• In a DNN, node activations encode nothing in particular, and
the network structure only encodes (trivially) how they derive
from each other.
Why the success of DNNs is surprising
Why the success of DNNs is surprising obvious
• Hierarchical representations are ubiquitous in AI. Computer
vision:
Why the success of DNNs is surprising obvious
• Natural language:
Why the success of DNNs is surprising obvious
• Human Learning: is deeply layered.
Deep expertise
Why the success of DNNs is surprising obvious
• What about greedy optimization?
• Less obvious, but it looks like many learning problems (e.g.
image classification) are actually “easy” i.e. have reliable
steepest descent paths to a good model.
Ian Goodfellow – ICLR 2015 Tutorial

More Related Content

Similar to lec01.pptx

Deep learning: the future of recommendations
Deep learning: the future of recommendationsDeep learning: the future of recommendations
Deep learning: the future of recommendationsBalázs Hidasi
 
Ml pluss ejan2013
Ml pluss ejan2013Ml pluss ejan2013
Ml pluss ejan2013CS, NcState
 
Deep Learning with Python (PyData Seattle 2015)
Deep Learning with Python (PyData Seattle 2015)Deep Learning with Python (PyData Seattle 2015)
Deep Learning with Python (PyData Seattle 2015)Alexander Korbonits
 
HML: Historical View and Trends of Deep Learning
HML: Historical View and Trends of Deep LearningHML: Historical View and Trends of Deep Learning
HML: Historical View and Trends of Deep LearningYan Xu
 
Data Con LA 2019 - State of the Art of Innovation in Computer Vision by Chris...
Data Con LA 2019 - State of the Art of Innovation in Computer Vision by Chris...Data Con LA 2019 - State of the Art of Innovation in Computer Vision by Chris...
Data Con LA 2019 - State of the Art of Innovation in Computer Vision by Chris...Data Con LA
 
Introduction to the Artificial Intelligence and Computer Vision revolution
Introduction to the Artificial Intelligence and Computer Vision revolutionIntroduction to the Artificial Intelligence and Computer Vision revolution
Introduction to the Artificial Intelligence and Computer Vision revolutionDarian Frajberg
 
Scene classification using Convolutional Neural Networks - Jayani Withanawasam
Scene classification using Convolutional Neural Networks - Jayani WithanawasamScene classification using Convolutional Neural Networks - Jayani Withanawasam
Scene classification using Convolutional Neural Networks - Jayani WithanawasamWithTheBest
 
Talk from NVidia Developer Connect
Talk from NVidia Developer ConnectTalk from NVidia Developer Connect
Talk from NVidia Developer ConnectAnuj Gupta
 
Promises of Deep Learning
Promises of Deep LearningPromises of Deep Learning
Promises of Deep LearningDavid Khosid
 
Introduction of Machine learning and Deep Learning
Introduction of Machine learning and Deep LearningIntroduction of Machine learning and Deep Learning
Introduction of Machine learning and Deep LearningMadhu Sanjeevi (Mady)
 
Deep learning 1.0 and Beyond, Part 1
Deep learning 1.0 and Beyond, Part 1Deep learning 1.0 and Beyond, Part 1
Deep learning 1.0 and Beyond, Part 1Deakin University
 
Intro to Neural Networks
Intro to Neural NetworksIntro to Neural Networks
Intro to Neural NetworksDean Wyatte
 
TensorFlow London: Cutting edge generative models
TensorFlow London: Cutting edge generative modelsTensorFlow London: Cutting edge generative models
TensorFlow London: Cutting edge generative modelsSeldon
 
Introduction to Deep Learning for Image Analysis at Strata NYC, Sep 2015
Introduction to Deep Learning for Image Analysis at Strata NYC, Sep 2015Introduction to Deep Learning for Image Analysis at Strata NYC, Sep 2015
Introduction to Deep Learning for Image Analysis at Strata NYC, Sep 2015Turi, Inc.
 
Improving search with neural ranking methods
Improving search with neural ranking methodsImproving search with neural ranking methods
Improving search with neural ranking methodsvoginip
 
Big Data Malaysia - A Primer on Deep Learning
Big Data Malaysia - A Primer on Deep LearningBig Data Malaysia - A Primer on Deep Learning
Big Data Malaysia - A Primer on Deep LearningPoo Kuan Hoong
 

Similar to lec01.pptx (20)

Deep learning: the future of recommendations
Deep learning: the future of recommendationsDeep learning: the future of recommendations
Deep learning: the future of recommendations
 
Ml pluss ejan2013
Ml pluss ejan2013Ml pluss ejan2013
Ml pluss ejan2013
 
Deep Learning with Python (PyData Seattle 2015)
Deep Learning with Python (PyData Seattle 2015)Deep Learning with Python (PyData Seattle 2015)
Deep Learning with Python (PyData Seattle 2015)
 
Phx dl meetup
Phx dl meetupPhx dl meetup
Phx dl meetup
 
HML: Historical View and Trends of Deep Learning
HML: Historical View and Trends of Deep LearningHML: Historical View and Trends of Deep Learning
HML: Historical View and Trends of Deep Learning
 
Data Con LA 2019 - State of the Art of Innovation in Computer Vision by Chris...
Data Con LA 2019 - State of the Art of Innovation in Computer Vision by Chris...Data Con LA 2019 - State of the Art of Innovation in Computer Vision by Chris...
Data Con LA 2019 - State of the Art of Innovation in Computer Vision by Chris...
 
Introduction to the Artificial Intelligence and Computer Vision revolution
Introduction to the Artificial Intelligence and Computer Vision revolutionIntroduction to the Artificial Intelligence and Computer Vision revolution
Introduction to the Artificial Intelligence and Computer Vision revolution
 
Scene classification using Convolutional Neural Networks - Jayani Withanawasam
Scene classification using Convolutional Neural Networks - Jayani WithanawasamScene classification using Convolutional Neural Networks - Jayani Withanawasam
Scene classification using Convolutional Neural Networks - Jayani Withanawasam
 
Talk from NVidia Developer Connect
Talk from NVidia Developer ConnectTalk from NVidia Developer Connect
Talk from NVidia Developer Connect
 
Promises of Deep Learning
Promises of Deep LearningPromises of Deep Learning
Promises of Deep Learning
 
Deeplearning in finance
Deeplearning in financeDeeplearning in finance
Deeplearning in finance
 
Introduction of Machine learning and Deep Learning
Introduction of Machine learning and Deep LearningIntroduction of Machine learning and Deep Learning
Introduction of Machine learning and Deep Learning
 
AlexNet
AlexNetAlexNet
AlexNet
 
Deep learning 1.0 and Beyond, Part 1
Deep learning 1.0 and Beyond, Part 1Deep learning 1.0 and Beyond, Part 1
Deep learning 1.0 and Beyond, Part 1
 
Semantic, Cognitive and Perceptual Computing -Deep learning
Semantic, Cognitive and Perceptual Computing -Deep learning Semantic, Cognitive and Perceptual Computing -Deep learning
Semantic, Cognitive and Perceptual Computing -Deep learning
 
Intro to Neural Networks
Intro to Neural NetworksIntro to Neural Networks
Intro to Neural Networks
 
TensorFlow London: Cutting edge generative models
TensorFlow London: Cutting edge generative modelsTensorFlow London: Cutting edge generative models
TensorFlow London: Cutting edge generative models
 
Introduction to Deep Learning for Image Analysis at Strata NYC, Sep 2015
Introduction to Deep Learning for Image Analysis at Strata NYC, Sep 2015Introduction to Deep Learning for Image Analysis at Strata NYC, Sep 2015
Introduction to Deep Learning for Image Analysis at Strata NYC, Sep 2015
 
Improving search with neural ranking methods
Improving search with neural ranking methodsImproving search with neural ranking methods
Improving search with neural ranking methods
 
Big Data Malaysia - A Primer on Deep Learning
Big Data Malaysia - A Primer on Deep LearningBig Data Malaysia - A Primer on Deep Learning
Big Data Malaysia - A Primer on Deep Learning
 

Recently uploaded

Types of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxTypes of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxEyham Joco
 
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Celine George
 
How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17Celine George
 
CELL CYCLE Division Science 8 quarter IV.pptx
CELL CYCLE Division Science 8 quarter IV.pptxCELL CYCLE Division Science 8 quarter IV.pptx
CELL CYCLE Division Science 8 quarter IV.pptxJiesonDelaCerna
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...Marc Dusseiller Dusjagr
 
Roles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceRoles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceSamikshaHamane
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptxVS Mahajan Coaching Centre
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsanshu789521
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationnomboosow
 
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTiammrhaywood
 
Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Celine George
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxthorishapillay1
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxSayali Powar
 
MARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupMARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupJonathanParaisoCruz
 
internship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developerinternship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developerunnathinaik
 
Biting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdfBiting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdfadityarao40181
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxOH TEIK BIN
 

Recently uploaded (20)

Types of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxTypes of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptx
 
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
 
How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17
 
CELL CYCLE Division Science 8 quarter IV.pptx
CELL CYCLE Division Science 8 quarter IV.pptxCELL CYCLE Division Science 8 quarter IV.pptx
CELL CYCLE Division Science 8 quarter IV.pptx
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
 
Roles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceRoles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in Pharmacovigilance
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha elections
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communication
 
ESSENTIAL of (CS/IT/IS) class 06 (database)
ESSENTIAL of (CS/IT/IS) class 06 (database)ESSENTIAL of (CS/IT/IS) class 06 (database)
ESSENTIAL of (CS/IT/IS) class 06 (database)
 
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
 
Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptx
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
 
MARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupMARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized Group
 
OS-operating systems- ch04 (Threads) ...
OS-operating systems- ch04 (Threads) ...OS-operating systems- ch04 (Threads) ...
OS-operating systems- ch04 (Threads) ...
 
internship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developerinternship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developer
 
Biting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdfBiting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdf
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptx
 

lec01.pptx

  • 1. CS294-129: Designing, Visualizing and Understanding Deep Neural Networks John Canny Fall 2016 CN: 34652 With considerable help from Andrej Karpathy, Fei-Fei Li, Justin Johnson, Ian Goodfellow and several others
  • 3. Hype: (n) “extravagant or intensive publicity or promotion” Hope: (n) “expectation of fulfillment or success” Deep Learning: Hype or Hope?
  • 4. LeNet 1989: recognize zip codes, Yann Lecun, Bernhard Boser and others, ran live in US postal service Milestones: Digit Recognition
  • 5. Convolutional NNs: AlexNet (2012): trained on 200 GB of ImageNet Data Milestones: Image Classification Human performance 5.1% error
  • 6. Recurrent Nets: LSTMs (1997): Milestones: Speech Recognition
  • 7. Sequence-to-sequence models with LSTMs and attention: Source Luong, Cho, Manning ACL Tutorial 2016. Milestones: Language Translation
  • 8. Milestones: Deep Reinforcement Learning In 2013, Deep Mind’s arcade player bests human expert on six Atari Games. Acquired by Google in 2014,. In 2016, Deep Mind’s alphaGo defeats former world champion Lee Sedol 8
  • 9. Deep Learning: Is it Hype or Hope?
  • 10. Yes ! Deep Learning: Is it Hype or Hope?
  • 11. Other Deep Learning courses at Berkeley: Fall 2016: CS294-43: Special Topics in Deep Learning: Trevor Darrell, Alexei Efros, Marcus Rohrbach, Mondays 10-12 in Newton room (250 SDH). Emphasis on computer vision research. Readings, presentations. CS294-YY: Special Topics in Deep Learning: Trevor Darrell, Sergey Levine, and Dawn Song, Weds 10-12, https://people.eecs.berkeley.edu/~dawnsong/cs294-dl.html Emphasis on security and other emerging applications. Readings, presentations, project. 11
  • 12. Other Deep Learning courses at Berkeley: Spring XX: CS294-YY: Deep Reinforcement Learning, John Shulman and Sergey Levine, http://rll.berkeley.edu/deeprlcourse/. Should happen in Spring 2017. Broad/introductory coverage, assignments, project. Stat 212B: Topics in Deep Learning: Joan Bruna (Statistics), last offered Spring 2016. Next ?? Paper reviews and course project. 12
  • 13. Learning about Deep Neural Networks Yann Lecun quote: DNNs require: “an interplay between intuitive insights, theoretical modeling, practical implementations, empirical studies, and scientific analyses” i.e. there isn’t a framework or core set of principles to explain everything (c.f. graphical models for machine learning). We try to cover the ground in Lecun’s quote. 13
  • 14. This Course (please interrupt with questions) Goals: • Introduce deep learning to a broad audience. • Review principles and techniques (incl. visualization) for understanding deep networks. • Develop skill at designing networks for applications. Materials: • Book(s) • Notes • Lectures • Visualizations 14
  • 15. The role of Animation From A. Karpathy’s cs231n notes. 15
  • 16. This Course Work: • Class Participation: 10% • Midterm: 20% • Final Project (in groups): 40% • Assignments : 30% Audience: EECS grads and undergrads. In the long run will probably be “normalized” to a 100/200- course numbered “mezzanine” offering. 16
  • 17. Prerequisites • Knowledge of calculus and linear algebra, Math 53/54 or equivalent. • Probability and Statistics, CS70 or Stat 134. CS70 is bare minimum preparation, a stat course is better. • Machine Learning, CS189. • Programming, CS61B or equivalent. Assignments will mostly use Python. 17
  • 18. Logistics • Course Number: CS 294-129 Fall 2016, UC Berkeley • On bCourses, publicly readable. • CCN: 34652 • Instructor: John Canny lastname@berkeley.edu • Time: MW 1pm - 2:30pm • Location: 306 Soda Hall • Discussion: Join Piazza for announcements and to ask questions about the course • Office hours: • John Canny - M 2:30-3:30, in 637 Soda 18
  • 19. • More info later • Can be combined with other course projects • Encourage “open-source” projects that can be archived somewhere. Course Project
  • 22. • Reading: the Deep Learning Book, Introduction Some History
  • 23. • 1940s-1960s: Cybernetics: Brain like electronic systems, morphed into modern control theory and signal processing. • 1960s-1980s: Digital computers, automata theory, computational complexity theory: simple shallow circuits are very limited… • 1980s-1990s: Connectionism: complex, non-linear networks, back-propagation. • 1990s-2010s: Computational learning theory, graphical models: Learning is computationally hard, simple shallow circuits are very limited… • 2006: Deep learning: End-to-end training, large datasets, explosion in applications. Phases of Neural Network Research
  • 24. • Recall the LeNet was a modern visual classification network that recognized digits for zip codes. Its citations look like this: • The 2000s were a golden age for machine learning, and marked the ascent of graphical models. But not so for neural networks. Citations of the “LeNet” paper Second phase Third phase Deep Learning “Winter”
  • 25. • From both complexity and learning theory perspectives, simple networks are very limited. • Can’t compute parity with a small network. • NP-Hard to learn “simple” functions like 3SAT formulae, and i.e. training a DNN is NP-hard. Why the success of DNNs is surprising
  • 26. • The most successful DNN training algorithm is a version of gradient descent which will only find local optima. In other words, it’s a greedy algorithm. Backprop: loss = f(g(h(y))) d loss/dy = f’(g) x g’(h) x h’(y) • Greedy algorithms are even more limited in what they can represent and how well they learn. • If a problem has a greedy solution, its regarded as an “easy” problem. Why the success of DNNs is surprising
  • 27. • In graphical models, values in a network represent random variables, and have a clear meaning. The network structure encodes dependency information, i.e. you can represent rich models. • In a DNN, node activations encode nothing in particular, and the network structure only encodes (trivially) how they derive from each other. Why the success of DNNs is surprising
  • 28. Why the success of DNNs is surprising obvious • Hierarchical representations are ubiquitous in AI. Computer vision:
  • 29. Why the success of DNNs is surprising obvious • Natural language:
  • 30. Why the success of DNNs is surprising obvious • Human Learning: is deeply layered. Deep expertise
  • 31. Why the success of DNNs is surprising obvious • What about greedy optimization? • Less obvious, but it looks like many learning problems (e.g. image classification) are actually “easy” i.e. have reliable steepest descent paths to a good model. Ian Goodfellow – ICLR 2015 Tutorial

Editor's Notes

  1. Around 5% error rate. Good enough to be useful.
  2. Founders: Demis Hassabis, Shane Legg, Mustafa Suleyman Why was this milestone particularly worrying from an ethics point of view?