deep learning intro

28
The History and Near Future of Deep Learning David Kammeyer Kammeyer Development [email protected] Big Data Beers 15.9.2015

Upload: kammeyer

Post on 17-Jan-2017

211 views

Category:

Data & Analytics


0 download

TRANSCRIPT

Page 1: Deep Learning Intro

The History and Near Future of Deep Learning

David Kammeyer Kammeyer Development

[email protected]

Big Data Beers 15.9.2015

Page 2: Deep Learning Intro

What’s the big Deal?

Page 3: Deep Learning Intro

Solving Problems that are Easy for Humans, Hard

for Computers

• Visual Recognition, including OCR • Speech Recognition • Natural Language Processing (Translation,

Sentiment Analysis

Page 4: Deep Learning Intro

Where did this all come from?

Page 5: Deep Learning Intro

1957: The PerceptronFrank Rosenblatt @ Cornell, MIT, ONR

Page 6: Deep Learning Intro

How the Perceptron Works

Page 7: Deep Learning Intro

Limitations and Winter #1

Perceptrons cannot learn the XOR function, or any nonmonotonic function.

Page 8: Deep Learning Intro

Multilayer Perceptrons1989: Cybenko’s Universal Approximation theorem for

Single Hidden Layer Perceptrons

Page 9: Deep Learning Intro

Backpropagation

Page 10: Deep Learning Intro

Training Methods and Winter #2

• Just because you can represent a function as a single hidden layer net doesn’t mean you can learn it (Might need more layers to be able to learn)

• SVMs provided better learning guarantees

Page 11: Deep Learning Intro

The Renaissance

Page 12: Deep Learning Intro

Convolutional Neural NetworksLeCun, 1993

Page 13: Deep Learning Intro

ImageNet 2012A. Krizhevsky’s AlexNet wins ImageNet Competition

Page 14: Deep Learning Intro

Image CaptioningKarpathy 2015

Page 15: Deep Learning Intro

What Changed?

Page 16: Deep Learning Intro

GPUs

• 40x Speedup relative to CPUs, allows the training of much larger models

than before

Page 17: Deep Learning Intro

Very Deep Models• Allows for Hierarchical Representation of Knowledge

Page 18: Deep Learning Intro

Big Data

Page 19: Deep Learning Intro

Newer TechniquesRNN, LSTM, Deep Q-Learning, New Activation

Functions, Max Pooling

Page 20: Deep Learning Intro

What’s Next?

Page 21: Deep Learning Intro

Faster Processing• Faster GPUs • FPGAs • ASICS

Page 22: Deep Learning Intro

More Recurrence, Bidirectional Hierarchies

• LSTM and RNN models have taken over at the state of the art.

• Next step is Deep Recurrent models to capture conceptual hierarchies

• Will Require new learning algorithms

Page 23: Deep Learning Intro

Hierarchical Representations in the Brain

Page 24: Deep Learning Intro

Attentional ModelsAllow the network to sequentially focus attention on a

particular part of the input

Page 25: Deep Learning Intro

Simulated (or Real) Worlds• Lots of Data Needed to Train Large Models • We’re going to have to Generate it, or Capture it from the Real World

Page 26: Deep Learning Intro

More Researchers

Page 27: Deep Learning Intro

Questions?

Dave Kammeyer Kammeyer Development

[email protected]

Page 28: Deep Learning Intro

Thanks!