representation learning 2016 - github pages · learning 2016 course introduction october 17, 2016...
TRANSCRIPT
RepresentationLearning 2016
RepresentationLearning 2016
Course IntroductionOctober 17, 2016
Sebastian Stober <[email protected]>
RepresentationLearning 2016
Introductions
2016-10-17Course Introduction 3
RepresentationLearning 2016
Representation LearningA Brief Intro to
2016-10-17Course Introduction 4
RepresentationLearning 2016
2016-10-17Course Introduction 5
My 1st ML Project
13. Juni 2005 Sebastian Stober - Verteidigung der Studienarbeit 11
Datenstruktur: Normalisierung• A coffee mug though the “eyes” of a road-
sign detector in 2003:
RepresentationLearning 2016
13. Juni 2005 Sebastian Stober - Verteidigung der Studienarbeit 11
Datenstruktur: Normalisierung• A coffee mug though the “eyes” of a road-sign detector in 2003:
2016-10-17Course Introduction 6
My 1st ML Project
Chapter 2. Database 11
Figure 2.4: Shape-primitives detected by the detection module.
Figure 2.5: Normalized object view representation of the input shown in figure 2.4.
representation to be used by classifier
RepresentationLearning 2016
Typical Machine Learning Workflow (for Classification)
(hand-crafted) feature
extraction
“simple” trainable classifier
data predictedlabels
train withlabeled data(supervised)
make use of domain knowledge from experts
2016-10-17Course Introduction 7
RepresentationLearning 2016
Typical Deep Learning Workflow (for Classification)
trainable feature
transform(s)trainable classifier
predictedlabels
train withlabeled data(supervised)
e.g., “pre-train” withunlabeled data(unsupervised)
data
make use of abundant data and (GPU) compute power
2016-10-17Course Introduction 8
RepresentationLearning 2016
2016-10-17 9Course Introduction[deeplearningbook.org]
RepresentationLearning 2016
2016-10-17 10Course Introduction
[deeplearningbook.org]
RepresentationLearning 2016
• learn suitable feature representationsalong with the actual learning task
• using a general-purpose learning procedure
2016-10-17Course Introduction 11
The Promise of Deep Learning
RepresentationLearning 2016
2016-10-17Course Introduction 12
An Example Deep Net
Visible layer(input pixels)
1st hidden layer
(edges)
2nd hidden layer(corners and
contours)
3rd hidden layer(object parts)
CAR PERSON ANIMALOutput
(object identity)
Figure 1.2: Illustration of a deep learning model. It is difficult for a computer to un-derstand the meaning of raw sensory input data, such as this image represented as acollection of pixel values. The function mapping from a set of pixels to an object identityis very complicated. Learning or evaluating this mapping seems insurmountable if tack-led directly. Deep learning resolves this difficulty by breaking the desired complicatedmapping into a series of nested simple mappings, each described by a different layer of
the model. The input is presented at the visible layer, so named because it contains thevariables that we are able to observe. Then a series of hidden layers extracts increasingly
abstract features from the image. These layers are called “hidden” because their valuesare not given in the data; instead the model must determine which concepts are usefulfor explaining the relationships in the observed data. The images here are visualizationsof the kind of feature represented by each hidden unit. Given the pixels, the first layercan easily identify edges, by comparing the brightness of neighboring pixels. Given thefirst hidden layer’s description of the edges, the second hidden layer can easily search forcorners and extended contours, which are recognizable as collections of edges. Given thesecond hidden layer’s description of the image in terms of corners and contours, the third
hidden layer can detect entire parts of specific objects, by finding specific collections of
contours and corners. Finally, this description of the image in terms of the object partsit contains can be used to recognize the objects present in the image. Images reproduced
with permission from Zeiler and Fergus (2014).
9
[Zeiler & Fergus 2014]
RepresentationLearning 2016
Course Rationale & Design
2016-10-17Course Introduction 13
RepresentationLearning 2016
• Think – Pair – Share
1. Think about your personal learning goals for
this course!
2. Discuss with your neighbor & create a ranking!
3. Name your most important one!
2016-10-17Course Introduction 14
Learning Goals
RepresentationLearning 2016
• in-depth theory (math) and background (why and how it works)
• hands-on experience with frameworks and design decisions, “craftsmanship”
• unsupervised RL approaches• “big picture” – applications to various topics / fields• understand inner workings• hierarchical feature representations for language• working system (portfolio)
2016-10-17Course Introduction 15
Learning Goals
RepresentationLearning 2016
• At the end of the course, you are able to …– confidently apply RL techniques
to develop a solution for a given problem– follow recent RL publications
and critically assess their contributions– formulate hypotheses and design & conduct
RL experiments to validate them– document progress & design decisions
for reproducibility and transparency
2016-10-17Course Introduction 16
Overall Learning Goals
RepresentationLearning 2016 Disclaimer
this course may not be suitable for …
– mere credit collectors– passive attendees– remote students– the lighthearted ;-)
2016-10-17Course Introduction 17
cc-by-nc-nd Lord-Psymonhttp://www.deviantart.com/art/Here-Be-Dragons-172141393
RepresentationLearning 2016
2016-10-17Course Introduction 18
Course DesignPreparation In Class Course Project
reading• book chapters• papers• blogs
(session summary)
forum discussions
weekly blog posts• new insights• hints, tricks & hacks• open questions
“last episode on RL”(3-min summary)
literaturediscussion / Q&A
small-group activity(25-60 min)
project / assignmentsdiscussion / Q&A
(25-60 min)
weekly warm-upassignments(until Nov. 20)
working in teams• up to 4 teams• scrum-style
weekly sprints
team progress blog
forum discussions
3-6h per week 3-6h per week150h totalgrading: *oral exam (20min)
RepresentationLearning 2016
• MLPs, Gradient Descent & Backpropagation• Autoencoders• Convolutional Neural Networks• Recurrent/Recursive Neural Networks• Visualization & Sonification• Regularization Techniques• Advanced Regularization Techniques• Introspection & Inception• Optimization Techniques• Advanced Training Strategies• Reinforcement Learning
2016-10-17Course Introduction 19
Topics (Tentative)
RepresentationLearning 2016
• Campus.UP (workspace “RL2016”)– blogs (course / personal / team)– forum– wiki– microblog / messaging (?)
• Slack channel (?)• GPU compute environment (medusa)
– shell access & jupyterhub for notebooks
2016-10-17Course Introduction 20
Online Tools
RepresentationLearning 2016
• 4 Maxwell Geforce Titan X GPUs• 128 GB RAM• 12 CPU cores• jupyterhub server
for notebooks
… more info in assignment #12016-10-17Course Introduction 21
GPU Compute Server
https://jupyter.org/
RepresentationLearning 2016 Jupyterhub
2016-10-17Course Introduction 22
RepresentationLearning 2016
https://campusup.uni-potsdam.de 2016-10-17Course Introduction 23
Campus.UP Workspace
RepresentationLearning 2016
• short summary blog post (in course blog) + 3-min intro recap at next session– key topics– results of the discussion– optional photos
2016-10-17Course Introduction 24
Session Summaries
last episode on“Representation Learning”
…
• rotating job!(one session per person, assignment by poll)
RepresentationLearning 2016
• document your learning / project progress– one post per week– share your experiences!– visible only to course participants
• examples:– https://deeprandommumbling.wordpress.com/– http://bartvanmerrienboer.nl/#blog
• guidelines:– https://www2.uwstout.edu/content/profdev/rubrics/blogrubric.html
2016-10-17Course Introduction 25
Your Personal Blog
RepresentationLearning 2016
• weekly progress reports for course project– similar to scrum– compare original goals with outcomes
• What has worked well?• What did not work / had to be changed?
– outline plan for next week• What would you like to try / investigate next?
• can be written up by one designated team member or in turns
2016-10-17Course Introduction 26
Team Blogs
RepresentationLearning 2016
• guide for what is covered in classdeadline: Monday morning 7am
• do not hesitate to post questions!(If you got one, you are probably not the only one!)
• post a comment if you know the answer
2016-10-17Course Introduction 27
Open Questions (Blog)
RepresentationLearning 2016
• ask – in your blog and the forum• comment / like / rate• answer• document – in your blog and the wiki
– hints, tricks & hacks• recommend
– additional readings (papers, blogs, etc.)• give (constructive) feedback
2016-10-17Course Introduction 28
Contribute!
RepresentationLearning 2016
Course Project
2016-10-17Course Introduction 29
RepresentationLearning 2016
• vision: speech-base interaction
real systems:– Siri (Apple)– Alexa (Amazon)– Cortana (Microsoft)– Google Speech– Skype Translate– …
2016-10-17Course Introduction 30
Course Project
fictional characters:– J.A.R.V.I.S. (Iron Man)– Samantha (Her)– Jane (Ender's Game)– ...
RepresentationLearning 2016
• Automatic Speech Recognition (ASR)– state-of-the-art systems use deep nets
• large-scale dataset (cc-by 4.0): 1000h corpus of read English speech: “LibriSpeech: an ASR corpus based on public domain audio books”, V. Panayotov, G. Chen, D. Povey and S. Khudanpur, ICASSP 2015.http://www.openslr.org/12/
2016-10-17Course Introduction 31
Course Project
RepresentationLearning 2016
• collaborative effort• “coopetition” (cooperative + competition)• up to 4 teams:
– form after course withdrawal deadline (Nov. 20)– self-organized (heterogeneous if possible)– scrum-like approach– focus on different aspects / strategies / tools
2016-10-17Course Introduction 32
Course Project
RepresentationLearning 2016
• available deep learning frameworks:– Theano (+ Blocks&Fuel or Keras)– Tensorflow (+ Keras)
2016-10-17Course Introduction 33
Course Project
RepresentationLearning 2016
• optional: collaborative overview-paper
• manuscript deadline: January 22, 20172016-10-17Course Introduction 34
Course Project
http://www.aes.org/conferences/2017/semantic/