cogs 118a: natural computation i angela yu ajyu@ucsd ta: he (crane) huang january 5, 2009
DESCRIPTION
Cogs 118A: Natural Computation I Angela Yu [email protected] TA: He (Crane) Huang January 5, 2009. www.cogsci.ucsd.edu/~ajyu/Teaching/Cogs118A_wi10/cogs118A.html. Course Website. Policy on academic integrity Scribe schedule. 118A Curriculum Overview. Regression (3). Graphical models (8). - PowerPoint PPT PresentationTRANSCRIPT
![Page 1: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/1.jpg)
Cogs 118A: Natural Computation I
Angela Yu
TA: He (Crane) Huang
January 5, 2009
www.cogsci.ucsd.edu/~ajyu/Teaching/Cogs118A_wi10/cogs118A.html
![Page 2: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/2.jpg)
Course Website
• Policy on academic integrity
• Scribe schedule
![Page 3: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/3.jpg)
118A Curriculum OverviewRegression (3)
Graphical models (8)
Approximate inference & sampling (10, 11)
Applications to cognitive science
midterm
finalprojectdue
assignments
No final!
Sequential data (13)
Information theory, reinforcement learning
![Page 4: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/4.jpg)
118B Curriculum (Tentative) Overview
Classification (4)
Neural networks (5)
Kernel methods & SVM (6, 7)
Final project
Mixture models & EM (9)
Continuous latent variables (12)
![Page 5: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/5.jpg)
Machine Learning: 3 Types
Supervised learning
The agent also observes a sequence of labels or outputs y1, y2, …,
and the goal is to learn the mapping f: x y.
Unsupervised learning
The agent’s goal is to learn a statistical model for the inputs, p(x),to be used for prediction, compact representation, decisions, …
Reinforcement learning
The agent can perform actions a1, a2, a3, …, which affect the
state of the world, and receives rewards r1, r2, r3, … Its goal is
to learn a policy : {x1, x2, …, xt} at that maximizes <reward>.
Imagine an agent observes a sequence of inputs: x1, x2, x3, …
![Page 6: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/6.jpg)
Supervised LearningClassification: the outputs y1, y2, … are discrete labels
Dog Cat
Regression: the outputs y1, y2, … are continuously valued
x
y
![Page 7: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/7.jpg)
Applications: Cognitive ScienceObject Categorization Speech Recognition
WET VET
Face Recognition Motor Learning
![Page 8: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/8.jpg)
• Noisy sensory inputs
• Incomplete information
• Excess inputs (irrelevant)
• Prior knowledge/bias
• Changing environment
• No (one) right answer, inductive inference (open-ended)
• Stochasticity (uncertainty), not deterministic
• …
Challenges
Face Recognition
![Page 9: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/9.jpg)
An Example: Curve-Fitting
x
y
![Page 10: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/10.jpg)
Polynomial Curve-Fitting
€
y(x;w) = w0 + w1x +w2x 2 +K + wn x n = w j xj
j= 0
n
∑
Linear model
Error function (root-mean-square)
€
E(w) = y(x j ,w) − y j( )2
j= 0
n
∑ N
![Page 11: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/11.jpg)
Linear Fit?
x
y
![Page 12: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/12.jpg)
Quadratic Fit?
x
y
![Page 13: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/13.jpg)
5th-Order Polynomial Fit?
x
y
![Page 14: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/14.jpg)
10th-Order Polynomial Fit?
x
y
![Page 15: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/15.jpg)
And the Answer is…
x
y
Quadratic
![Page 16: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/16.jpg)
Training Error vs. Test (Generalization) Error
x
y
y
x
1st-order
5th-order 10th-order
2nd-order
![Page 17: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/17.jpg)
• Model complexity important
• Minimizing training error overtraining
• A better error metric: test (generalization) error
• But test error costs extra data (precious)
• Fix 1: regularization (3.1)
• Fix 2: Bayesian model comparison (3.3)
What Did We Learn?
![Page 18: Cogs 118A: Natural Computation I Angela Yu ajyu@ucsd TA: He (Crane) Huang January 5, 2009](https://reader034.vdocuments.us/reader034/viewer/2022051401/56815097550346895dbe92de/html5/thumbnails/18.jpg)
Math Review