Beruflich Dokumente
Kultur Dokumente
Learning
Section 1:
Section 3:
John Biddle: IBM's Watson, the Jeopardy! Playing Robot (delayed from Nov. 6)
Project
Quiz
Labs
A3 Competitions
Franz J. Kurfess
CourseOverview
Introduction
Knowledge
Intelligent
Agents
Search
Games
and Reasoning
reasoning agents
propositional logic
predicate logic
knowledge-based systems
Learning
Conclusions
Learning
ChapterOverview
Learning
Motivation
Learning
Objectives
Learning
from Observation
Learning Agents
Inductive Learning
Computational
Learning
Theory
Probably Approximately
Correct (PAC) Learning
in Neural Networks
Important
Concepts and
Terms
Chapter Summary
Learning
BridgeIn
knowledge
adaptivity
an
improved
through
performance
learning the agent can make better decisions
autonomy
without
Learning
Motivation
learning
unknown
environments
changes
the
Learning
Objectives
be
Learning
Learning
an
from experience
forecasting
theories
Learning
LearningfromObservation
Learning
Agents
Inductive Learning
Learning Decision Trees
Learning
LearningAgents
based
usually
most
inductive
learning
Learning
LearningAgentModel
Performance
Standard
Sensors
Critic
Feedback
Changes
Learning Element
Knowledge
Performance Element
Learning
Goals
Problem Generator
Agent Effectors
Environment
Learning
FormsofLearning
supervised
learning
unsupervised
reinforcement
learning
learning
the agent does not know the exact output for an input, but it receives
feedback on the desirability of its behavior
the feedback can come from an outside entity, the environment, or the
agent itself
the feedback may be delayed, and not follow the respective action
immediately
2000-2012 Franz Kurfess
Learning
Feedback
provides
both the input and the output of a component can be perceived by the
agent directly
reinforcement
learning
the agent or component knows that it did something right (or wrong),
but not what action caused it
Learning
PriorKnowledge
background
tackled
can increase performance or decrease learning time
considerably
many learning schemes assume that no prior
knowledge is available
in reality, some prior knowledge is almost always
available
but
Learning
InductiveLearning
to find a function h (the hypothesis) that
approximates a set of samples defining a function f
tries
the
supervised
learning method
relies on inductive inference, or induction
conclusions
Learning
Hypotheses
finding
since
the
e.g.
Learning
ExampleInductiveLearning1
f(x)
input-output pairs
displayed as points
in a plane
various
performance
measures
number of points
connected
minimal surface
lowest tension
x
2000-2012 Franz Kurfess
Learning
ExampleInductiveLearning2
f(x)
hypothesis is a
function consisting
of linear segments
fully incorporates
all sample pairs
x
2000-2012 Franz Kurfess
very easy to
calculate
has discontinuities
at the joints of the
segments
moderate
predictive
performance
Learning
ExampleInductiveLearning3
f(x)
hypothesis
x
2000-2012 Franz Kurfess
expressed as a
polynomial
function
incorporates all
samples
more
complicated to
calculate than
linear segments
no
discontinuities
better predictive
power
Learning
ExampleInductiveLearning4
f(x)
hypothesis is a
linear functions
does not
incorporate all
samples
extremely easy to
compute
low predictive
power
x
2000-2012 Franz Kurfess
Learning
LearningandDecisionTrees
based
Boolean
or binary classification
output
decisions
Learning
BooleanDecisionTrees
compute
branches
the
Learning
Terminology
example
or sample
describes
a positive sample has the value true for the goal predicate, a negative
sample false
sample
set
collection
training
the
validation
the
Learning
RestaurantSampleSet
Example
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
Attributes
GoalExam
Alt
Bar
Yes
Yes
No
Yes
Yes
No
No
No
No
Yes
No
Yes
No
No
Yes
No
No
Yes
Yes
No
Yes
Yes
No
Yes
No
No
No
Yes
Yes
No
No
No
Yes
Yes
No
Yes
Yes
Yes
No
Yes
No
Yes
No
Yes
No
Yes
No
Yes
Some $$$
Full
$
Some
$
Full
$
Full $$$
Some $$
None
$
Some $$
Full
$
Full $$$
None
$
Full
$
No
No
No
No
No
Yes
Yes
Yes
Yes
No
No
No
Yes
No
No
No
Yes
Yes
No
Yes
No
Yes
No
No
French 0-10
Thai 30-60
Burger 0-10
Thai 10-30
French >60
Italian 0-10
Burger 0-10
Thai 0-10
Burger >60
Italian 10-30
Thai 0-10
Burger 30-60
Yes
No
Yes
Yes
No
Yes
No
Yes
No
No
No
Yes
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
Learning
DecisionTreeExample
Patrons?
So
m
e
No
Full
EstWait?
Yes
0-10
Hungry?
Yes
Alternative?
Yes
Yes
No
No
Yes
Yes
No
60
No
Alternative?
Driveable?
No
Ye
s
>
Yes
No
Walkable?
Bar?
Yes
No
10-30
No
Yes
No
30
60
Yes
None
Yes
Ye
s
No
Learning
DecisionTreeExercise
Formulate
possible criteria
Learning
ExpressivenessofDecisionTrees
decision
trees
parity
Learning
LearningDecisionTrees
problem:
Learning
OckhamsRazor
The most likely hypothesis is the simplest one
that is consistent with all observations.
general
Learning
ConstructingDecisionTrees
in
hopefully
Learning
DecisionTreeAlgorithm
recursive
formulation
select
if
Learning
RestaurantSampleSet
Example
Alt
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
Yes
Yes
No
Yes
Yes
No
No
No
No
Yes
No
Yes
Attributes
Bar Fri
No
No
Yes
No
No
Yes
Yes
No
Yes
Yes
No
Yes
No
No
No
Yes
Yes
No
No
No
Yes
Yes
No
Yes
Goal Exam
Some $$$
Full
$
Some
$
Full
$
Full
$$$
Some $$
None
$
Some $$
Full
$
Full
$$$
None
$
Full
$
No
No
No
No
No
Yes
Yes
Yes
Yes
No
No
No
Yes
No
No
No
Yes
Yes
No
Yes
No
Yes
No
No
French 0-10
Thai 30-60
Burger 0-10
Thai 10-30
French >60
Italian 0-10
Burger 0-10
Thai 0-10
Burger >60
Italian 10-30
Thai 0-10
Burger 30-60
WillWait
Yes
No
Yes
Yes
No
Yes
No
Yes
No
No
No
Yes
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
Learning
RestaurantSampleSet
Example
Alt
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
Yes
Yes
No
Yes
Yes
No
No
No
No
Yes
No
Yes
select
Attributes
Bar Fri
No
No
Yes
No
No
Yes
Yes
No
Yes
Yes
No
Yes
No
No
No
Yes
Yes
No
No
No
Yes
Yes
No
Yes
Goal Example
Some $$$
Full
$
Some
$
Full
$
Full
$$$
Some $$
None
$
Some $$
Full
$
Full
$$$
None
$
Full
$
No
No
No
No
No
Yes
Yes
Yes
Yes
No
No
No
Yes
No
No
No
Yes
Yes
No
Yes
No
Yes
No
No
French 0-10
Thai 30-60
Burger 0-10
Thai 10-30
French >60
Italian 0-10
Burger 0-10
Thai 0-10
Burger >60
Italian 10-30
Thai 0-10
Burger 30-60
WillWait
Yes
No
Yes
Yes
No
Yes
No
Yes
No
No
No
Yes
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
best attribute
candidate 1: Pat
candidate 2: Type
PartialDecisionTree
X1, X3, X4, X6, X8, X12
X2, X5, X7, X9, X10,
X11
None
X7, X11
No
Patrons?
e
m
So
Yes
Full
X4, X12
X2, X5, X9, X10
Patrons
needs further
discrimination only for
the Full value
None and Some agree
with the WillWait goal
predicate
the next step will be
performed on the
remaining samples for
the Full value of
Patrons
Learning
RestaurantSampleSet
Example
Alt
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
Yes
Yes
No
Yes
Yes
No
No
No
No
Yes
No
Yes
select
Attributes
Bar Fri
No
No
Yes
No
No
Yes
Yes
No
Yes
Yes
No
Yes
No
No
No
Yes
Yes
No
No
No
Yes
Yes
No
Yes
Goal Example
Some $$$
Full
$
Some
$
Full
$
Full
$$$
Some $$
None
$
Some $$
Full
$
Full
$$$
None
$
Full
$
No
No
No
No
No
Yes
Yes
Yes
Yes
No
No
No
Yes
No
No
No
Yes
Yes
No
Yes
No
Yes
No
No
French 0-10
Thai 30-60
Burger 0-10
Thai 10-30
French >60
Italian 0-10
Burger 0-10
Thai 0-10
Burger >60
Italian 10-30
Thai 0-10
Burger 30-60
WillWait
Yes
No
Yes
Yes
No
Yes
No
Yes
No
No
No
Yes
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
candidate 1: Hungry
candidate 2: Type
PartialDecisionTree
Hungry
None
X7, X11
Yes
Patrons?
e
m
So
Full
X4, X12
X2, X5, X9, X10
No
Hungry?
Y
X4, X12
N
X5, X9
needs further
discrimination only for
the Yes value
No agrees with the
WillWait goal predicate
the next step will be
performed on the
remaining samples for
the Yes value of
Hungry
X2, X10
No
2000-2012 Franz Kurfess
Learning
RestaurantSampleSet
Example
Alt
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
Yes
Yes
No
Yes
Yes
No
No
No
No
Yes
No
Yes
select
Attributes
Bar Fri
No
No
Yes
No
No
Yes
Yes
No
Yes
Yes
No
Yes
No
No
No
Yes
Yes
No
No
No
Yes
Yes
No
Yes
Goal Example
Some $$$
Full
$
Some
$
Full
$
Full
$$$
Some $$
None
$
Some $$
Full
$
Full
$$$
None
$
Full
$
No
No
No
No
No
Yes
Yes
Yes
Yes
No
No
No
Yes
No
No
No
Yes
Yes
No
Yes
No
Yes
No
No
French 0-10
Thai 30-60
Burger 0-10
Thai 10-30
French >60
Italian 0-10
Burger 0-10
Thai 0-10
Burger >60
Italian 10-30
Thai 0-10
Burger 30-60
WillWait
Yes
No
Yes
Yes
No
Yes
No
Yes
No
No
No
Yes
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
candidate 1: Type
candidate 2: Friday
Learning
PartialDecisionTree
X1, X3, X4, X6, X8, X12
X2, X5, X7, X9, X10,
X11
Patrons?
None
X7, X11
Yes
e
m
So
Hungry
Full
X4, X12
X2, X5, X9, X10
No
Hungry?
Y
X4, X12
N
X5, X9
X2, X10
Yes
X10
No
.
Ital
i
Tha
h
Frenc
Type?
X4
X2
Burge
r
No
needs further
discrimination only for
the Yes value
No agrees with the
WillWait goal predicate
the next step will be
performed on the
remaining samples for
the Yes value of
Hungry
X12
Yes
Learning
RestaurantSampleSet
Example
Alt
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
Yes
Yes
No
Yes
Yes
No
No
No
No
Yes
No
Yes
select
Attributes
Bar Fri
No
No
Yes
No
No
Yes
Yes
No
Yes
Yes
No
Yes
No
No
No
Yes
Yes
No
No
No
Yes
Yes
No
Yes
Goal Example
Some $$$
Full
$
Some
$
Full
$
Full
$$$
Some $$
None
$
Some $$
Full
$
Full
$$$
None
$
Full
$
No
No
No
No
No
Yes
Yes
Yes
Yes
No
No
No
Yes
No
No
No
Yes
Yes
No
Yes
No
Yes
No
No
French 0-10
Thai 30-60
Burger 0-10
Thai 10-30
French >60
Italian 0-10
Burger 0-10
Thai 0-10
Burger >60
Italian 10-30
Thai 0-10
Burger 30-60
WillWait
Yes
No
Yes
Yes
No
Yes
No
Yes
No
No
No
Yes
X1
X2
X3
X4
X5
X6
X7
X8
X9
X10
X11
X12
candidate 1: Friday
Learning
DecisionTree
Patrons?
None
X7, X11
Yes
e
m
So
Full
X4, X12
X2, X5, X9, X10
No
Hungry?
N
Y
X4, X12
X5, X9
X2, X10
Yes
.
Ital
Burge
r
i
Tha
h
Frenc
Type?
X4
X10
X4
Yes
No
Yes
Friday?
Y
two remaining
samples can be made
consistent by selecting
Friday as the next
predicate
no more samples left
X12
X2
No
the
X2
No
Learning
PerformanceofDecisionTree
Learning
quality
of predictions
predictions
size
learn the training set, and assess the performance via the test set
of the tree
Learning
NoiseandOverfitting
the
overfitting
decision
cross-validation
Learning
EnsembleLearning
hypotheses (an ensemble) are generated,
and their predictions combined
multiple
by
each
Learning
ComputationalLearningTheory
relies
it will most likely generate a false prediction even for small numbers
of examples
hypothesis
Learning
ProbablyApproximatelyCorrect
(PAC)Learning
approximately
correct hypothesis
its
assumption
training
Learning
LearninginNeuralNetworks
Neurons
Learning
NeuralNetworks
complex
collection
Learning
NeuralNetworksandtheBrain
brain
performs information
processing operations at
various levels
neurons
NeuronDiagram
soma
dendrites
outgoing
branch
synapse
incoming
branches
axon
cell body
junction
between a
dendrite and
an axon from
another neuron
Learning
Computervs.Brain
Computer
1-1000 CPUs
Computational units
107 gates/CPU
Brain
1011 neurons
Storage units
Cycle time
Bandwidth
109 sec
1011 neurons
1014 synapses
1014 sec
1014
Learning
ComputerBrainvs.CatBrain
in
has announced a
software simulation of a
mammalian cerebral cortex
that's significantly more
complex than the cortex of a
cat. And, just like the actual
brain that it simulates, they
still have to figure out how it
works.
http://arstechnica.com/science/news/2009/11/ibm-makes-supercomput
er-significantly-smarter-than-cat.ars?utm_source=rss&utm_medium=r
2000-2012 Franz Kurfess
ss&utm_campaign=rss
http://static.arstechnica.com/cat_computer_ars.jpg
Learning
GoogleNeuralNetworklearnsabout
???
What
ale-brain-simulations-for.html
Learning
CatDiscovery
cat
discovery in NN
learned
to identify a category
of images with cats
Google blog post
https://plus.google.com/u/0/+Rese
archatGoogle/posts/EMyhnBetd2F
published
paper
http://static.googleusercontent.com/
external_content/untrusted_dlcp/res
earch.google.com/en/us/archive/uns
upervised_icml2012.pdf
http://1.bp.blogspot.com/-VENOsYD1uJc/T-nkLAiANtI/AA
AAAAAAJWc/2KCTl3OsI18/s1600/cat+detection.jpeg
Learning
ArtificialNeuronDiagram
Learning
CommonActivationFunctions
Stept(x)
Sign(x)
=
=
+1
if x >= 0, else 1
1/(1+e-x)
Sigmoid(x)
2000-2012 Franz Kurfess
if x >= t, else 0
Learning
NeuralNetworksandLogicGates
simple
Learning
NetworkStructures
in
occasionally
semantic networks
logical sentences
makes
layered
structures
networks
Learning
Perceptrons
late 1950s
capable of learning
simple functions
linearly separable
Learning
PerceptronsandLinearSeparability
0,1
0,0
AND
1,1
0,1
1,1
1,0
0,0
1,0
XOR
[Russell & Norvig, 1995]
perceptrons
XOR function
Learning
PerceptronsandLinearSeparability
Learning
PerceptronsandLearning
perceptrons
calculate
adjust
this
Learning
GenericNeuralNetworkLearning
basic
Learning
MultiLayerNetworks
research
Learning
DiagramMultiLayerNetwork
Oi
two-layer
aj
Wkj
Ik
input units Ik
Wji
network
hidden units aj
output units Oi
usually
Learning
BackPropagationAlgorithm
assigns
layers
essentially
Learning
CapabilitiesofMultiLayerNeuralNetworks
expressiveness
weaker
efficiency
training
generalization
works
Learning
CapabilitiesofMultiLayerNeural
Networks(cont.)
sensitivity
to noise
very
tolerant
they perform nonlinear regression
transparency
neural
knowledge
very
Learning
Applications
domains
recognition
control problems
series
prediction
categorization
Learning
ImportantConceptsandTerms
axon
back-propagation learning
algorithm
bias
decision tree
dendrite
feedback
function approximation
generalization
gradient descent
hypothesis
inductive learning
learning element
linear separability
machine learning
multi-layer neural network
neural network
neuron
noise
Ockhams razor
perceptron
performance element
prior knowledge
sample
synapse
test set
training set
transparency
Learning
ChapterSummary
learning
most
a
decision
Learning