1.concept Learning
1.concept Learning
Concept Learning
H1(x)
H2(x)
determines a hypothesis
H3(x) that best fits the training
.
.
examples, by searching
. space of potential
hypotheses
Concept Space
Target concept
to be learned
Other concepts
Boolean valued function is able to identity target concept over concept space
Target Concept:
The set of items/objects over which the concept is defined is called the set of
instances and denoted by X.
c: X -> {0, 1}
we can make our machine to learn from past data and make them intelligent to
identify whether an object falls into a specific category of our interest or not.
Machines can also learn from the concepts to identify whether an object
belongs to a specific category or not by processing past/training data to find a
hypothesis that best fits the training examples.
Assume the following:
Some attributes/features of the day can be:
Sky, Air Temperature, Humidity, Wind, Water, Forecast
X = set of instances
Many concepts can defined over the X.
For example, the concepts can be
- Days on which my friend Rama enjoys his favorite water sport
- Days on which my friend Rama will not go outside of his house.
- etc
Target concept —
-the concept or function to be learned
-denoted by c
-a boolean valued function defined over X
-represented as c: X → {0, 1}.
e.g., c = Days on which Rama will enjoy sports
To indicate whether Rama enjoys sports on that day, one more attribute
EnjoySport is included in dataset as shown below
X, set of instances
Our target concept is EnjoySport.
It is defined as EnjoySport : X -> {0,1}
Our goal is to predict EnjoySport for given arbitrary day with new sample values
for attributes of day, based on previous learning(from training examples)
H denotes set of all possible hypotheses that computer can consider regarding
the identify of target concept.
Examples of hypotheses:
Hypothesis Target Concept
<?, Cold, High, ?, ?, ?> Rama will enjoy sports on cold days with high
humidity
<?, ?, ?, ?, ?, ?> Every day is good day for enjoying sports
<0, 0, 0, 0, 0, 0> No day is good day for enjoying sports
Most General/Most Specific Hypothesis
X — The set of items over which the concept is defined is called the set of instances,
which we denote by X. In the current example, X is the set of all possible days, each
represented by the attributes Sky, AirTemp, Humidity, Wind, Water, and Forecast.
H — Given a set of training examples of the target concept c, the problem faced by the
learner is to hypothesize, or estimate, c. We use the symbol H to denote the set of all
possible hypotheses that the learner may consider regarding the identity of the target
concept.
For example, the below 2 hypothesis says the same but they look different.
All such hypothesis having same semantic is counted as 1. So we can have total number
of combinations as below.
Many algorithms for concept learning organize the search through the hypothesis
space by relying on a general-to-specific ordering of hypotheses.
For any instance x in X and hypothesis h in H, we say that x satisfies h if and only if
h(x) = 1.
More-General-Than-Or-Equal Relation:
more-specific-than:
FIND-S Algorithm starts from the most specific hypothesis and generalize it by
considering only positive examples.
• FIND-S algorithm ignores negative examples.
FIND-S algorithm finds the most specific hypothesis within H that is consistent
with the positive training examples.
1.Has FIND-S converged to the correct target concept?. Although FIND-S will find a
hypothesis consistent with the training data, it has no way to determine whether it
has found the only hypothesis in H consistent with the data (i.e., the correct target
concept), or whether there are many other consistent hypotheses as well.
2.Why prefer the most specific hypothesis ?. In case there are multiple hypotheses
consistent with the training examples, FIND-S will find the most specific. It is unclear
whether we should prefer this hypothesis over, say, the most general, or some other
hypothesis of intermediate generality.
3.Are the training examples consistent ?. In most practical learning problems there
is some chance that the training examples will contain at least some errors or noise.
Such inconsistent sets of training examples can severely mislead FIND-S, given the
fact that it ignores negative examples.
4.What if there are several maximally specific consistent hypotheses?. There can be
several maximally specific hypotheses consistent with the data. Find S finds only one.
Consistent Hypothesis
Definition
A hypothesis h is consistent with a set of training examples D if and only if h(x) =
c(x) for each example (x, c(x)) in D.
Assume
hypothesis h = < Sunny, Warm, ?, Strong, ?, ?>
Now for each example (x, c(x)) in D, we will evaluate h(x)
All the training examples hold h(x) = c(x). So hypothesis h2 is consistent with D.
In above example, we have two hypothesis from H and they are consistent with D.
h1=< Sunny, Warm, ?, Strong, ?, ?> and h2=< ?, Warm, ?, Strong, ?, ?>
So this set of hypothesis { h1, h2} is called a Version Space.
List-Then-Eliminate Algorithm
List-Then-Eliminate algorithm initializes the version space to contain all hypotheses
in H, then eliminates any hypothesis found inconsistent with any training example.
In the list of hypothesis, there are two extremes representing general (h1 and h2)
and specific (h6) hypothesis.
A version space can be represented with its general and specific boundary sets.
Definition — G
The general boundary G, with respect to hypothesis space H and training data D, is
the set of maximally general members of H consistent with D.
Definition — S
The specific boundary S, with respect to hypothesis space H and training data D, is
the set of minimally general (i.e., maximally specific) members of H consistent with
D.
The Candidate-Elimination algorithm represents the version space by storing only its
most general members G and its most specific members S.
Given only these two sets S and G, it is possible to enumerate all members of a
version space by generating hypotheses that lie between these two sets in general-
to-specific partial ordering over hypotheses.
A similar symptom will appear when the training examples are correct, but the
target concept cannot be described in the hypothesis representation.
What will happen if the training data contains errors ?.
Suppose, for example, that the second training example above is incorrectly presented
as a negative example instead of a positive example.
Lets run the candidate elimination algorithm on this data and see the result.
After processing all the training examples, the algorithm removes the correct target
concept from the version space.
Find-S Candidate Elimination
Only consider the most specific one. Consider all possible consistent
hypothesis.