Machine Learning Reftest
Machine Learning Reftest
programs
have been developed that successfully learn to recognize spoken words
(Waibel 1989; Lee 1989
We
are beginning to obtain initial models of human and animal learning and
to understand
their relationship to learning algorithms developed for computers (e.g.,
Laird et al. 1986; Anderson 1991; Qin et al. 1992; Chi and Bassock
1989; Ahn
and Brewer 1993)
Supervised learning algorithms are generally used for solving classification and
regression problems.
Unsupervised Learning:
Unsupervised learning is a type of machine learning that uses unlabeled data to train
machines. Unlabeled data doesn’t have a fixed output variable. The model learns from
the data, discovers the patterns and features in the data, and returns the output.
The training process used in unsupervised learning techniques does not need any
supervision to build models. They learn on their own and predict the output.
Reinforcement Learning
Reinforcement Learning trains a machine to take suitable actions and maximize its
rewards in a particular situation. It uses an agent and an environment to produce
actions and rewards. The agent has a start and an end state. But, there might be
different paths for reaching the end state, like a maze. In this learning technique, there is
no predefined target variable.
An example of reinforcement learning is to train a machine that can identify the shape
of an object, given a list of different objects. In the example shown, the model tries to
predict the shape of the object, which is a square in this case.
Reinforcement learning follows trial and error methods to get the desired result. After
accomplishing a task, the agent receives an award. An example could be to train a dog to catch
the ball. If the dog learns to catch a ball, you give it a reward, such as a biscuit.
Reinforcement Learning methods do not need any external supervision to train models.
Reinforcement learning algorithms are widely used in the gaming industries to build games. It is
also used to train robots to do human tasks.
Many algorithms for concept learning organize the search through the
hypothesis
space by relying on a very useful structure that exists for any concept
learning
problem: a general-to-specific ordering of hypotheses. By taking
advantage of this
naturally occurring structure over the hypothesis space, we can design
learning
algorithms that exhaustively search even infinite hypothesis spaces
without explicitly
enumerating every hypothesis. To illustrate the general-to-specific
ordering,
consider the two hypotheses
hi = (Sunny, ?, ?, Strong, ?, ?)
h2 = (Sunny, ?, ?, ?, ?, ?)
Now consider the sets of instances that are classified positive by hl and
by h2.
Because h2 imposes fewer constraints on the instance, it classifies more
instances
as positive. In fact, any instance classified positive by hl will also be
classified
positive by h2. Therefore, we say that h2 is more general than hl.
This intuitive "more general than" relationship between hypotheses can
be
defined more precisely as follows. First, for any instance x in X and
hypothesis
h in H, we say that x satisjies h if and only if h(x) = 1. We now define
the
more-general~han_or.-equalr~eola tion in terms of the sets of instances
that satisfy
the two hypotheses: Given hypotheses hj and hk, hj is more-general-
than--
equal to hk if and only if any instance that satisfies hk also satisfies hj.
NAIVE BAYES CLASSIFIER:
One highly practical Bayesian learning method is the naive Bayes learner,
often
called the naive Bayes classijier. In some domains its performance has been
shown
to be comparable to that of neural network and decision tree learning. This
section
introduces the naive Bayes classifier; the next section applies it to the
practical
problem of learning to classify natural language text documents.
The naive Bayes classifier applies to learning tasks where each instance x
is described by a conjunction of attribute values and where the target function
f ( x ) can take on any value from some finite set V. A set of training
examples of
the target function is provided, and a new instance is presented, described by
the
tuple of attribute values (al, a2.. .a,). The learner is asked to predict the target
value, or classification, for this new instance.
The Bayesian approach to classifying the new instance is to assign the most
probable target value, VMAPg, iven the attribute values (al,a 2 . . .a ,) that
describe
the instance.
VNB = argmax P (vj) n P (ai 1vj)