0% found this document useful (0 votes)
14 views

PUSHKAR

Machine Learning is a subfield of artificial intelligence focused on enabling computers to learn from data and improve their performance without explicit programming. It encompasses various types, including supervised, unsupervised, and reinforcement learning, and has applications in areas such as image recognition and natural language processing. The document discusses the importance, methodologies, challenges, and future prospects of machine learning, emphasizing its growing significance in various industries.

Uploaded by

anmolsujata368
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
14 views

PUSHKAR

Machine Learning is a subfield of artificial intelligence focused on enabling computers to learn from data and improve their performance without explicit programming. It encompasses various types, including supervised, unsupervised, and reinforcement learning, and has applications in areas such as image recognition and natural language processing. The document discusses the importance, methodologies, challenges, and future prospects of machine learning, emphasizing its growing significance in various industries.

Uploaded by

anmolsujata368
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 15

INTRODUCTION

Machine Learning is the science of getting computers to learn without being explicitly
programmed. It is closely related to computational statistics, which focuses on making
prediction using computer. In its application across business problems, machine learning
is also referred as predictive analysis. Machine Learning is closely related to
computational statistics. Machine Learning focuses on the development of computer
programs that can access data and use it to learn themselves. The process of learning
begins with observations or data, such as examples, direct experience, or instruction, in
order to look for patterns in data and make better decisions in the future based on the
examples that we provide. The primary aim is to allow the computers learn automatically
without human intervention or assistance and adjust actions accordingly.

Arthur Samuel, an early American leader in the field of computer gaming and artificial
intelligence, coined the term “Machine Learning ” in 1959 while at IBM. He defined
machine learning as “the field of study that gives computers the ability to learn without
being explicitly programmed “. However, there is no universally accepted definition for
machine learning. Different authors define the term differently. We give below two more
definitions.

 Machine learning is programming computers to optimize a performance criterion


using example data or past experience . We have a model defined up to some
parameters, and learning is the execution of a computer program to optimize the
parameters of the model using the training data or past experience. The model may
be predictive to make predictions in the future, or descriptive to gain knowledge
from data.
 The field of study known as machine learning is concerned with the question of how
to construct computer programs that automatically improve with experience.
Machine learning is a subfield of artificial intelligence that involves the development of
algorithms and statistical models that enable computers to improve their performance in
tasks through experience. These algorithms and models are designed to learn from data
and make predictions or decisions without explicit instructions. There are several types of
machine learning, including supervised learning, unsupervised learning, and

1
reinforcement learning. Supervised learning involves training a model on labeled data,
while unsupervised learning involves training a model on unlabeled data. Reinforcement
learning involves training a model through trial and error. Machine learning is used in a
wide variety of applications, including image and speech recognition, natural language
processing, and recommender systems.

2
REVIEW OF LITERATURE

A core objective of a learner is to generalize from its experience. The computational


analysis of machine learning algorithms and their performance is a branch of theoretical
computer science known as computational learning theory. Because training sets are
finite and the future is uncertain, learning theory usually does not yield guarantees of the
performance of algorithms. Instead, probabilistic bounds on the performance are quite
common. The bias–variance decomposition is one way to quantify generalization error.
For the best performance in the context of generalization, the complexity of the
hypothesis should match the complexity of the function underlying the data. If the
hypothesis is less complex than the function, then the model has underfit the data. If the
complexity of the model is increased in response, then the training error decreases. But if
the hypothesis is too complex, then the model is subject to overfitting and generalization
will be poorer. In addition to performance bounds, learning theorists study the time
complexity and feasibility of learning. In computational learning theory, a computation is
considered feasible if it can be done in polynomial time. There are two kinds of time
complexity results. Positive results show that a certain class of functions can be learned
in polynomial time. Negative results show that certain classes cannot be learned in
polynomial time.

While there has been much progress in machine learning, there are also challenges. For
example, the mainstream machine learning technologies are black-box approaches,
making us concerned about their potential risks. To tackle this challenge, we may want to
make machine learning more explainable and controllable. As another example, the
computational complexity of machine learning algorithms is usually very high and we
may want to invent lightweight algorithms or implementations. Furthermore, in many
domains such as physics, chemistry, biology, and social sciences, people usually seek
elegantly simple equations (e.g., the Schrödinger equation) to uncover the underlying
laws behind various phenomena. Machine learning takes much more time. You have to
gather and prepare data, then train the algorithm. There are much more uncertainties.
That is why, while in traditional website or application development an experienced team
can estimate the time quite precisely, a machine learning project used for example to

3
provide product recommendations can take much less or much more time than expected.
Why? Because even the best machine learning engineers don’t know how the deep
learning networks will behave when analyzing different sets of data. It also means that
the machine learning engineers and data scientists cannot guarantee that the training
process of a model can be replicated.

4
NEED OF THE STUDY

Machine learning is a type of artificial intelligence that allows software application to


become more accurate at predicting outcomes without being explicitly programmed to do
so. Machine learning algorithms use historical data as input to predict new output values.

Machine learning is important because it gives enterprises a view of trends in customer


behavior and business operational pattern, as well as supports the development of new
products. Many of today leading companies, such as Facebook, Google , make machine
learning a central part of their operations.

Future of Machine Learning is as vast as the limits of human mind. We can always keep
learning, and teaching the computers how to learn. And at the same time, wondering how
some of the most complex machine learning algorithms have been running in the back of
our own mind so effortlessly all the time. There is a bright future for machine learning.
Companies like Google, Quora, and Facebook hire people with machine learning. There
is intense research in machine learning at the top universities in the world. The global
machine learning as a service market is rising expeditiously mainly due to the Internet
revolution. The process of connecting the world virtually has generated vast amount of
data which is boosting the adoption of machine learning solutions. Considering all these
applications and dramatic improvements that ML has brought us, it doesn't take a genius
to realize that in coming future we will definitely see more advanced applications of ML,
applications that will stretch the capabilities of machine learning to an unimaginable
level.

Machine learning is one of the most exciting technologies that one would have ever come
across. As it is evident from the name, it gives the computer that which makes it more
similar to humans: The ability to learn. Machine learning is actively being used today,
perhaps in many more places than one would expect. We probably use a learning
algorithm dozen of time without even knowing it. Applications of Machine Learning
include:

5
• Web Search Engine: One of the reasons why search engines like google, bing etc
work so well is because the system has learnt how to rank pages through a complex
learning algorithm.

• Photo tagging Applications: Be it facebook or any other photo tagging application, the
ability to tag friends makes it even more happening. It is all possible because of a face
recognition algorithm that runs behind the application.

• Spam Detector: Our mail agent like Gmail or Hotmail does a lot of hard work for us
in classifying the mails and moving the spam mails to spam folder. This is again achieved
by a spam classifier running in the back end of mail application.

There are many types of Machine Learning Algorithms specific to different use cases.
As we work with datasets, a machine learning algorithm works in two stages. We
usually split the data around 20%-80% between testing and training stages. Under
supervised learning, we split a dataset into a training data and test data in Python ML.
Followings are the Algorithms of Python Machine Learning -

1. Linear Regression-
Linear regression is one of the supervised Machine learning algorithms in Python that
observes continuous features and predicts an outcome. Depending on whether it runs
on a single variable or on many features, we can call it simple linear regression or
multiple linear regression.
This is one of the most popular Python ML algorithms and often under-appreciated. It
assigns optimal weights to variables to create a line ax+b to predict the output. We
often use linear regression to estimate real values like a number of calls and costs of
houses based on continuous variables. The regression line is the best line that fits
Y=a*X+b to denote a relationship between independent and dependent variables.

6
2. Logistic Regression -
Logistic regression is a supervised classification is unique Machine Learning
algorithms in Python that finds its use in estimating discrete values like 0/1, yes/no,
and true/false. This is based on a given set of independent variables. We use a logistic
function to predict the probability of an event and this gives us an output between 0
and 1. Although it says ‘regression’, this is actually a classification algorithm. Logistic
regression fits data into a logit function and is also called logit regression.

3. Support Vector Machine (SVM)-


SVM is a supervised classification is one of the most important Machines Learning
algorithms in Python, that plots a line that divides different categories of your data.
In this ML algorithm, we calculate the vector to optimize the line. This is to ensure
7
that the closest point in each group lies farthest from each other. While you will almost
always find this to be a linear vector, it can be other than that. An SVM model is a
representation of the examples as points in space, mapped so that the examples of the
separate categories are divided by a clear gap that is as wide as possible. In addition to
performing linear classification, SVMs can efficiently perform a non-linear
classification using what is called the kernel trick, implicitly mapping their inputs into
high-dimensional feature spaces. When data are unlabeled, supervised learning is not
possible, and an unsupervised learning approach is required, which attempts to find
natural clustering of the data to groups, and then map new data to these formed groups.

4.Naïve Bayes Algorithm -


Naive Bayes is a classification method which is based on Bayes’ theorem. This
assumes independence between predictors. A Naive Bayes classifier will assume that a
feature in a class is unrelated to any other. Consider a fruit. This is an apple if it is
round, red, and 2.5 inches in diameter. A Naive Bayes classifier will say these
characteristics independently contribute to the probability of the fruit being an apple.
This is even if features depend on each other. For very large data sets, it is easy to
build a Naive Bayesian model. Not only is this model very simple, it performs better
than many highly sophisticated classification methods. Naïve Bayes classifiers are

8
highly scalable, requiring a number of parameters linear in the number of variables
(features/predictors) in a learning problem. Maximum-likelihood training can be done
by evaluating a closed-form expression, which takes linear time, rather than by
expensive iterative approximation as used for many other types of classifiers.

9
OBJECTIVES OF THE STUDY

DreamUny Education was created with a mission to create skilled software engineers for
our country and the world. It aims to bridge the gap between the quality of skills
demanded by industry and the quality of skills imparted by conventional institutes. With
assessments, learning paths and courses authored by industry experts, DreamUny helps
businesses and individuals benchmark expertise across roles, speed up release cycles and
build reliable, secure products.

Main objectives were to learn:

• How to determine and measure program complexity,

• Python Programming

• ML Library Scikit, Numpy , Matplotlib, Pandas , Theano , TensorFlow

• Statistical Math for the Algorithms.

• Learning to solve statistics and mathematical concepts.

• Supervised and Unsupervised Learning

• Classification and Regression

• ML Algorithms

• Machine Learning Programming and Use Cases.

10
METHODOLOGY

There were several facilitation techniques used by the trainer which included question
and answer, brainstorming, group discussions, case study discussions and practical
implementation of some of the topics by trainees on flip charts and paper sheets. The
multitude of training methodologies was utilized in order to make sure all the participants
get the whole concepts and they practice what they learn, because only listening to the
trainers can be forgotten, but what the trainees do by themselves they will never forget.
After the post-tests were administered and the final course evaluation forms were filled in
by the participants, the trainer expressed his closing remarks and reiterated the
importance of the training for the trainees in their daily activities and their readiness for
applying the learnt concepts in their assigned tasks. Certificates of completion were
distributed among the participants at the end.

The analytical parameter model requires human ground work, where the model designer
usually first visualizes the data and considers the science behind the system to create a
parametric model template, without specified parameters. Some popular examples would
be a polynomial model, an exponential model or a Fourier series. Then that model
template is being fitted to a set of data by minimizing error by some definition,
commonly least square error. The choice of parameter model template and error
definition causes a human bias in how the final model might turn out based on the
experiences and preferences of the designer. A way to minimize such bias would be to do
a cross-over analysis to determine the most accurate model. However, this usually
requires a lot more work and there would still be a human bias since different designer
may have different preferences when it comes to performance evaluation. Since two
different designer always could design two different analytical parametric models for the
same problem, there is no one analytical model that alone can be designed to represent
the entirety of the analytical model approach. However, by comparing the machine
learning approach with at least one of these infinite number of potential analytical
models, there could still be some important insights to gain.

By having this reference method, discussion concerning the relative performance and
suitability of different models can be done. An analytical parameter model was chosen as
11
reference method since this approach is the usual method used when creating models. It
was also the method used by ABB prior to this project to create their primary model. In
the case of the primary model, it was created by doing a pre-study to determine the first
principles of the model, then fit it to the available experimental data. These two methods
are fundamentally different and most machine learning models can be considered and
nonparametric.

12
FINDINGS and ANALYSIS

Machine Learning algorithms don’t work so well with processing raw data. Before we
can feed such data to an ML algorithm, we must preprocess it. We must apply some
transformations on it. With data preprocessing, we convert raw data into a clean data set.
To perform data this, there are 7 techniques –

1. Rescaling Data - For data with attributes of varying scales, we can rescale attributes to
possess the same scale. We rescale attributes into the range 0 to 1 and call it
normalization. We use the MinMaxScaler class from scikitlearn. This gives us values
between 0 and 1.

2. Standardizing Data - With standardizing, we can take attributes with a Gaussian


distribution and different means and standard deviations and transform them into a
standard Gaussian distribution with a mean of 0 and a standard deviation of 1.

3. Normalizing Data - In this task, we rescale each observation to a length of 1 (a unit


norm). For this, we use the Normalizer class

. 4. Binarizing Data - Using a binary threshold, it is possible to transform our data by


marking the values above it 1 and those equal to or below it, 0. For this purpose, we use
the Binarizer class.

5. Mean Removal - We can remove the mean from each feature to center it on zero.

6. One Hot Encoding - When dealing with few and scattered numerical values, we may
not need to store these. Then, we can perform One Hot Encoding. For k distinct values,
we can transform the feature into a k-dimensional vector with one value of 1 and 0 as the
rest values.

7. Label Encoding - Some labels can be words or numbers. Usually, training data is
labelled with words to make it readable. Label encoding converts word labels into
numbers to let algorithms work on them.

13
CONCLUSION

Machine learning approaches applied in systematic reviews of complex research fields


such as quality improvement may assist in the title and abstract inclusion screening
process. Machine learning approaches are of particular interest considering steadily
increasing search outputs and accessibility of the existing evidence is a particular
challenge of the research field quality improvement. Increased reviewer agreement
appeared to be associated with improved predictive performance.

Machine learning is quickly growing field in computer science. It has applications in


nearly every other field of study and is already being implemented commercially because
machine learning can solve problems too difficult or time consuming for humans to
solve. To describe machine learning in general terms, a variety models are used to learn
patterns in data and make accurate predictions based on the patterns it observes.

Machine learning is a field of artificial intelligence that deals with the design and
development of algorithms that can learn from and make predictions on data. The aim of
machine learning is to automate analytical model building and enable computers to learn
from data without being explicitly programmed to do so.

14
REFERENCE

• https://expertsystem.com/
• https://www.geeksforgeeks.org/
• https://www.wikipedia.org/
• https://www.coursera.org/learn/machine-learning
• https://machinelearningmastery.com/
• https://towardsdatascience.com/machine-learning/home

BOOKS
 Barros, Rodrigo C., Basgalupp, M. P., Carvalho, A. C. P. L. F., Freitas,
Alex A. (2012). A Survey of Evolutionary Algorithms for Decision-Tree
Induction. IEEE Transactions on Systems, Man and Cybernetics, Part C:
Applications and Reviews, 42(3), 291-312.

 Chang, C. C., and Lin, C. J. (2011). LIBSVM: A library for support


vector machines. ACM Transactions on Intelligent Systems and
Technology. 2 (3).

15

You might also like