Module-5 (1)
Module-5 (1)
Applications of AI –
Natural Language Processing, Text
Classification, and Information Retrieval,
Speech Recognition, Image processing and
computer vision, Robotics
Natural language processing
• Natural language processing (NLP) is a subfield
of Artificial Intelligence (AI). This is a widely used
technology for personal assistants that are used
in various business fields/areas. This technology
works on the speech provided by the user
breaks it down for proper understanding and
processes it accordingly. This is a very recent and
effective approach due to which it has a really
high demand in today’s market.
• Natural Language Processing (NLP) is a field of artificial
intelligence and linguistics that focuses on the interaction
between computers and humans through natural language.
• It encompasses the development of algorithms and
techniques that enable computers to understand, interpret,
and generate human language in a way that is meaningful
and contextually relevant
Language modeling
• Language modeling, or LM, is the use of various statistical
and probabilistic techniques to determine the probability
of a given sequence of words occurring in a sentence.
• Language models analyze bodies of text data to provide a
basis for their word predictions.
we need language models because they help computers
understand and generate human language better. Here's why:
• Communication: Language models enable computers to
understand what humans are saying or writing. For
example, they help virtual assistants like Siri or Alexa
understand spoken commands.
• Text Generation: Language models help computers
generate human-like text, such as auto-completion
suggestions when typing or generating responses for
chatbots.
• Translation: Language models assist in translating text from
one language to another, making communication easier
between people who speak different languages.
N-gram Character Models:
System Approach:
• Task Definition: The system understands that it needs to provide a short
response to the question.
• Using the Web: It searches the internet for information related to the
capital of France.
• Focus on Accuracy: Instead of giving a list of websites, it aims to give a
single correct answer.
• Rewriting Questions: It changes the question into a search query that a
search engine can understand, like "capital of France."
• Searching and Checking: It looks at short summaries of websites in the
search results to find the answer.
• Scoring the Answers: It rates each potential answer based on how often it
appears and where it appears in the search results.
• Choosing the Best Answer: It picks the answer that seems most likely to
be correct based on the question and the search results.
Speech recognition
• Speech recognition involves identifying spoken words from an acoustic signal.
• It's a mainstream AI application used in various everyday scenarios like voice
mail systems, mobile web searches, and hands-free operation.
• Challenges include ambiguity, noise, segmentation (lack of pauses between
words), coarticulation (blending sounds), and homophones.
• Speech recognition is viewed as a problem of finding the most likely
sequence of words given the observed sounds, using Bayes' rule.
• The approach involves an acoustic model (describing sounds) and a language
model (specifying word probabilities).
• Claude Shannon's noisy channel model inspired this approach,
demonstrating the possibility of recovering the original message despite
noise.
• The Viterbi algorithm is commonly used to find the most likely word
sequence.
a phone model is like a map that helps a computer understand and recognize the different
sounds in spoken language. It breaks down speech into small units called "phones,"
• Markov Model: Think of it like a series of states where you move from one
state to another based on probabilities. Each state represents a different
sound or word.
• Hidden: The states are "hidden" because we don't directly observe them.
Instead, we hear the speech but don't know exactly which state produced
each sound.
• Modeling Speech: HMMs help us model how speech transitions between
different sounds or words. They learn the probabilities of going from one
sound to another and use this knowledge to recognize spoken words.
• Key Component: HMMs are a key component of many speech recognition
systems because they can handle the variability and uncertainty present in
real-world speech. They're like a smart guesser that figures out the most
likely sequence of words based on the sounds it hears
Image processing and computer vision
• Image processing is the manipulation and analysis of images using
computational techniques. It involves converting images into digital form
and performing various operations on them to extract useful information,
enhance visual quality, or facilitate further analysis. Image processing can be
used in a wide range of applications including medical imaging, satellite
imaging, surveillance, remote sensing, and digital photography.
• [Converting images into digital form involves capturing an image through a digital
device, such as a digital camera or a scanner, and representing it in a format that can be
processed and manipulated by computers]
• Early Vision Operations: These are the initial steps in image processing,
including:
• Edge Detection: Identifying boundaries between objects or regions in an
image.
• Texture Analysis: Analyzing patterns and textures within the image.
• Computation of Optical Flow: Tracking the movement of objects or features
between frames in a sequence of images.
• Edge Detection:
definition: Edge detection is the process of identifying abrupt changes
in brightness in an image to locate boundaries between objects or
regions
– What it does: Identifies significant changes in brightness in an image,
indicating boundaries between objects or regions.
– How it works: Examines the rate of change in brightness across pixels and
detects sharp transitions.
– Example: If you have a picture of a stapler on a desk, edge detection
would outline the edges of the stapler, desk, and any other prominent
features..
• Texture:
definition: Texture refers to the repetitive patterns or visual characteristics of
surfaces in an image, aiding in object recognition based on unique surface
qualities.
– What it is: The repeating patterns or visual feel of a surface in an image.
– How it's used: Helps in recognizing objects or surfaces based on their
unique textures.
– Example: Texture could distinguish between the smooth surface of a desk
and the rough texture of a brick wall in an image.
• Optical Flow:
definition: Optical flow describes the apparent motion of objects in a
sequence of images or video by measuring their direction and speed of
movement
– What it describes: Apparent motion in a sequence of images or video.
– How it works: Measures the direction and speed of movement of features in an
image sequence.
– Example: In a video of a moving tennis player, optical flow would track the
direction and speed of the player's racket or limbs.
• Segmentation of Images:
definition: Image segmentation is the process of partitioning an image into
regions with similar visual characteristics, facilitating object recognition and
analysis.
– What it does: Divides an image into regions of similar visual properties.
– How it's done: Pixels with similar attributes (brightness, color, texture) are
grouped together.
– Example: In a picture of a beach, segmentation could separate the sand, water,
sky, and objects like umbrellas or people into distinct regions.