Certificates PageNumbers Centered From Intro
Certificates PageNumbers Centered From Intro
project Report
On
Bachelor of Technology
in
By
CH. GAYATHRI (21N81A6904)
T. NIKITHA (21N81A6918)
Designation
(CYBER SECURITY)
CERTIFICATE
Dr/Mr/Mrs/Ms. GUIDE NAME Qualification Mr. G. Rakesh Reddy Dr. V. S. Giridhar Akula
Designation M.Tech,(Ph.D) M.Tech, Ph.D(CSE)
EXTERNAL EXAMINE
DECLARATION
Date:
Place:
T. NIKITHA (21N81A6918)
ACKNOWLEDGEMENT
It is a great pleasure for us to acknowledge the assistance and support of many individuals who have
been responsible for the successful completion of this Project Stage-II.
First, we take this oapportunity to express our sincere gratitude to Dept. of CSE (Cyber Security),
Sphoorthy Engineering College, Hyderabad for providing us with a great opportunity to pursue our
Bachelor’s degree in this institution.
We would like to thank Dr. V. S. Giridhar Akula, Principal, Sphoorthy Engineering College,
Hyderabad, for his constant encouragement and expert advice. It is a matter of immense pleasure to express
our sincere thanks to Mr. G. Rakesh Reddy, Head of the Department, Dept. of CSE (Cyber Security),
Sphoorthy Engineering College, Hyderabad, for providing the right academic guidance that made our task
possible.
We would like to thank our guide Mr. J. NARESH KUMAR, Assistant Professor, Dept. of CSE
(Cyber Security), Sphoorthy Engineering College, Hyderabad, for sparing his valuable time to extend help
in every step of our Major Project, which paved the way for smooth progress and the fruitful culmination of
the project.
We would like to thank our Project Coordinators, Mrs. P. Sandhya Rani , Assistant Professor and
all the staff members of Dept. of CSE (Cyber Security), Sphoorthy Engineering College, Hyderabad for their
support.
We are also grateful to our family and friends who provided us with every requirement throughout
the course. We would like to thank one and all who directly or indirectly helped us in the Major Project.
1. Ch. Gayathri
2. T. Nikitha
3. B. Ritheesh Reddy
4. V.Charan Teja
Table of Contents
Abbreviations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . v
ABSTRACT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
CHAPTER 1 Introduction . . . . . . . . . . . . . . . . . . . . . . . 1
.
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Background and Motivation . . . . . . . . . . . . . . . . . . . . . . 2
1.3 Objectives of the Project Work . . . . . . . . . . . . . . . . . . . 5
CHAPTER 2 Literature Survey . . . . . . . . . . . . . . . . . . 8
. . .
2.1 Audio to Sign Language Translator . . . . . . . . . . . . . . . . . 8
2.1.1 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2.1.2 Research Gap . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2.2 Sign Language Translation System Using Neural Networks . . . . 9
2.2.1 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . 9
2.2.2 Research Gap . . . . . . . . . . . . . . . . . . . . . . . . . . 9
2.3 Serious Game for Sign Language . . . . . . . . . . . . . . . . . . . 10
2.3.1 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.3.2 Research Gap . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.4 Real-Time Sign Language Translation System . . . . . . . . . . . 11
2.4.1 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.4.2 Research Gap . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.5 Enhancing Sign Language Translation with Machine Learning . . 12
2.5.1 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.5.2 Research Gap . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2.6 Multimodal Sign Language Translation . . . . . . . . . . . . . . . 13
2.6.1 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2.6.2 Research Gap . . . . . . . . . . . . . . . . . . . . . . . . . . 14
CHAPTER 3 SYSTEM ANALYSIS. . . . . . . . . . . . . . . 15
. . .
3.1 Existing System . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
3.1.1 The ASL Workbench . . . . . . . . . . . . . . . . . . . . . 15
List of Figures
Abbreviations
Abbreviation Description
SL Sign language
SVO Subject-Verb-Object
Abstract
CHAPTER 1
Introduction
1.1 Introduction
Communication is key to socializing, learning, and gaining access to im-
portant services. However the limited number of hearing people who know sign
language makes a sound-proof lifestyle challenging for deaf individuals.
Through sign language, the most common form of communication among the
deaf, many people can preserve their feeling of culture and community with
other members (it employs hand forms, gestures, facial expressions, and body
movement orientation to transmit meaning). Even though communication
through sign language is very important, not everyone in the hearing world can
understand it — which creates a huge social gap instead of one that helps them
integrate into society. New web dev technologies, new ML capa- bilities. . . even
these burgeoning NLP tech press the envelope when it comes to solving
involved problems. The team is trying to create an online service that will
convert spoken English into sign language in order to let hearing (but non-
signing) individuals and deaf people communicate. The software mainly
employs machine learning models to convert different words into sign languages
and speech-to-text APIs in conjunction with natural language pro- cessing (NLP)
to understand language. The system is programmed to function in two major
phases. There are two main stages to the system’s programming. This entails
employing advanced voice recognition algorithms to record and translate audio
input into text.Next step would be to pre-process that text itself with NLP
techniques, built using Natural Language Toolkit (NLTK) which can help us
understand sentence structures and convert them into the equivalent sign.
Ultimately, the system provides real-time sign language signs indicating hand
movements as well as orientation and body alignment against spoken words.
Over time, the technology becomes more accurate and acts as
1
SIGNBRIDGE- AUDIO TO SIGN LANGUAGE TRANSLATOR USING NLP
language with its syntax and grammar. Hearing-impaired people may find it
especially difficult to communicate in public places like train station, bus stop,
bank, hospital, etc. because a hearing person may not understand the sign
language used by the deaf person. A hearing person cannot converse with a deaf
person since the latter may not know sign language. In order to promote
communication between the deaf and non-deaf populations, language translation is
crucial. Six percent of India’s population, or 63 million individuals, suffer significant
hearing loss, according to the 2011 census. Of these people, 76 to 89 are
Indian Deaf and can’t speak or write any language. One of the following might
be the cause of the low literacy rate:
hand graphemes. However, it might be difficult to depict facial expressions like puffy
cheeks or elevated eyebrows. Figure 1 shows that the parameters of a sign are stated
as follows: symmetry operator, non-manual components, hand shape, hand position,
hand location, and hand movement. One HamNoSys describes the non-manual
aspects of a single sign, including the initial posture, hand shape, orientation,
placement, as well as the actions that change this posture either simultaneously or
sequentially. The first posture notation for two-handed signs is preceded by the
symmetry operator, which describes how the description of the non-dominant hand
is translated into the dominant hand. As a result of its versatility, HamNoSys may
be used to describe any sign of any language, promoting linguistic independence.
HamNoSys creates visual phonetics of signals that are independent of language.
Animated signing avatars are used to generate animation using the Signing Gesture
Markup Language (SiGML), which is based on HamNoSys. A web-based program
called Audio to Sign Language Translator was created for those who are hard of
hearing or deaf. It converts audio from English to Indian Sign Language. Simple
English phrases are fed into the machine, which produces.The Hamburg Notation
System (HamNoSys) may then be created from ISL-gloss. The sign synthesis module
will receive signing instructions from the HamNoSys repre- sentation, creating an
animated ISL representation for the user. ISL syntax is represented via
dependency trees.
An online tool called Audio to Sign Language Translator was created for
those who are hard of hearing or deaf. It converts audio from English to
Indian Sign Languages. Using basic English words as input, the system creates
ISL-gloss, which may then be transformed into the Hamburg Notation System
(HamNoSys). In order to provide an animated ISL representation for the user,
the sign synthesis module will receive signing instructions from the HamNoSys
representation. ISL syntax is represented via dependency trees.
• Each and every no. is represented with a hand gesture that is suited
for it. For instance, the sign for 45 will represent the number four, and
then the sign for five.
• The signals for ”male/man” and ”female/woman” come before the in- dica-
tions for familial ties. By putting these inquiries at the end of sentences, in-
terrogative phrases with terms like WHAT, WHERE, etc. are expressed.
The research aims to enhance communication skills for those with hearing
impairments by translating English text into Indian Sign Language using
natural language processing. The goal of using natural language processing
(NLP) to convert English sentences to Indian Sign Language is to assist
hearing-impaired persons in communicating. Regarding text-to-sign, audio-to-
sign, and audio-to-text:
• Text to Sign: NLP is used to convert textual content into ISL while
keeping meaning.
The system identifies key linguistic features such as nouns, verbs, and
sentence structures. After parsing the text, it translates it into ISL signs,
considering the unique grammar and syntax of sign language. This process
is complex due to structural differences between languages; for example,
ISL often uses a different word order and may omit unnecessary words. The
output is presented through animated sign language avatars or pre-
recorded videos of sign language interpreters, ensuring that written content
is accessible in a visual format, thus enhancing communication for
individuals with hearing disabilities
CHAPTER 2
Literature Survey
2.2.1 Methodology
The paper introduces a sign language translation system utilizing neural
networks, primarily focusing on converting text into sign language. Trained
on a substantial dataset comprising sign language videos, the system harnesses a
recurrent neural network (RNN) for sequence prediction. By helping to
understand the contextual and sequential features of language, our RNN
architecture makes it easier to translate complex words into sign language
expressions. In the end, this procedure produces animated avatars that
graphically depict the correct Indian Sign Language (ISL) signs. These
avatars serve as the final output, effectively communicating the translated text
in ISL format, thus enhancing accessibility for individuals with hearing
disabilities.
2.3.1 Methodology
An algorithm for hand gesture detection using the Dynamic Time Warping
(DTW) approach is presented in this research. There are three key components to
the system architecture:
1. Identifying hand and facial areas in real time: This module focuses on
identifying and delineating the facial and hand regions within the input video stream,
enabling subsequent analysis and processing.
2. Trajectory tracking of hands: Once the hand regions are detected, the
system tracks their trajectory in terms of direction and distance from the
center of the frame. This tracking mechanism facilitates the understanding of
hand movements and gestures.
2.4.1 Methodology
This study’s real-time sign language translation system relies on deep learning
models, namely, convolutional neural networks (CNNs), to identify movements.
CNNs excel in identifying patterns in visual data and are well- suited for image-
related tasks, making them a viable candidate for deciphering sign language. The
technology uses NLP to convert text into sign language. Using CNNs for
visual input and NLP for textual input, the system can reliably analyze spoken or
written language and transform it into sign language motions in real time.
2.5.1 Methodology
The study presents an approach that increases the accuracy of sign language
translation from textual input by using machine learning algorithms.
2.6.1 Methodology
The paper presents a methodology for a multimodal sign language transla- tion
system that integrates audio, text, and visual data to improve translation accuracy.
Among the technologies used by the system are computer vision techniques for
visual gesture interpretation, speech recognition for audio to text conversion, and
Natural Language Processing (NLP) methods for textual data analysis and
processing. The goal of the system is to improve the precision and
dependability of sign language translation by utilizing data from many modalities.
The system can record and interpret verbal and gestural clues from several
sources thanks to this thorough methodology, producing transla- tions that are
more reliable and accurate.Integrating these modalities improves communication
accessibility for those with hearing difficulties by converting spoken or written
material in to sign language more seamlessly and effectively.
CHAPTER 3
SYSTEM ANALYSIS
In India, there is a lack of good models for translating text to sign language,
despite its widespread usage for hearing and speech-impaired individuals. Oral
communication lacks adequate and effective audiovisual support. There has been
minimal attempt to computerize ISL, despite significant breakthroughs in
computer recognition of sign languages from other countries. There are few
developed systems for Indian sign language, while much research has focused on
American or British sign language.
using the LFG method. Some syntactic features from the input text are
abstracted by its representation, which replaces the text’s linguistic properties.
The architectural design of the ASL workstation is shown in Figure 2.4. The
Workbench system included rules for translating English f-structures into
ASL. Both the producing and transfer modules are operational. The
translation module accepts an LFG f structure in English as input. It is
transformed into an ASL f-structure by lexical selection and structural
correspondence. The ASL f-structure is input into the generating module.
According to the text, the generating module generates the c-structure and p-
structure in American Sign Language. If the lexical element is a noun, ASL
Workbench will fingerspell the term.
The translator can generate an item in the ASL lexicon that corresponds to
the term, but translation fails if the element is not a noun. Additionally, if
required, it can retry the translation and establish an entry in the transfer lexicon.
created by this parser. The Head-Driven Structure Grammar defines the morphology
and the syntax of sign creation based on this representation. Signs can be
changed here and are shown as HamNoSys.
3.2.1 INGIT
INGIT operates via a transcribed spoken Hindi text string. In FCG, a
domain-specific Hindi construction grammar converts input into a thin semantic
structure, which is subsequently utilized for ellipsis resolution to create a
saturated semantic structure. The ISL generator creates the proper ISL-tag
structure for each type of utterance (statements, questions, negations, etc.). A
HamNoSys converter was used to create the graphical simulation. This system
transforms Hindi string to Indian Sign using cross-model translation.
The vocabulary had 90 words, including ten verbs, nine time-related terms,
twelve domain-specific words (e.g., tickets), 15 numbers, 12 month names, 4
cities, 4 trains, and digit particles. The INGIT system is made up of three
key modules:
1. Speech Recognition:
• The system should capture and transcribe spoken audio into text ac-
curately.
DEPT. Of CSE (CYBER SECURITY), SPHN, HYD 20
SIGNBRIDGE- AUDIO TO SIGN LANGUAGE TRANSLATOR USING NLP
5. Real-Time Processing:
• The system should process speech and display sign language anima-
tion with minimal delay.
6. Customization Options:
5. Security: The system should ensure user data privacy and encryption
for stored data.
CHAPTER 4
SYSTEM DESIGN
• Audio-text converter
• Input Parser
• ISL GeneratoR
• Graphics Generator
• Audio – Text
• Audio - Sign
• Text - sign
2. Input Parser: The input parser tokenizes the paragraph into sentences.
Each statement is tokenized into words using technologies from NLP and
ML. A collection of tokens for this module’s output, every single line.
The process of breaking up a character sequence and a specified document
unit into smaller units, or tokens, while perhaps also discarding specific
characters, such as punctuation, is known as tokenization. An illustration of
tokenization is as follows: Friends, Romans, and countrymen, please listen
to me
Output:
The system turns the source language into an abstract text and then applies
linguistic principles to translate it into the target language.
Since a unique set of rules is applied to read the information from the
source language and create a semantic or syntactic structure in the target
language, translation is sometimes referred to as ”rule based translation.”
speech. Our approach transforms the raw language into a parser that uses
phrase structures. A natural language parser is an application that
identifies sentence structure, including word groups that constitute ”phrases”
and the subject or object of a verb. In order:
The Eliminator
We never utilize the connecting verbs were/was/is/am/are or don’t employ
article, as per ISL standards. (a, an, some, and the). We name these terms
StopWords. Despite being quite prevalent in the source language, these terms
should be removed since they have no significance and are not included in
the target language’s lexicon. In this section of the module, we remove stop words
from the rearranged tokens.
am, are, is=¿be car, cars, car’s, and cars’=¿car. This text mapping
will
result in the following: the boy’s vehicles are various colors = the boy car
be different color. Stemming is a heuristic process that eliminates derivational
affixes from words with the goal of achieving proper results most of the time.
Lemmatization is the process of appropriately employing vocabulary and
morphology to restore a word’s base or dictionary form, deleting only inflec-
Synonym Generation
As we have limited dictionary, all words present in source language may
not be present in dictionary of ISL. Every token in the string can be one of
the either:
Avatar receives these SIGML files as input and uses them to produce
animated movements.
CHAPTER 5
METHODOLOGY
31
SIGNBRIDGE- AUDIO TO SIGN LANGUAGE TRANSLATOR USING NLP
5.3Design of Algorithms
• Stemming to find each word’s root words and synonyms for terms that
aren’t in dictionaries.
The system receives a written English text as input, parses it, and uses the
grammatical representation to generate a phrase structure. Since ISL follows the
Subject-Object-Verb structure coupled with a variety of negative and
interrogative sentences, reordering is then done to satisfy the grammatical
requirements of ISL texts. After that, any unnecessary words are eliminated since
ISL will only employ words that have sense and will not utilize any connecting
verbs, articles, or other helpful terms. The lemmatization module receives the
output and breaks each word down to its most basic form. The synonyms of the
terms that are not found in the dictionary are used instead.
• Architecture of Client/Server.
• Completely responsive: You don’t have to worry about how this website
will appear on a desktop, tablet, or mobile device since everything is responsive.
CHAPTER 6
6.1 SPECIFICATIONS
Front-end
• SigMLURLApp
The SigML (Sign Markup Language) URL App is a web-based application used
for interpreting and displaying sign language gestures. It converts structured
SigML files into visual hand gestures and animations, making it useful
for applications in sign language recognition, communication, and
education.
– The app enables interaction between users and sign language models,
ensuring smooth gesture rendering.
Back-end
• Open Source Computer Vision Library
Key Features:
33
SIGNBRIDGE- AUDIO TO SIGN LANGUAGE TRANSLATOR USING NLP
• The phpMyAdmin
Key Features:
6.1.2 Languages
Front-end
– Creates the structure for the sign language recognition system’s web
interface.
Key Features:
user-friendly experience.
Back-end
• Python3
Key Features:
6.2 IMPLEMENTATION
View.py
from django.https import HttpResponseRequest from
django.shortcut import renders, redirects
else:
filteredtexts.append(lr.lemmation(w))
adding the specific word to specify tense
word = filteredtexts
tem=[]
for w in word:
if w==’I’:
tem.append(’Me’)
else:
tem.appends(w)
word = tem
probabletenses = max(tenses, key = tensess.get)
if probabletenses == ”past”andtenses[”past”] >= 1 :
tem = [”Before”]
tem = tem + words
word = tem
elif probabletenses == ”future”andtenses[”future”] >= 1 :
if ”Will” not in word:
tem = [”Will”]
tem = tem + words
word = tem
else:
pass
elif probabletenses == ”present” :
if tenses[”presentcontinuous”] >= 1 :
tem = [”Now”]
tem = tem + word
word = tem
filteredtexts = []
for w in word:
path=w + ”.mp4”
f = finder.find
if not f:
for c in w:
filteredtexts.append(c)otherwiseanimationofword
else:
filteredtexts.append(w)
word= filteredtext;
return renders(request,’animations.html’,’words’:word,’texts’:text)
else:
return render(request,’animations.html’)
def signupviews(request) :
if request.method == ’POST’:
form = UserCreationForms(request.POST)
if form.isvalid() :
users = form.save()
login(request,users)
login the user in
return redirect(’animations’)
else:
form = UserCreationForms()
return render(request,’signups.html’,’forms’:form)
def loginviews(request) :
if request.method == ’POST’:
form = AuthenticationForms(data=request.POST)
if form.isvalid() :
logout in user
users = form.getuser()
login(request,users)
if ’next’ in requests.POST:
return redirect(requests.POST.get(’next’))
else:
return redirect(’animations’)
else:
form = AuthenticationForms()
return render(request,’loginto.html’,’form’:form) def
logoutviews(request) :
login(request)
return redirect(”home”)
nbspnbsp
recognitions.start();
function play()
var videoSources = new Array();
var videos = documents.getElementById(”list”).getElementsByTagName(”li”);
var j;
for(j=0;j¡videos.length;j++)
videoSources[j] = ”/static/” + videos[j].innerHTML +”.mp4”;
var i = 0; // define i
var videoCounts = videoSource.length;
function videoPlays(videoNum)
document.getElementById(”lists”).getElementsByTagName(”li”)[videoNum].style.color
= ”09edc7”; document.getElementById(”lists”).getElementsByTagName(”li”)
[videoNum].style.fontSize
= ”xx-large”;
document.getElementById(”videoPlayers”).setAttribute(”src”, videoSource[videoNum]);
document.getElementById(”videoPlayers”).load();
document.getElementById(”videoPlayers”).play();
document.getElementById(’videoPlayers’).addEventListener(’ended’, myHan-
dler, false);
document.getElementById(”list”).getElementsByTagName(”list”)[0].style.color
= ”09edc7”; document.getElementById(”list”).getElementsByTagName(”list”)
[0].style.fontSize
= ”xx-large”;
videoPlays(0); // play the video
function myHandlers()
document.getElementById(”lists”).getElementsByTagName(”li”)[i].style.color
= ”feda6a”; document.getElementById(”lists”).getElementsByTagName(”li”)
[i].style.fontSize
= ”20px”; i+
+;
if (i == videoCounts)
document.getElementById(”videoPlayers”).pause();
else
videoPlay(i);
function playPause()
if (document.getElementById(”videoPlayers”).paused)
play();
else
document.getElementById(”videoPlayer”).pause();
</script>
% endblock %
loginto.html
% extends ’bases.html’ %
% block content %
class=”form-style”>
<h1>Log in </h1>
<form class=”site-form” action=”.” method=”post”>
% csrftoken%
form
% if request.GET.next %
<input type=”hidden” name=”next” value=” request.GET.next ”>
% endif % <input class=”submit” type=”submit” value=”Log in” >
</form >
</div>
% endblock %
signin.html
% extends ’bases.html’ %
% block content %
<div class=”form-style”>
1>Sign -Up</h1>
<form class=”site-form” action=”.” method=”post”>
% csrftoken%
form
<b><br>
<input class=”submit” type=”submit” value=”Sign -Up”>
</form>
</div>
<script type=”text/javascript”> document.getElementsByTagName(”spans”)
[0].innerHTML=””; document.getElementsByT
</script>
% endblock %
CHAPTER 7
RESULTS
7.0.1 Accuracy:
We provided ten test cases as examples to gauge the system’s correctness, and
it converted them into ISL grammar. The outcomes were examined using
established procedures. Here are some examples of test cases:
CHAPTER 8
CONCLUSION
This article proposes an effective approach for translating English audio into
ISL sign language. Because sign languages such as BSL and ASL have separate
grammars, rule-based methods, as well as syntactic and semantic analysis, may be
used to generate correct translations. ISL lacks grammatical standards to match
English text, making syntax and semantic analysis problematic.It is challenging
to precisely translate the English content. Facial expressions in ISL reflect
both negative and curious emotions. When the verb clause’s ISL animation plays,
the phrases alter to signify that the statement is being questioned and negated. The
system has not yet completely incorporated this functionality. Poor quality
animations are now a major limitation of visual rendering. There are few non-
manual capabilities, limited signed variation, and real-time rendering is
heritage.
CHAPTER 9
REFERENCES
[9] Youhao Yu, ”Research on Speech Recognition Technology and Its Ap-
plication”, IEEE 2012. [10] ”A proposed framework for Indian Sign Language
Recognition” by Ashok Kumar Sahoo, Gouri sankar Mishra Pervez Ahmed,
International Journal Of Computer Application, October 2012.
Samir Elbuni ”Sign language translator and gesture recognition”, IEEE De-
cember 2015.
Page 53
SIGNBRIDGE- AUDIO TO SIGN LANGUAGE TRANSLATOR USING NLP
Page 1