Viritual Mouse Using Yolo
Viritual Mouse Using Yolo
By
KALLURI VENKATESH(REGNO-40110524)
BEZAWADA YASWANTH(REGNO-40110175)
SATHYABAMA
INSTITUTE OF SCIENCE AND TECHNOLOGY
(DEEMED TO BE UNIVERSITY)
CATEGORY -1 UNIVERSITY BY UGC
Accredited with Grade “A++” by NAAC I 12B Status by UGC I Approved by AICTE
JEPPIAAR NAGAR, RAJIV GANDHI SALAI, CHENNAI - 600 119
APRIL - 2024
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING
BONAFIDE CERTIFICATE
This is to certify that this Project Report is the bonafide work of “BEZADAWADA
YASWANTH([Link]-40110175) AND KALLURI VENKATESH([Link]-
40110524)who carried out the Project entitled “VIRITUAL MOUSE” under my
supervision from November 2023 to April 2024
Internal Guide
Dr. [Link] Mayan, M.E., Ph.D
InternalExaminer ExternalExaminer
DECLARATION
DATE:
I wish to express my thanks to all Teaching and Non-teaching staff members of the
Department of Computer Science and Engineering who were helpful in many
ways for the completion of the project.
ABSTRACT
This abstract presents a novel approach for developing a virtual mouse using You
Only Look Once (YOLO) as the underlying object detection algorithm. The virtual
mouse system functions by accurately detecting the user's hand movements and
gestures in real-time through a webcam, allowing for intuitive and precise control
over a simulated mouse cursor on a computer screen. The YOLO algorithm plays
a crucial role in effectively recognizing and tracking the user's hand and fingers,
enabling seamless interaction with the virtual environment. By leveraging the
speed and accuracy of YOLO, the virtual mouse provides a responsive and natural
user interface that can be utilized for various applications such as virtual reality,
gaming, and computer accessibility. The integration of YOLO with the virtual
mouse system showcases the potential for enhancing user experience and
interaction with digital interfaces, paving the way for more intuitive and efficient
human-computer interaction paradigms. This innovative approach opens up new
possibilities for immersive computing experiences and offers a promising direction
for the development of advanced virtual input devices.
v
TABLE OF CONTENT
1 INTRODUCTION 1-4
6 TESTING 28-31
6.1 TYPES OF TESTING 28-31
7 CONCLUSION 32
8 REFERENCES 33-34
APPENDIX 35-38
vi
SCREENSHOTS 39-41
10 CERTIFICATE 48
vii
LIST OF FIGURES
SCREENSHOTS 39-41
viii
CHAPTER-1
INTRODUCTION
1.1 INTRODUCTION TO VIRITUAL MOUSE
The Virtual Mouse system incorporates state-of-the-art You Only Look Once
(YOLO) technology to enable users to interact with their computers in a novel and
intuitive way. Leveraging the power of real-time object detection, the Virtual Mouse
software detects and tracks hand movements in the camera feed, allowing users
to control their cursor and interact with applications without the need for a physical
mouse. By employing deep learning algorithms, YOLO enables precise and
responsive tracking, ensuring smooth and accurate cursor movement for users.
This innovative approach to virtual mouse control opens up new possibilities for
individuals seeking a more natural and hands-free computing experience. With
Virtual Mouse powered by YOLO, users can navigate through interfaces, click on
buttons, scroll through documents, and perform a wide range of actions simply by
moving their hands in front of a camera. This technology holds immense potential
for enhancing accessibility, enabling users with mobility impairments or physical
limitations to interact with computers effectively. As the field of computer vision
continues to advance, Virtual Mouse represents a promising application that
showcases the fusion of artificial intelligence and human-computer interaction.
Overall, the combination of virtual mouse technology and YOLO has opened up a
3
myriad of possibilities across various domains, from entertainment and gaming to
healthcare and education. By harnessing the power of computer vision and
machine learning, this technology is reshaping how users interact with digital
content and devices, offering a more natural and immersive computing
experience.
4
CHAPTER-2
LITERATURE SURVEY
5
Vasanthagokul, S., Kamakshi, K. V. G., Mudbhari, G., & Chithrakumar, T. (2022)
developed a Virtual Mouse system to enhance user experience and increase
accessibility. Presented at the 2022 4th International Conference on Inventive
Research in Computing Applications (ICIRCA) hosted by IEEE, the system
Utilizes YOLO technology to provide a seamless and efficient interaction for users.
The innovative approach aims to revolutionize accessibility features through a
virtual mouse interface, catering to a diverse range of user needs and preferences.
4. Bhakare, M. A., Rathod, M. A., Kolhe, M. A., Shardul, M. K., & Kumar, N. TO
STUDY DIFFERENT TYPES OF MACHINE LEARNING ALGORITHM FOR
GESTURE RECOGNITION.
Bhakare, M. A., Rathod, M. A., Kolhe, M. A., Shardul, M. K., and Kumar, N. are
conducting a study on different types of machine learning algorithms for gesture
recognition to develop a virtual mouse. They are implementing the YOLO (You
Only Look Once) algorithm as part of their research. This project aims to enable
users to control a computer mouse using hand gestures, potentially improving the
accessibility and user experience of computing devices. The team's work could
lead to innovative advancements in human-computer interaction technology.
5. Bahri, H., Krčmařík, D., & Kočí, J. (2019, December). Accurate object
detection system on hololens using yolo algorithm. In 2019 International
Conference on Control, Artificial Intelligence, Robotics & Optimization
(ICCAIRO) (pp. 219-224). IEEE.
Bahri, H., Krčmařík, D., & Kočí, J. (2019) developed an accurate object detection
system for virtual mouse using the YOLO algorithm. Their work was presented at
the 2019 International Conference on Control, Artificial Intelligence, Robotics &
Optimization (ICCAIRO) organized by IEEE. The system was designed to work
specifically on Hololens, demonstrating the potential of YOLO for real-time object
detection in augmented reality applications. By leveraging YOLO's efficiency and
accuracy, the researchers showcased the capability of their system to enhance
user interactions in AR environments.
6
6. Pansare, J., Aochar, G., Salvi, T., Braganza, J., Tiwari, A., & Kesharkar, D.
Effective Computer Vision Techniques for Real-Time Hand Gesture
Recognition and Detection.
Pansare, J., Aochar, G., Salvi, T., Braganza, J., Tiwari, A., and Kesharkar, D.
collaborated on a research project titled "Effective Computer Vision Techniques for
Real-Time Hand Gesture Recognition and Detection." The study explores the use
of YOLO (You Only Look Once) for creating a virtual mouse through hand
gestures in real-time. The team's work aims to enhance user experience and
accessibility by providing an innovative alternative to traditional input
[Link] findings showcase the potential of computer vision technologies
Inrevolutionizing human-computer interaction.
[Link]-Oliveira, P., Arriaga, P., Paiva, A., & Hoffman, G. (2019). Guide to
build YOLO, a creativity-stimulating robot for children. HardwareX, 6,
e00074.
The study by Alves-Oliveira et al. (2019) introduces a guide for constructing a
creativity-stimulating robot named YOLO for children. This robot is intended to
serve as a virtual mouse using YOLO technology. The research, published in
HardwareX, provides detailed instructions on the hardware setup and
implementation of YOLO for creating an interactive and engaging tool for young
learners. By following this guide, educators and researchers can support children's
learning and development through hands-on experiences with innovative
technology.
8. Chua, S. D., Chin, K. R., Lim, S. F., & Jain, P. (2022). Hand gesture control
for human–computer interaction with Deep Learning. Journal of Electrical
Engineering & Technology, 17(3), 1961-1970.
Chua, S. D., Chin, K. R., Lim, S. F., & Jain, P. (2022) discussed hand gesture
control for human-computer interaction using Deep Learning in their article
published in the Journal of Electrical Engineering & Technology. The study
focused on developing a virtual mouse system that leverages YOLO (You Only
Look Once) technology for real-time hand detection and tracking. The proposed
system aims to enhance the user experience by enabling intuitive interaction with
computers through hand gestures. The research findings highlight the potential of
7
Deep Learning algorithms for improving human-computer interaction in various
applications.
10. Sun, G., Lyu, C., Cai, R., Yu, C., Sun, H., Schriver, K. E., ... & Li, X. (2021).
DeepBhvTracking: a novel behavior tracking method for laboratory animals based
on deep learning. Frontiers in behavioral neuroscience, 15, 750894.
8
CHAPTER 3
REQUIRMENT ANALYSIS
3.1EXISTING SYSTEM
The existing system for human trafficking identification and prediction is currently
fragmented and inefficient, relying on manual data analysis and information
sharing by law enforcement agencies and non-profit organizations. This approach
lacks data integration and predictive analysis capabilities, resulting in delays in
identifying trafficking patterns and perpetrators. Moreover, the system often
overlooks crucial indicators and relies on human intuition and limited resources for
decision-making. This leads to many instances of human trafficking going
undetected or not being addressed promptly. The lack of standardized protocols
and technology integration hampers collaboration and information sharing among
stakeholders. The absence of a centralized system for data collection, analysis,
and prediction further exacerbates these challenges. To address these limitations,
there is an urgent need for a comprehensive human trafficking identification and
prediction system that utilizes machine learning techniques like YOLO to analyze
vast amounts of data, identify patterns, and predict potential trafficking activities.
Such a system would enable proactive intervention, improve resource allocation,
and enhance collaboration among stakeholders for more effective combating of
human [Link] of the key disadvantages of existing human trafficking
identification and prediction systems is their heavy reliance on manual and
subjective processes. The lack of standardized methods and criteria often results
in inconsistencies and potential biases. Moreover, these systems struggle to
efficiently analyze large volumes of data, leading to delays in detecting trafficking
activities. Additionally, the underutilization of advanced technologies like machine
learning limits the accuracy and effectiveness of these systems. To address these
shortcomings, a comprehensive human trafficking identification and prediction
9
significantly enhance effectiveness and efficiency in combating human trafficking.
Through the integration of advanced technologies and collaborative efforts among
various agencies and organizations, a more streamlined and accurate approach
can be achieved in identifying and predicting human trafficking activities.
address the complex challenges posed by human trafficking with the ultimate goal
of safeguarding vulnerable populations and combating modern-day [Link]
proposed Comprehensive Human Trafficking Identification and Prediction System
leveraging machine learning techniques like YOLO offers a groundbreaking
approach with numerous benefits. By utilizing advanced algorithms, the system
can identify subtle patterns and indicators imperceptible to human analysts,
10
significantly enhancing the accuracy and efficiency in detecting potential human
trafficking cases. Through the integration of diverse data sources such as social
media, financial transactions, and law enforcement records, the system provides a
holistic overview of potential trafficking activities, enabling a more comprehensive
strategy to combat this serious issue. Moreover, its predictive capabilities
empower authorities to take preemptive actions to prevent trafficking incidents,
thereby offering better protection to at-risk individuals and potential victims. The
system's ability to continuously learn and adapt based on new data and emerging
trends ensures its effectiveness in dynamically addressing human trafficking
challenges. Additionally, its automated nature streamlines operations, reducing the
strain on human resources and enabling law enforcement agencies and NGOs to
focus their efforts more efficiently on investigating and combating human
trafficking. Overall, the Comprehensive Human Trafficking Identification and
Prediction System is a powerful tool in the fight against human trafficking,
providing a proactive, data-driven approach that can significantly enhance
identification, prevention, and intervention efforts.
effective predictive model for identifying potential human trafficking activities. With
a growing demand for solutions to combat human trafficking, there are economic
opportunities for funding and partnerships. Operationally, the system can
streamline the identification process and offer valuable insights for law
enforcement agencies and organizations working in this field. By dividing the
project into manageable phases, timely delivery and implementation can be
ensured. Overall, the feasibility study will provide a comprehensive analysis of the
project's viability and establish a roadmap for the successful development and
deployment of the Human Trafficking Identification and Prediction System using
11
Machine Learning techniques.
Additionally, the utilization of advanced algorithms like YOLO can improve the
effectiveness of anti-human trafficking efforts by detecting intricate patterns and
relationships. This enhanced capability can augment traditional methods and
increase the system's overall impact.
12
In conclusion, the economic feasibility of implementing a machine learning-based
human trafficking identification and prediction system like YOLO involves
analyzing cost savings, benefits, revenue streams, and organizational reputation.
Demonstrating a substantial return on investment and long-term sustainability is
crucial for establishing the value of such a system in combating human trafficking.
In the context of using YOLO for human trafficking identification and prediction,
several factors need to be considered. Firstly, it's important to assess the
compatibility of YOLO with the data sources available for training the model.
YOLO is known for its real-time object detection capabilities, but the data sources
used must align with the required input format and quality to ensure accurate
results.
Additionally, the technical feasibility analysis should focus on the scalability of the
system to handle large volumes of data. Human trafficking patterns can vary
across regions and populations, so the system needs to be able to process
diverse datasets efficiently. Considerations like data security, privacy, and
regulatory compliance should also be taken into account to meet ethical and legal
standards.
Integrating multiple data sources, such as law enforcement databases and social
media data, is crucial for enhancing the system's effectiveness. YOLO's ability to
detect objects in images and videos can be utilized to analyze diverse data
streams and identify potential human trafficking activities.
13
The system's architecture should be designed to support real-time data processing
and continuous learning, enhancing the predictive model's accuracy over time. By
carefully evaluating the technical feasibility of implementing YOLO in a Human
Trafficking Identification and Prediction System, you can address challenges,
leverage opportunities, and ultimately contribute to the fight against human
trafficking using advanced machine learning techniques.
Scalability and sustainability are key factors to consider. The system should adapt
to evolving needs in the fight against human trafficking. Regular maintenance and
updates will be necessary for long-term effectiveness.
I'm sorry, but it seems like there might have been a misunderstanding in your
request. The information provided regarding social feasibility in the context of
implementing a Comprehensive Human Trafficking Identification and Prediction
14
System using Machine Learning techniques seems detailed and relevant, but the
addition of "in 300 words for virtual mouse using YOLO" appears to be a technical
specification or requirement related to object detection with the YOLO algorithm,
which may not directly correspond to the social feasibility aspects discussed.
If you need assistance with incorporating YOLO (You Only Look Once) object
detection methodology into the development of a human trafficking identification
system, please clarify your requirements or context so I can provide information or
guidance tailored to your needs. Thank you for your understanding.
Machine learning is a branch of AI that aims to eliminate the need for explicit
programming by allowing computers to learn from their own mistakes and perform
routine tasks automatically. However, "artificial intelligence" (AI) encompasses a
broader definition of "machine learning," which is the method through which
computers are trained to recognize visual and auditory cues, understand spoken
language, translate between languages, and ultimately make significant decisions
15
on their own.
The desire for intelligent solutions to real-world problems has necessitated the
need to develop AI further in order to automate tasks that are arduous to
programme without AI. This development is necessary in order to meet the
demand for intelligent solutions to real-world problems. Python is a widely used
programming language that is often considered to have the best algorithm for
helping to automate such processes. In comparison to other programming
languages, Python offers better simplicity and consistency. In addition, the
•Reliability
Most software developers seek simplicity and consistency in Python. Python code
is concise and readable, simplifying presentation. Compared to other programming
languages, developers can write code quickly. Developers can get community
feedback to improve their product or app. Python is simpler than other
programming languages, therefore beginners may learn it quickly. Experienced
developers may focus on innovation and solving real-world problems with machine
learning because they can easily design stable and trustworthy solutions.
•Easily Executable
16
Developers choose Python because it works on many platforms without change.
Python runs unmodified on Windows, Linux, and macOS. Python is supported on
all these platforms, therefore you don't need a Python expert to comprehend it.
Python's great executability allows separate applications. Programming the app
requires only Python. Developers benefit from this because some programming
languages require others to complete the job. Python's portability cuts project
execution time and effort.
CHAPTER-4
SYSTEM DESIGN
4.1SYSTEM ARCHITECTURE
17
System architecture is the foundation of a system that outlines the components,
interactions, and relationships within it. It dictates how these elements collaborate
to fulfill the system's functionality, managing data flow, communication, and overall
structure. This blueprint is crucial for achieving optimal performance, scalability,
security, and reliability. By providing a roadmap for development, system
architecture guides developers in building and maintaining the system effectively.
A well-designed architecture ensures that the system meets its requirements
efficiently and serves as a valuable tool for understanding the system's inner
workings.
18
diagram, it visualizes the interactions between different components and actors,
providing a blueprint for understanding how the system functions. This diagram
helps in illustrating the tasks and functionalities required to effectively identify and
predict human trafficking incidents using YOLO's machine learning algorithms.
19
the flow of processes involved. The diagram would illustrate activities such as data
collection through a virtual mouse using YOLO (You Only Look Once) for object
detection, data preprocessing, model training with machine learning algorithms,
prediction generation based on identified patterns, and outcome evaluation for
20
Identification and Prediction System using Machine Learning
techniques would visually depict how different system components
like data collection, preprocessing, feature extraction, model training,
and prediction are connected. This diagram showcases the flow of
events, messages, and data exchanges among the modules,
showing how data moves through the system and machine learning
algorithms are employed to analyze and predict human trafficking
patterns. Stakeholders can grasp the system's operations and the
process of identifying and forecasting human trafficking incidents
from this diagram. Virtual mouse employing YOLO can enhance
real-time object detection and tracking capabilities in such a system.
CHAPTER-5
MODULE DESCRIPTION
21
5.1 DATA COLLECTION AND ANNOTATION
22
[Link] DETECTION WITH YOLO
Object detection using YOLO for a virtual mouse involves utilizing the YOLO (You Only
Look Once) algorithm to detect and track the movements of a user's hand or pointer on
the screen. YOLO is a popular real-time object detection system that processes images
quickly and accurately by dividing the image into grids and predicting bounding boxes
and class probabilities for each grid cell. In the context of a virtual mouse, YOLO can be
trained to recognize gestures or movements associated with hand movements, such as
pointing, clicking, or dragging. By continuously analyzing the video feed from a camera
or input device, YOLO can accurately detect the hand movements and translate them
into corresponding actions on the screen, effectively creating a virtual mouse interface.
This technology can be applied in various interactive applications, such as virtual reality
environments, smart whiteboards, or touchless interactions with displays for improved
user experience and accessibility.
23
filtering and cleaning processes can further optimize the system's performance and
ensure its effectiveness in various scenarios.
24
virtual mouse application to enable real-time detection and tracking of virtual mouse
movements with high precision and efficiency.
25
Only Look Once) could include regular model retraining with updated and diverse
datasets to enhance accuracy and performance. Secondly, optimizing the network
architecture by fine-tuning hyperparameters, layers, and activation functions can result
in better detection capabilities. Additionally, implementing data augmentation
techniques like rotation, scaling, and flipping can help the model generalize better to
various virtual mouse interactions. Lastly, incorporating post-processing techniques
such as non-maximum suppression to filter out redundant detections and improve
overall precision. By iteratively refining the model through these strategies, the virtual
mouse detection system can become more robust, efficient, and capable of accurately
detecting the mouse in different virtual environments.
DATABASE
To prepare the training data for a YOLO model for a virtual mouse, start by
26
collecting a diverse dataset of images containing various hand and finger positions that
simulate mouse movements. Annotate these images with bounding boxes around the
hands or fingers to indicate the region of interest. Ensure that the annotations are
consistent and accurately represent the positions of the mouse cursor. Augment the
dataset by applying transformations such as rotation, scaling, and flipping to increase the
variability of the training data. Normalize the input images and labels to ensure uniformity
in the data. Finally, split the dataset into training and validation sets to evaluate the
model's performance. By following these steps, you can effectively prepare the training
data for training a YOLO model for a virtual mouse.
SECURITY
Three model performance evaluation metrics for a virtual mouse system using YOLO (You
Only Look Once) can include Average Precision (AP), Intersection over Union (IoU), and
Mean Average Precision (mAP). Average Precision assesses the accuracy of object
detection by calculating the area under the precision-recall curve. IoU measures the
overlap between the predicted bounding boxes and the ground truth boxes to evaluate
localization accuracy. Lastly, mAP provides an overall performance score by averaging
AP across different categories. These metrics are crucial for understanding the efficiency
of the YOLO model in accurately identifying and tracking the virtual mouse in a computer
vision environment, ensuring optimal performance and reliability for the system.
27
CHAPTER 6
Discovering and fixing such problems is what testing is all about. The purpose of testing is
to find and correct any problems with the final product. It's a method for evaluating the
quality of the operation of anything from a whole product to a single component. The goal
of stress testing software is to verify that it retains its original functionality under extreme
circumstances. There are several different tests from which to pick. Many tests are
available since there is such a vast range of assessment options.
Who Performs the Testing: All individuals who play an integral role in the software
development process are responsible for performing the testing. Testing the software is
the responsibility of a wide variety of specialists, including the End Users, Project
Manager, Software Tester, and Software Developer.
When it is recommended that testing begin: Testing the software is the initial step in the
process. begins with the phase of requirement collecting, also known as the Planning
phase, and ends with the stage known as the Deployment phase. In the waterfall model,
the phase of testing is where testing is explicitly arranged and carried out. Testing in the
incremental model is carried out at the conclusion of each increment or iteration, and the
entire application is examined in the final test. When it is appropriate to halt testing:
Testing the programme is an ongoing activity that will never end. Without first
putting the software through its paces, it is impossible for anyone to guarantee that it is
completely devoid of errors. Because the domain to which the input belongs is so
expansive, we are unable to check every single input.
UNIT TESTING
Implementing a virtual mouse using YOLO (You Only Look Once) involves detecting hand
gestures and movements to control the cursor on a screen in real-time.
28
Testcase1: Testing the YOLO object detection module to accurately detect and classify
hand gestures such as pointing, dragging, and clicking.
Testcase2: Testing the cursor control module with simulated hand movement data to
verify that the cursor responds correctly and smoothly to different gestures.
Testcase3: Testing the overall system performance by tracking the cursor's accuracy and
responsiveness while interacting with different applications and interfaces.
By conducting these test cases during unit testing, developers can ensure that each
component of the virtual mouse system functions correctly, leading to a user-friendly and
efficient control mechanism for computer systems.
INTEGRATION TESTING
To develop a virtual mouse using YOLO (You Only Look Once) for comprehensive human
trafficking identification and prediction system, integration testing is crucial to ensure
seamless functionality. Here are 3 test cases for testing the integration of the virtual
mouse with the system:
Testcase1: Validate that the virtual mouse accurately detects and tracks human
movement within the system's user interface, ensuring precise input and interaction with
the machine learning model for human trafficking identification.
Testcase2: Test the integration of the virtual mouse with external components, such as
web browsers or applications, to verify smooth navigation and control for users accessing
the system for human trafficking prediction purposes.
Testcase3: Verify the compatibility of the virtual mouse with various operating systems
and devices to ensure consistent performance and user experience across different
platforms, enhancing the accessibility and usability of the system.
29
FUNCTIONAL TESTING
Functional testing for a virtual mouse system using YOLO (You Only Look Once) involves
testing the system's functionality to ensure it accurately detects and tracks the movement
of an on-screen cursor. Testcases are crucial for evaluating the system's performance
and user interaction experience.
Testcase1: Verify that the virtual mouse accurately tracks the movement of the user's
hand gestures in real-time using YOLO object detection.
Testcase2: Validate the accuracy of the virtual mouse's clicking functionality by assessing
its ability to click on specific on-screen elements with precision.
Testcase3: Test the responsiveness of the virtual mouse system to various hand
movements and gestures to ensure smooth and accurate cursor control for users.
By executing these testcases, the functional testing process can help ensure that the
virtual mouse system using YOLO operates effectively accurately, and
responsively, providing users with a seamless and intuitive interaction experience.
Testcase2: Input different lighting conditions and background environments to test the
system's robustness in detecting the virtual mouse pointer with high accuracy.
30
By testing the Virtual Mouse application using Black Box Testing methodologies, we aim
to ensure the system's functionality and performance without the need to understand its
internal algorithms and processes, focusing on its overall behavior and user interaction.
When applying White box testing to a comprehensive human trafficking identification and
prediction system that utilizes machine learning techniques, it is crucial to examine the
internal workings of the system meticulously. Here are the test cases for a virtual mouse
using YOLO.
Testcase1: Input validation - Confirm that the virtual mouse system effectively handles
invalid inputs, such as incorrect image formats or non-existent objects, providing
appropriate error messages and preventing crashes.
Testcase2: Model accuracy - Evaluate the accuracy of the YOLO machine learning model
by inputting images containing various objects and assessing if the virtual mouse can
accurately detect and track the objects in real-time.
Testcase3: Algorithm flow - Trace the flow of algorithms implemented in the virtual mouse
system, ensuring that image processing, object detection, and virtual cursor movement
occur in the expected sequence, leading to smooth and accurate interaction with the
virtual environment.
31
CHAPTER 7
CONCLUSION
7.1 CONCLUSION
32
REFERENCES
[1] Krishnamoorthi, M., Gowtham, S., Sanjeevi, K., & Vishnu, R. R. (2022, July). Virtual
Mouse Using YOLO. In 2022 International Conference on Innovative Computing,
Intelligent Communication and Smart Electrical Systems (ICSES) (pp. 1-7). IEEE.
[2] Karthick, S., Dinesh, M., & Jayapandian, N. (2023, December). Artificial Intelligence
Based Enhanced Virtual Mouse Hand Gesture Tracking Using Yolo Algorithm. In 2023
IEEE 2nd International Conference on Data, Decision and Systems (ICDDS) (pp. 1-6).
IEEE.
[3] Vasanthagokul, S., Kamakshi, K. V. G., Mudbhari, G., & Chithrakumar, T. (2022,
September). Virtual Mouse to Enhance User Experience and Increase Accessibility. In
2022 4th International Conference on Inventive Research in Computing Applications
(ICIRCA) (pp. 1266-1271). IEEE.
[4] Bhakare, M. A., Rathod, M. A., Kolhe, M. A., Shardul, M. K., & Kumar, N. TO
STUDY DIFFERENT TYPES OF MACHINE LEARNING ALGORITHM FOR GESTURE
RECOGNITION.
[5] Bahri, H., Krčmařík, D., & Kočí, J. (2019, December). Accurate object detection
system on hololens using yolo algorithm. In 2019 International Conference on Control,
Artificial Intelligence, Robotics & Optimization (ICCAIRO) (pp. 219-224). IEEE.
[6] Pansare, J., Aochar, G., Salvi, T., Braganza, J., Tiwari, A., & Kesharkar, D.
Effective Computer Vision Techniques for Real-Time Hand Gesture Recognition and
Detection.
33
[7] Alves-Oliveira, P., Arriaga, P., Paiva, A., & Hoffman, G. (2019). Guide to build
YOLO, a creativity-stimulating robot for children. HardwareX, 6, e00074.
34
[8] Chua, S. D., Chin, K. R., Lim, S. F., & Jain, P. (2022). Hand gesture control
for human–computer interaction with Deep Learning. Journal of Electrical
Engineering & Technology, 17(3), 1961-1970.
[10] Sun, G., Lyu, C., Cai, R., Yu, C., Sun, H., Schriver, K. E., ... & Li, X. (2021).
DeepBhvTracking: a novel behavior tracking method for laboratory animals
based on deep learning. Frontiers in behavioral neuroscience, 15, 750894.
35
APPENDIX
[Link] CODE
import streamlit as st
import cv2
import mediapipe as mp
import pyautogui
import math
[Link] = False
video_capture = [Link](0)
[Link]("Live Video Stream with Hand Cursor, Click")
video_placeholder = [Link]()
prev_palm_x, prev_palm_y = 0, 0
prev_index_thumb_touching = False
prev_index_middle_touching = False
prev_thumb_middle_touching = False
while True:
36
ret, frame = video_capture.read()
if not ret:
[Link]("Error: Unable to capture video.")
break
if results.multi_hand_landmarks:
for hand_landmarks in results.multi_hand_landmarks:
for landmark in hand_landmarks.landmark:
x, y = int(landmark.x * [Link][1]), int(landmark.y *
[Link][0])
[Link](frame, (x, y), 5, (255, 0, 0), -1)
video_placeholder.image(frame, channels='BGR')
video_capture.release()
# Run the function to start the live video stream with hand cursor control, click, and
scroll
capture_video()
38
39
SCREENSHOTS:
40
41
42
RESEARCH PAPER
43
44
45
46
47
48
CERTIFICATE
49