Page 358 - Kaleidoscope Academic Conference Proceedings 2024
P. 358

2024 ITU Kaleidoscope Academic Conference




               1.  The user interface is designed to be intuitive,
                  engaging, and visually appealing, ensuring ease of
                  navigation and accessibility for all users.
               2.  Features clear and concise information on
                  sustainable practices, making it easy for users to
                  adopt and incorporate these practices into their
                  daily routines.
               3.  The application leverages data analytics to provide
                  personalized recommendations and insights based
                  on user behavior and preferences.
               4.  Utilizes data-driven approaches to track and
                  measure the user's environmental and social
                  impact, providing feedback and rewards for
                  positive contributions.
               5.  Offers a wealth of educational resources such as
                  articles, videos, and interactive modules to raise     Figure 1– Landmarks
                  awareness and educate users about sustainable
                  yoga practices.

                             2.  DATASETS
                                                                    5.  OBJECTIVE AND METHODOLOGY
           Yoga-82 is a new Dataset for Fine-grained Classification of   Objective: Development of DNN based detection model
           Human Poses. It is one of the largest data set which is used   using transfer learning technique to detect the yoga pose,
           to train the yoga classification techniques. The data set   capture landmarks on real-time image or video to detect the
           contains a three-level hierarchy including body positions,   posture and integrate it in application to alert user for
           variations in body positions, and the actual pose names. We   correctness. The application shall also have encyclopedia on
           present the classification accuracy of the state-of-the-art   yogasana’s with relevant information.
           convolutional neural network architectures on Yoga-82. We
           also present several hierarchical variants of Dense Net in   Methodology: For examination of different yoga poses,
           order to utilize the hierarchical labels. On similar lines, we   pre-trained model MoveNet is used, which detects 17
           have also created custom data set to train the model for   landmarks on each images and creates csv file and prepared
           most common poses used by the public to demonstrate the   train and test data. These manually created csv files are fed
           capability of the model.
                                                              to train the model for classification of poses. We have
                                                              developed react.js front-end application and used
                               3.  LOGIC
                                                              tensorflow.js for importing the model. At first it shall detect
                                                              the pose and capture 17 landmarks from input image or
           Human pose estimation is a Computer Vision technique   video for comparison with master landmarks. If current
           used to predict a person’s body parts or joints position. This   posture reaches the accuracy of threefold above 70%
           can be done by defining the human body joints like wrist,   probability then, the colour turns to the green and also it
           shoulder, knees, eyes, ears, ankles, arms, also called key   counts the time for holding the posture correctly and
           points in images and videos. Then, when a picture or video   notifies us by playing sound.
           comes in as input to the pose estimator model, it identifies
           the coordinates of those detected body parts as output and a   6.  LITERATURE SURVEY
           confidence score indicating continuity of the estimations.
           The MoveNet model is based on 3D estimation. The   Due to the increase of stress in the modern lifestyle, yoga
           operation takes place in a phase-wise  manner like; first, the   has become popular throughout the world. Although
           RGB image is fed to the convolutional network as input,   considered to be a great way to bring physical, mental, and
           then the pose model is applied to detect the poses, key   spiritual harmony, it can come with its own set of problems,
           points, pose confidence score and key point confidence   if not done properly. This gave rise to the need for Yoga
           score from the model outputs.
                                                              Pose Detection and Classification, to omit the necessity of a
                                                              human instructor. But, for an AI agent to behave like a
                            4.  LANDMARKS
                                                              human, it is necessary to have a strong model to detect and
                                                              classify various yoga poses. OpenPose was the pioneer
           The landmarks of a human body used in this project are   project in the field of Landmark detection using Image
           nose, ears, eyes, shoulders, elbows, wrists, hips, knees, and   Processing. Using 18 landmarks detected on the human
           ankles. This gives us 17 landmarks, which will be used to   body, it was able to achieve 78% accuracy in detecting the
           detect various poses by the model.                 yoga poses. PifPaf uses a Part Intensity Field (PIF) and a
                                                              Part Association Field (PAF) for body part limitation, and
                                                              relationship of body parts to shape full human stances,
                                                              respectively. This technique is dependent on the base up




                                                          – 314 –
   353   354   355   356   357   358   359   360   361   362   363