Summary: | The advancement of artificial intelligence (AI) has bought many advances to human society as a whole. By using daily activities and integrating the technology from the fruits of AI, we can manage to gain further access to knowledge we can only begin to imagine. In identifying human action recognition (HAR); processing photos and videos to discern whether a human is present, then mapping the subject classified, which lastly determines the action being carried out is the objective. To achieve this, various steps are taken and careful approach is required, with the extensive amount of research, numerous troubleshooting and experimentation is required. The AI architecture has to learn from dataset collected for it to discern the identification of action properly. HAR is achieved by using Python code using real-time webcam feed. Human pose detection library known as MediaPipe Pose Detection detects human anatomy from input through joints key-points. MediaPipe algorithm that extract features in x-y-z axis with visibility (four variables) and the extracted data is trained using CNN-LSTM based on the trained and tested algorithm classifier model. The output obtained produced an RGB-skeleton and an action label on the detected subject as standing, waving, walking and sitting, has yielded good results. © 2023 IEEE.
|