Unified Human Activity and Sign Recognition System
- Version
- Download 0
- File Size 369.22 KB
- File Count 1
- Create Date 17 December 2025
- Last Updated 17 December 2025
Unified Human Activity and Sign Recognition System
Divyashree Nayak1, Hamsini M2, Harshitha R3 , Nikita Shantappa Biradar4
1Student, Dept. of Information Science Engineering, AMC Engineering College, Karnataka, India
2Student, Dept. of Information Science Engineering, AMC Engineering College, Karnataka, India
3Student, Dept. of Information Science Engineering, AMC Engineering College, Karnataka, India
4Student, Dept. of Information Science Engineering, AMC Engineering College, Karnataka, India
Abstract
Human Activity Recognition (HAR) and Sign Language Recognition (SLR) play a crucial role in advancing intelligent human–computer interaction systems. However, most existing approaches address these tasks independently, which limits their real-world usability and scalability. To overcome this limitation, this paper presents a unified deep learning–based framework that integrates Human Activity Recognition (HAR), Sign Language Recognition (SLR), and American Sign Language (ASL) gesture classification into a single, cohesive system.The proposed framework utilizes video-based datasets for HAR and SLR stored in .npy format, along with an image-based ASL dataset in .jpg format. For effective feature extraction, 3D Convolutional Neural Networks (3D CNNs) are employed to capture spatio-temporal patterns from video data, while MobileNetV2 is used for lightweight and efficient feature extraction from image-based gestures. The extracted features are further processed using Long Short-Term Memory (LSTM) networks to model temporal dependencies and improve recognition accuracy.To enhance societal applicability, the system provides real-time feedback through a text-to-speech module and generates automated alerts via Telegram for detected activities or recognized signs. Experimental evaluations show high accuracy, stable convergence, and efficient performance across all tasks, demonstrating the robustness of the unified architecture. This work contributes to the development of an intelligent multimodal recognition system that enhances accessibility and bridges communication gaps between humans and machines.
Key Words: Human Activity Recognition (HAR), Sign Language Recognition (SLR), American Sign Language (ASL), Deep Learning, 3D CNN, MobileNetV2, LSTM, Text-to-Speech, Telegram Alert.
Download