Alexandria Engineering Journal (Dec 2024)
Enhancing fitness action recognition with ResNet-TransFit: Integrating IoT and deep learning techniques for real-time monitoring
Abstract
With the growing popularity of fitness, the demand for real-time action recognition and feedback is increasing. Current research faces challenges in handling complex actions, real-time processing, and system integration. To address these issues, we propose a novel fitness action recognition model that integrates ResNet, Transformer, and transfer learning techniques. Specifically, ResNet is used for image feature extraction, Transformer handles time-series data processing, and transfer learning accelerates the model’s adaptation to new data. We evaluated our model on the NTU RGB+D action recognition dataset, achieving 48.5 ms latency, 29.1 fps throughput, and 93.7% accuracy, significantly outperforming other models. Our model achieved an accuracy improvement of 5% over existing methods, demonstrating significant potential for real-time fitness monitoring. By incorporating IoT technology, our system enables real-time data processing and action recognition, making it ideal for smart fitness monitoring. Although the model has high complexity and memory usage, its efficiency and accuracy demonstrate its potential for widespread adoption. Future work will focus on optimizing the model structure and training methods to enhance applicability in resource-constrained environments, ensuring broader usability and efficiency in various real-world applications.