Human-action-classification-with-still-images-using-CNN
In recent years, one of the most difficult problems in computer vision has been automatic human action and activity recognition. It is critical in a variety of artificial intelligence applications, including video surveillance, computer games, robotics, and human-computer interactions. Research into intelligent surveillance has resulted from the growing demand for safety and protection. Feature extraction, action representation, and classification are the three major steps of an action recognition system. As a result, each step is critical to achieving a high recognition rate. Raw data is fed into a feature vector by the feature extraction and action representation models. The classification result would be greatly influenced by the proper extraction and selection of features. In this research seven human activities are chosen from a custom dataset made by the author in order to create an optimal human action detection model. The activities are yawning, phoning, sitting, standing, walking, running and hand waving. In order to get a benchmark result for our model conventional methods such as Naive Bayes, Support Vector Machine (SVM), K-nearest Neighbor (KNN), Random Forest, Gradient Boosting and Decision Tree were used. After that deep learning models such as ResNet, AlexNet, VGG16 and DenseNet were used. Next I used ResNet50 for feature extraction and then classified the dataset using SVM. Finally, in order to outperform the other methods, a custom CNN model was introduced. The best result came from the custom CNN model. The custom Convolutional Neural Network (CNN) model gave 80\% accuracy.