• Title/Summary/Keyword: Action recognition

Search Result 405, Processing Time 0.025 seconds

Two person Interaction Recognition Based on Effective Hybrid Learning

  • Ahmed, Minhaz Uddin;Kim, Yeong Hyeon;Kim, Jin Woo;Bashar, Md Rezaul;Rhee, Phill Kyu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.2
    • /
    • pp.751-770
    • /
    • 2019
  • Action recognition is an essential task in computer vision due to the variety of prospective applications, such as security surveillance, machine learning, and human-computer interaction. The availability of more video data than ever before and the lofty performance of deep convolutional neural networks also make it essential for action recognition in video. Unfortunately, limited crafted video features and the scarcity of benchmark datasets make it challenging to address the multi-person action recognition task in video data. In this work, we propose a deep convolutional neural network-based Effective Hybrid Learning (EHL) framework for two-person interaction classification in video data. Our approach exploits a pre-trained network model (the VGG16 from the University of Oxford Visual Geometry Group) and extends the Faster R-CNN (region-based convolutional neural network a state-of-the-art detector for image classification). We broaden a semi-supervised learning method combined with an active learning method to improve overall performance. Numerous types of two-person interactions exist in the real world, which makes this a challenging task. In our experiment, we consider a limited number of actions, such as hugging, fighting, linking arms, talking, and kidnapping in two environment such simple and complex. We show that our trained model with an active semi-supervised learning architecture gradually improves the performance. In a simple environment using an Intelligent Technology Laboratory (ITLab) dataset from Inha University, performance increased to 95.6% accuracy, and in a complex environment, performance reached 81% accuracy. Our method reduces data-labeling time, compared to supervised learning methods, for the ITLab dataset. We also conduct extensive experiment on Human Action Recognition benchmarks such as UT-Interaction dataset, HMDB51 dataset and obtain better performance than state-of-the-art approaches.

Analysis of Table Tennis Swing using Action Recognition (동작인식을 이용한 탁구 스윙 분석)

  • Heo, Geon;Ha, Jong-Eun
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.21 no.1
    • /
    • pp.40-45
    • /
    • 2015
  • In this paper, we present an algorithm for the analysis of poses while playing table-tennis using action recognition. We use Kinect as the 3D sensor and 3D skeleton data provided by Kinect for further processing. We adopt a spherical coordinate system and feature selected using k-means clustering. We automatically detect the starting and ending frame and discriminate the action of table-tennis into two groups of forehand and backhand swing. Each swing is modeled using HMM(Hidden Markov Model) and we used a dataset composed of 200 sequences from two players. We can discriminate two types of table tennis swing in real-time. Also, it can provide analysis according to similarities found in good poses.

Decomposed "Spatial and Temporal" Convolution for Human Action Recognition in Videos

  • Sediqi, Khwaja Monib;Lee, Hyo Jong
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2019.05a
    • /
    • pp.455-457
    • /
    • 2019
  • In this paper we study the effect of decomposed spatiotemporal convolutions for action recognition in videos. Our motivation emerges from the empirical observation that spatial convolution applied on solo frames of the video provide good performance in action recognition. In this research we empirically show the accuracy of factorized convolution on individual frames of video for action classification. We take 3D ResNet-18 as base line model for our experiment, factorize its 3D convolution to 2D (Spatial) and 1D (Temporal) convolution. We train the model from scratch using Kinetics video dataset. We then fine-tune the model on UCF-101 dataset and evaluate the performance. Our results show good accuracy similar to that of the state of the art algorithms on Kinetics and UCF-101 datasets.

Human Action Recognition Using Deep Data: A Fine-Grained Study

  • Rao, D. Surendra;Potturu, Sudharsana Rao;Bhagyaraju, V
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.6
    • /
    • pp.97-108
    • /
    • 2022
  • The video-assisted human action recognition [1] field is one of the most active ones in computer vision research. Since the depth data [2] obtained by Kinect cameras has more benefits than traditional RGB data, research on human action detection has recently increased because of the Kinect camera. We conducted a systematic study of strategies for recognizing human activity based on deep data in this article. All methods are grouped into deep map tactics and skeleton tactics. A comparison of some of the more traditional strategies is also covered. We then examined the specifics of different depth behavior databases and provided a straightforward distinction between them. We address the advantages and disadvantages of depth and skeleton-based techniques in this discussion.

Optimization of Action Recognition based on Slowfast Deep Learning Model using RGB Video Data (RGB 비디오 데이터를 이용한 Slowfast 모델 기반 이상 행동 인식 최적화)

  • Jeong, Jae-Hyeok;Kim, Min-Suk
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.8
    • /
    • pp.1049-1058
    • /
    • 2022
  • HAR(Human Action Recognition) such as anomaly and object detection has become a trend in research field(s) that focus on utilizing Artificial Intelligence (AI) methods to analyze patterns of human action in crime-ridden area(s), media services, and industrial facilities. Especially, in real-time system(s) using video streaming data, HAR has become a more important AI-based research field in application development and many different research fields using HAR have currently been developed and improved. In this paper, we propose and analyze a deep-learning-based HAR that provides more efficient scheme(s) using an intelligent AI models, such system can be applied to media services using RGB video streaming data usage without feature extraction pre-processing. For the method, we adopt Slowfast based on the Deep Neural Network(DNN) model under an open dataset(HMDB-51 or UCF101) for improvement in prediction accuracy.

Consideration of the Early Action in the GHG Emission Reduction (국내 온실가스 감축의 조기행동 인정 방안)

  • Song, Bo-Yun;Park, Su-Mi;Chung, Jin-Do
    • Journal of Korean Society for Atmospheric Environment
    • /
    • v.27 no.2
    • /
    • pp.209-213
    • /
    • 2011
  • Enforcement Decree of the Framework Act on Low Carbon, Green Growth for achieving the country's GHG emission reduction goal of 30% was in effect. The remarkable content of the Act is the managements of targets for GHG reduction. So, the entities that have reduced voluntarily have much interest in the recognition of 'early action'. The recognition of early action is necessary to induce the fair competence of business entities and promote the voluntary GHG reduction. The definite and concrete guidance should be prepared. The important principles for this are the environmental integrity and the additionality. Based on this, the early action activities must be restricted to the voluntary, real, permanent, quantifiable, verifiable reduction. In the early action recognition, its credit should be allocated additionally set aside from the GHG target allocation in the national total allowance. Through this, the reward for the early reduction should be realized on market mechanism. The effective period to award the early action should be addressed. This can be the period after the enactment of framework on GHG reduction in effect and before the beginning year of GHG target control. It should be set with flexibility through the collection and consultation of the sector's opinions. The appropriate allowance reserve of early action was estimated as approximately 1~1.5% by using the data from the 'Pilot GHG Emission Trading Program' operated by Ministry of Environment. Also, the concrete and detail guidance to construct the necessary infra which is used to register the related information of early action activities should be prepared.

Probabilistic Graph Based Object Category Recognition Using the Context of Object-Action Interaction (물체-행동 컨텍스트를 이용하는 확률 그래프 기반 물체 범주 인식)

  • Yoon, Sung-baek;Bae, Se-ho;Park, Han-je;Yi, June-ho
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.40 no.11
    • /
    • pp.2284-2290
    • /
    • 2015
  • The use of human actions as context for object class recognition is quite effective in enhancing the recognition performance despite the large variation in the appearance of objects. We propose an efficient method that integrates human action information into object class recognition using a Bayesian appraoch based on a simple probabilistic graph model. The experiment shows that by using human actions ac context information we can improve the performance of the object calss recognition from 8% to 28%.

Snake Robot Motion Scheme Using Image and Voice (감각 정보를 이용한 뱀 로봇의 행동구현)

  • 강준영;김성주;조현찬;전홍태
    • Proceedings of the IEEK Conference
    • /
    • 2002.06c
    • /
    • pp.127-130
    • /
    • 2002
  • Human's brain action can divide by recognition and intelligence. recognition is sensing voice, image and smell and Intelligence is logical judgment, inference, decision. To this concept, Define function of cerebral cortex, and apply the result. Current expert system is lack, that reasoning by cerebral cortex and thalamus, hoppocampal and so on. In this paper, With human's brain action, wish to embody human's action artificially Embody brain mechanism using Modular Neural Network, Applied this result to snake robot.

  • PDF

Intelligent Activity Recognition based on Improved Convolutional Neural Network

  • Park, Jin-Ho;Lee, Eung-Joo
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.6
    • /
    • pp.807-818
    • /
    • 2022
  • In order to further improve the accuracy and time efficiency of behavior recognition in intelligent monitoring scenarios, a human behavior recognition algorithm based on YOLO combined with LSTM and CNN is proposed. Using the real-time nature of YOLO target detection, firstly, the specific behavior in the surveillance video is detected in real time, and the depth feature extraction is performed after obtaining the target size, location and other information; Then, remove noise data from irrelevant areas in the image; Finally, combined with LSTM modeling and processing time series, the final behavior discrimination is made for the behavior action sequence in the surveillance video. Experiments in the MSR and KTH datasets show that the average recognition rate of each behavior reaches 98.42% and 96.6%, and the average recognition speed reaches 210ms and 220ms. The method in this paper has a good effect on the intelligence behavior recognition.

Human Action Recognition via Depth Maps Body Parts of Action

  • Farooq, Adnan;Farooq, Faisal;Le, Anh Vu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.5
    • /
    • pp.2327-2347
    • /
    • 2018
  • Human actions can be recognized from depth sequences. In the proposed algorithm, we initially construct depth, motion maps (DMM) by projecting each depth frame onto three orthogonal Cartesian planes and add the motion energy for each view. The body part of the action (BPoA) is calculated by using bounding box with an optimal window size based on maximum spatial and temporal changes for each DMM. Furthermore, feature vector is constructed by using BPoA for each human action view. In this paper, we employed an ensemble based learning approach called Rotation Forest to recognize different actions Experimental results show that proposed method has significantly outperforms the state-of-the-art methods on Microsoft Research (MSR) Action 3D and MSR DailyActivity3D dataset.