• Title/Summary/Keyword: Pose classification

Search Result 80, Processing Time 0.023 seconds

A Study on Improvement of the Human Posture Estimation Method for Performing Robots (공연로봇을 위한 인간자세 추정방법 개선에 관한 연구)

  • Park, Cheonyu;Park, Jaehun;Han, Jeakweon
    • Journal of Broadcast Engineering
    • /
    • v.25 no.5
    • /
    • pp.750-757
    • /
    • 2020
  • One of the basic tasks for robots to interact with humans is to quickly and accurately grasp human behavior. Therefore, it is necessary to increase the accuracy of human pose recognition when the robot is estimating the human pose and to recognize it as quickly as possible. However, when the human pose is estimated using deep learning, which is a representative method of artificial intelligence technology, recognition accuracy and speed are not satisfied at the same time. Therefore, it is common to select one of a top-down method that has high inference accuracy or a bottom-up method that has high processing speed. In this paper, we propose two methods that complement the disadvantages while including both the advantages of the two methods mentioned above. The first is to perform parallel inference on the server using multi GPU, and the second is to mix bottom-up and One-class Classification. As a result of the experiment, both of the methods presented in this paper showed improvement in speed. If these two methods are applied to the entertainment robot, it is expected that a highly reliable interaction with the audience can be performed.

Parking Lot Occupancy Detection using Deep Learning and Fisheye Camera for AIoT System

  • To Xuan Dung;Seongwon Cho
    • Smart Media Journal
    • /
    • v.13 no.1
    • /
    • pp.24-35
    • /
    • 2024
  • The combination of Artificial Intelligence and the Internet of Things (AIoT) has gained significant popularity. Deep neural networks (DNNs) have demonstrated remarkable success in various applications. However, deploying complex AI models on embedded boards can pose challenges due to computational limitations and model complexity. This paper presents an AIoT-based system for smart parking lots using edge devices. Our approach involves developing a detection model and a decision tree for occupancy status classification. Specifically, we utilize YOLOv5 for car license plate (LP) detection by verifying the position of the license plate within the parking space.

A Novel Algorithm for Fault Classification in Transmission Lines Using a Combined Adaptive Network and Fuzzy Inference System

  • Yeo, Sang-Min;Kim, Chun-Hwan
    • KIEE International Transactions on Power Engineering
    • /
    • v.3A no.4
    • /
    • pp.191-197
    • /
    • 2003
  • Accurate detection and classification of faults on transmission lines is vitally important. In this respect, many different types of faults occur, such as inter alia low impedance faults (LIF) and high impedance faults (HIF). The latter in particular pose difficulties for the commonly employed conventional overcurrent and distance relays, and if undetected, can cause damage to expensive equipment, threaten life and cause fire hazards. Although HIFs are far less common than LIFs, it is imperative that any protection device should be able to satisfactorily deal with both HIFs and LIFs. Because of the randomness and asymmetric characteristics of HIFs, their modeling is difficult and numerous papers relating to various HIF models have been published. In this paper, the model of HIFs in transmission lines is accomplished using the characteristics of a ZnO arrester, which is then implemented within the overall transmission system model based on the electromagnetic transients program (EMTP). This paper proposes an algorithm for fault detection and classification for both LIFs and HIFs using Adaptive Network-based Fuzzy Inference System (ANFIS). The inputs into ANFIS are current signals only based on Root-Mean-Square (RMS) values of 3-phase currents and zero sequence current. The performance of the proposed algorithm is tested on a typical 154 kV Korean transmission line system under various fault conditions. Test results demonstrate that the ANFIS can detect and classify faults including LIFs and HIFs accurately within half a cycle.

Multi-Frame Face Classification with Decision-Level Fusion based on Photon-Counting Linear Discriminant Analysis

  • Yeom, Seokwon
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.14 no.4
    • /
    • pp.332-339
    • /
    • 2014
  • Face classification has wide applications in security and surveillance. However, this technique presents various challenges caused by pose, illumination, and expression changes. Face recognition with long-distance images involves additional challenges, owing to focusing problems and motion blurring. Multiple frames under varying spatial or temporal settings can acquire additional information, which can be used to achieve improved classification performance. This study investigates the effectiveness of multi-frame decision-level fusion with photon-counting linear discriminant analysis. Multiple frames generate multiple scores for each class. The fusion process comprises three stages: score normalization, score validation, and score combination. Candidate scores are selected during the score validation process, after the scores are normalized. The score validation process removes bad scores that can degrade the final output. The selected candidate scores are combined using one of the following fusion rules: maximum, averaging, and majority voting. Degraded facial images are employed to demonstrate the robustness of multi-frame decision-level fusion in harsh environments. Out-of-focus and motion blurring point-spread functions are applied to the test images, to simulate long-distance acquisition. Experimental results with three facial data sets indicate the efficiency of the proposed decision-level fusion scheme.

Spatiotemporal Patched Frames for Human Abnormal Behavior Classification in Low-Light Environment (저조도 환경 감시 영상에서 시공간 패치 프레임을 이용한 이상행동 분류)

  • Widia A. Samosir;Seong G. Kong
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.11a
    • /
    • pp.634-636
    • /
    • 2023
  • Surveillance systems play a pivotal role in ensuring the safety and security of various environments, including public spaces, critical infrastructure, and private properties. However, detecting abnormal human behavior in lowlight conditions is a critical yet challenging task due to the inherent limitations of visual data acquisition in such scenarios. This paper introduces a spatiotemporal framework designed to address the unique challenges posed by low-light environments, enhancing the accuracy and efficiency of human abnormality detection in surveillance camera systems. We proposed the pre-processing using lightweight exposure correction, patched frames pose estimation, and optical flow to extract the human behavior flow through t-seconds of frames. After that, we train the estimated-action-flow into autoencoder for abnormal behavior classification to get normal loss as metrics decision for normal/abnormal behavior.

A Study on Hand-signal Recognition System in 3-dimensional Space (3차원 공간상의 수신호 인식 시스템에 대한 연구)

  • 장효영;김대진;김정배;변증남
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.41 no.3
    • /
    • pp.103-114
    • /
    • 2004
  • This paper deals with a system that is capable of recognizing hand-signals in 3-dimensional space. The system uses 2 color cameras as input devices. Vision-based gesture recognition system is known to be user-friendly because of its contact-free characteristic. But as with other applications using a camera as an input device, there are difficulties under complex background and varying illumination. In order to detect hand region robustly from a input image under various conditions without any special gloves or markers, the paper uses previous position information and adaptive hand color model. The paper defines a hand-signal as a combination of two basic elements such as 'hand pose' and 'hand trajectory'. As an extensive classification method for hand pose, the paper proposes 2-stage classification method by using 'small group concept'. Also, the paper suggests a complementary feature selection method from images from two color cameras. We verified our method with a hand-signal application to our driving simulator.

Multimodal Biometrics Recognition from Facial Video with Missing Modalities Using Deep Learning

  • Maity, Sayan;Abdel-Mottaleb, Mohamed;Asfour, Shihab S.
    • Journal of Information Processing Systems
    • /
    • v.16 no.1
    • /
    • pp.6-29
    • /
    • 2020
  • Biometrics identification using multiple modalities has attracted the attention of many researchers as it produces more robust and trustworthy results than single modality biometrics. In this paper, we present a novel multimodal recognition system that trains a deep learning network to automatically learn features after extracting multiple biometric modalities from a single data source, i.e., facial video clips. Utilizing different modalities, i.e., left ear, left profile face, frontal face, right profile face, and right ear, present in the facial video clips, we train supervised denoising auto-encoders to automatically extract robust and non-redundant features. The automatically learned features are then used to train modality specific sparse classifiers to perform the multimodal recognition. Moreover, the proposed technique has proven robust when some of the above modalities were missing during the testing. The proposed system has three main components that are responsible for detection, which consists of modality specific detectors to automatically detect images of different modalities present in facial video clips; feature selection, which uses supervised denoising sparse auto-encoders network to capture discriminative representations that are robust to the illumination and pose variations; and classification, which consists of a set of modality specific sparse representation classifiers for unimodal recognition, followed by score level fusion of the recognition results of the available modalities. Experiments conducted on the constrained facial video dataset (WVU) and the unconstrained facial video dataset (HONDA/UCSD), resulted in a 99.17% and 97.14% Rank-1 recognition rates, respectively. The multimodal recognition accuracy demonstrates the superiority and robustness of the proposed approach irrespective of the illumination, non-planar movement, and pose variations present in the video clips even in the situation of missing modalities.

Training Network Design Based on Convolution Neural Network for Object Classification in few class problem (소 부류 객체 분류를 위한 CNN기반 학습망 설계)

  • Lim, Su-chang;Kim, Seung-Hyun;Kim, Yeon-Ho;Kim, Do-yeon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.21 no.1
    • /
    • pp.144-150
    • /
    • 2017
  • Recently, deep learning is used for intelligent processing and accuracy improvement of data. It is formed calculation model composed of multi data processing layer that train the data representation through an abstraction of the various levels. A category of deep learning, convolution neural network is utilized in various research fields, which are human pose estimation, face recognition, image classification, speech recognition. When using the deep layer and lots of class, CNN that show a good performance on image classification obtain higher classification rate but occur the overfitting problem, when using a few data. So, we design the training network based on convolution neural network and trained our image data set for object classification in few class problem. The experiment show the higher classification rate of 7.06% in average than the previous networks designed to classify the object in 1000 class problem.

Human Skeleton Keypoints based Fall Detection using GRU (PoseNet과 GRU를 이용한 Skeleton Keypoints 기반 낙상 감지)

  • Kang, Yoon Kyu;Kang, Hee Yong;Weon, Dal Soo
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.22 no.2
    • /
    • pp.127-133
    • /
    • 2021
  • A recent study of people physically falling focused on analyzing the motions of the falls using a recurrent neural network (RNN) and a deep learning approach to get good results from detecting 2D human poses from a single color image. In this paper, we investigate a detection method for estimating the position of the head and shoulder keypoints and the acceleration of positional change using the skeletal keypoints information extracted using PoseNet from an image obtained with a low-cost 2D RGB camera, increasing the accuracy of judgments about the falls. In particular, we propose a fall detection method based on the characteristics of post-fall posture in the fall motion-analysis method. A public data set was used to extract human skeletal features, and as a result of an experiment to find a feature extraction method that can achieve high classification accuracy, the proposed method showed a 99.8% success rate in detecting falls more effectively than a conventional, primitive skeletal data-use method.

Fall Detection Based on Human Skeleton Keypoints Using GRU

  • Kang, Yoon-Kyu;Kang, Hee-Yong;Weon, Dal-Soo
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.12 no.4
    • /
    • pp.83-92
    • /
    • 2020
  • A recent study to determine the fall is focused on analyzing fall motions using a recurrent neural network (RNN), and uses a deep learning approach to get good results for detecting human poses in 2D from a mono color image. In this paper, we investigated the improved detection method to estimate the position of the head and shoulder key points and the acceleration of position change using the skeletal key points information extracted using PoseNet from the image obtained from the 2D RGB low-cost camera, and to increase the accuracy of the fall judgment. In particular, we propose a fall detection method based on the characteristics of post-fall posture in the fall motion analysis method and on the velocity of human body skeleton key points change as well as the ratio change of body bounding box's width and height. The public data set was used to extract human skeletal features and to train deep learning, GRU, and as a result of an experiment to find a feature extraction method that can achieve high classification accuracy, the proposed method showed a 99.8% success rate in detecting falls more effectively than the conventional primitive skeletal data use method.