• Title/Summary/Keyword: learning through the image

Search Result 925, Processing Time 0.028 seconds

Using Skeleton Vector Information and RNN Learning Behavior Recognition Algorithm (스켈레톤 벡터 정보와 RNN 학습을 이용한 행동인식 알고리즘)

  • Kim, Mi-Kyung;Cha, Eui-Young
    • Journal of Broadcast Engineering
    • /
    • v.23 no.5
    • /
    • pp.598-605
    • /
    • 2018
  • Behavior awareness is a technology that recognizes human behavior through data and can be used in applications such as risk behavior through video surveillance systems. Conventional behavior recognition algorithms have been performed using the 2D camera image device or multi-mode sensor or multi-view or 3D equipment. When two-dimensional data was used, the recognition rate was low in the behavior recognition of the three-dimensional space, and other methods were difficult due to the complicated equipment configuration and the expensive additional equipment. In this paper, we propose a method of recognizing human behavior using only CCTV images without additional equipment using only RGB and depth information. First, the skeleton extraction algorithm is applied to extract points of joints and body parts. We apply the equations to transform the vector including the displacement vector and the relational vector, and study the continuous vector data through the RNN model. As a result of applying the learned model to various data sets and confirming the accuracy of the behavior recognition, the performance similar to that of the existing algorithm using the 3D information can be verified only by the 2D information.

Indoor positioning method using WiFi signal based on XGboost (XGboost 기반의 WiFi 신호를 이용한 실내 측위 기법)

  • Hwang, Chi-Gon;Yoon, Chang-Pyo;Kim, Dae-Jin
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.1
    • /
    • pp.70-75
    • /
    • 2022
  • Accurately measuring location is necessary to provide a variety of services. The data for indoor positioning measures the RSSI values from the WiFi device through an application of a smartphone. The measured data becomes the raw data of machine learning. The feature data is the measured RSSI value, and the label is the name of the space for the measured position. For this purpose, the machine learning technique is to study a technique that predicts the exact location only with the WiFi signal by applying an efficient technique to classification. Ensemble is a technique for obtaining more accurate predictions through various models than one model, including backing and boosting. Among them, Boosting is a technique for adjusting the weight of a model through a modeling result based on sampled data, and there are various algorithms. This study uses Xgboost among the above techniques and evaluates performance with other ensemble techniques.

A Method of Detecting Character Data through a Adaboost Learning Method (에이다부스트 학습을 이용한 문자 데이터 검출 방법)

  • Jang, Seok-Woo;Byun, Siwoo
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.18 no.7
    • /
    • pp.655-661
    • /
    • 2017
  • It is a very important task to extract character regions contained in various input color images, because characters can provide significant information representing the content of an image. In this paper, we propose a new method for extracting character regions from various input images using MCT features and an AdaBoost algorithm. Using geometric features, the method extracts actual character regions by filtering out non-character regions from among candidate regions. Experimental results show that the suggested algorithm accurately extracts character regions from input images. We expect the suggested algorithm will be useful in multimedia and image processing-related applications, such as store signboard detection and car license plate recognition.

Recognition System of Car License Plate using Fuzzy Neural Networks (퍼지 신경망을 이용한 자동차 번호판 인식 시스템)

  • Kim, Kwang-Baek;Cho, Jae-Hyun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.12 no.5
    • /
    • pp.313-319
    • /
    • 2007
  • In this paper, we propose a novel method to extract an area of car licence plate and codes of vehicle number from a photographed car image using features on vertical edges and a new Fuzzy neural network algorithm to recognize extracted codes. Prewitt mask is used in searching for vertical edges for detection of an area of vehicle number plate and feature information of vehicle number palate is used to eliminate image noises and extract the plate area and individual codes of vehicle number. Finally, for recognition of extracted codes, we use the proposed Fuzzy neural network algorithm, in which FCM is used as the learning structure between input and middle layers and Max_Min neural network is used as the learning structure within inhibition and output layers. Through a variety of experiments using real 150 images of vehicle, we showed that the proposed method is more efficient than others.

  • PDF

A Smart Refrigerator System based on Internet of Things (IoT 기반 스마트 냉장고 시스템)

  • Kim, Hanjin;Lee, Seunggi;Kim, Won-Tae
    • Journal of IKEEE
    • /
    • v.22 no.1
    • /
    • pp.156-161
    • /
    • 2018
  • Recently, as the population rapidly increases, food shortages and waste are emerging serious problem. In order to solve this problem, various countries and enterprises are trying research and product development such as a study of consumers' purchasing patterns of food and a development of smart refrigerator using IoT technology. However, the smart refrigerators which currently sold have high price issue and another waste due to malfunction and breakage by complicated configurations. In this paper, we proposed a low-cost smart refrigerator system based on IoT for solving the problem and efficient management of ingredients. The system recognizes and registers ingredients through QR code, image recognition, and speech recognition, and can provide various services of the smart refrigerator. In order to improve an accuracy of image recognition, we used a model using a deep learning algorithm and proved that it is possible to register ingredients accurately.

Emotion Recognition and Expression System of User using Multi-Modal Sensor Fusion Algorithm (다중 센서 융합 알고리즘을 이용한 사용자의 감정 인식 및 표현 시스템)

  • Yeom, Hong-Gi;Joo, Jong-Tae;Sim, Kwee-Bo
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.18 no.1
    • /
    • pp.20-26
    • /
    • 2008
  • As they have more and more intelligence robots or computers these days, so the interaction between intelligence robot(computer) - human is getting more and more important also the emotion recognition and expression are indispensable for interaction between intelligence robot(computer) - human. In this paper, firstly we extract emotional features at speech signal and facial image. Secondly we apply both BL(Bayesian Learning) and PCA(Principal Component Analysis), lastly we classify five emotions patterns(normal, happy, anger, surprise and sad) also, we experiment with decision fusion and feature fusion to enhance emotion recognition rate. The decision fusion method experiment on emotion recognition that result values of each recognition system apply Fuzzy membership function and the feature fusion method selects superior features through SFS(Sequential Forward Selection) method and superior features are applied to Neural Networks based on MLP(Multi Layer Perceptron) for classifying five emotions patterns. and recognized result apply to 2D facial shape for express emotion.

A Study on the Development of an Automatic Classification System for Life Safety Prevention Service Reporting Images through the Development of AI Learning Model and AI Model Serving Server (AI 학습모델 및 AI모델 서빙 서버 개발을 통한 생활안전 예방 서비스 신고 이미지 자동분류 시스템 개발에 대한 연구)

  • Young Sic Jeong;Yong-Woon Kim;Jeongil Yim
    • Journal of the Society of Disaster Information
    • /
    • v.19 no.2
    • /
    • pp.432-438
    • /
    • 2023
  • Purpose: The purpose of this study is to enable users to conveniently report risks by automatically classifying risk categories in real time using AI for images reported in the life safety prevention service app. Method: Through a system consisting of a life safety prevention service platform, life safety prevention service app, AI model serving server and sftp server interconnected through the Internet, the reported life safety images are automatically classified in real time, and the AI model used at this time An AI learning algorithm for generation was also developed. Result: Images can be automatically classified by AI processing in real time, making it easier for reporters to report matters related to life safety.Conclusion: The AI image automatic classification system presented in this paper automatically classifies reported images in real time with a classification accuracy of over 90%, enabling reporters to easily report images related to life safety. It is necessary to develop faster and more accurate AI models and improve system processing capacity.

Two-Stream Convolutional Neural Network for Video Action Recognition

  • Qiao, Han;Liu, Shuang;Xu, Qingzhen;Liu, Shouqiang;Yang, Wanggan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.10
    • /
    • pp.3668-3684
    • /
    • 2021
  • Video action recognition is widely used in video surveillance, behavior detection, human-computer interaction, medically assisted diagnosis and motion analysis. However, video action recognition can be disturbed by many factors, such as background, illumination and so on. Two-stream convolutional neural network uses the video spatial and temporal models to train separately, and performs fusion at the output end. The multi segment Two-Stream convolutional neural network model trains temporal and spatial information from the video to extract their feature and fuse them, then determine the category of video action. Google Xception model and the transfer learning is adopted in this paper, and the Xception model which trained on ImageNet is used as the initial weight. It greatly overcomes the problem of model underfitting caused by insufficient video behavior dataset, and it can effectively reduce the influence of various factors in the video. This way also greatly improves the accuracy and reduces the training time. What's more, to make up for the shortage of dataset, the kinetics400 dataset was used for pre-training, which greatly improved the accuracy of the model. In this applied research, through continuous efforts, the expected goal is basically achieved, and according to the study and research, the design of the original dual-flow model is improved.

SSD-based Fire Recognition and Notification System Linked with Power Line Communication (유도형 전력선 통신과 연동된 SSD 기반 화재인식 및 알림 시스템)

  • Yang, Seung-Ho;Sohn, Kyung-Rak;Jeong, Jae-Hwan;Kim, Hyun-Sik
    • Journal of IKEEE
    • /
    • v.23 no.3
    • /
    • pp.777-784
    • /
    • 2019
  • A pre-fire awareness and automatic notification system are required because it is possible to minimize the damage if the fire situation is precisely detected after a fire occurs in a place where people are unusual or in a mountainous area. In this study, we developed a RaspberryPi-based fire recognition system using Faster-recurrent convolutional neural network (F-RCNN) and single shot multibox detector (SSD) and demonstrated a fire alarm system that works with power line communication. Image recognition was performed with a pie camera of RaspberryPi, and the detected fire image was transmitted to a monitoring PC through an inductive power line communication network. The frame rate per second (fps) for each learning model was 0.05 fps for Faster-RCNN and 1.4 fps for SSD. SSD was 28 times faster than F-RCNN.

Anthropomorphic Animal Face Masking using Deep Convolutional Neural Network based Animal Face Classification

  • Khan, Rafiul Hasan;Lee, Youngsuk;Lee, Suk-Hwan;Kwon, Oh-Jun;Kwon, Ki-Ryong
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.5
    • /
    • pp.558-572
    • /
    • 2019
  • Anthropomorphism is the attribution of human traits, emotions, or intentions to non-human entities. Anthropomorphic animal face masking is the process by which human characteristics are plotted on the animal kind. In this research, we are proposing a compact system which finds the resemblance between a human face and animal face using Deep Convolutional Neural Network (DCNN) and later applies morphism between them. The whole process is done by firstly finding which animal most resembles the particular human face through a DCNN based animal face classification. And secondly, doing triangulation based morphing between the particular human face and the most resembled animal face. Compared to the conventional manual Control Point Selection system using an animator, we are proposing a Viola-Jones algorithm based Control Point selection process which detects facial features for the human face and takes the Control Points automatically. To initiate our approach, we built our own dataset containing ten thousand animal faces and a fourteen layer DCNN. The simulation results firstly demonstrate that the accuracy of our proposed DCNN architecture outperforms the related methods for the animal face classification. Secondly, the proposed morphing method manages to complete the morphing process with less deformation and without any human assistance.