• Title/Summary/Keyword: object detection and classification

Search Result 296, Processing Time 0.035 seconds

Estimation of User Activity States for Context-Aware Computing in Mobile Devices (모바일 디바이스에서 상황인식 컴퓨팅을 위한 사용자 활동 상태 추정)

  • Baek Jonghun;Yun Byoung-Ju
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.1 s.307
    • /
    • pp.67-74
    • /
    • 2006
  • Contort-aware computing technology is one of the key technology of ubiquitous computing in the mobile device environment. Context recognition computing enables computer applications that automatically respond to user's everyday activity to be realized. In this paper, We use accelerometer could sense activity states of the object and apply to mobile devices. This method for estimating human motion states utilizes various statistics of accelerometer data, such as mean, standard variation, and skewness, as features for classification, and is expected to be more effective than other existing methods that rely on only a few simple statistics. Classification algorithm uses simple decision tree instead of existing neural network by considering mobile devices with limited resources. A series of experiments for testing the effectiveness of the our context detection system for mobile applications and ubiquitous computing has been performed, and its result is presented.

An Efficient Face Region Detection for Content-based Video Summarization (내용기반 비디오 요약을 위한 효율적인 얼굴 객체 검출)

  • Kim Jong-Sung;Lee Sun-Ta;Baek Joong-Hwan
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.30 no.7C
    • /
    • pp.675-686
    • /
    • 2005
  • In this paper, we propose an efficient face region detection technique for the content-based video summarization. To segment video, shot changes are detected from a video sequence and key frames are selected from the shots. We select one frame that has the least difference between neighboring frames in each shot. The proposed face detection algorithm detects face region from selected key frames. And then, we provide user with summarized frames included face region that has an important meaning in dramas or movies. Using Bayes classification rule and statistical characteristic of the skin pixels, face regions are detected in the frames. After skin detection, we adopt the projection method to segment an image(frame) into face region and non-face region. The segmented regions are candidates of the face object and they include many false detected regions. So, we design a classifier to minimize false lesion using CART. From SGLD matrices, we extract the textual feature values such as Inertial, Inverse Difference, and Correlation. As a result of our experiment, proposed face detection algorithm shows a good performance for the key frames with a complex and variant background. And our system provides key frames included the face region for user as video summarized information.

De-interlacing and Block Code Generation For Outsole Model Recognition In Moving Picture (동영상에서 신발 밑창 모델 인식을 위한 인터레이스 제거 및 블록 코드 생성 기법)

  • Kim Cheol-Ki
    • Journal of Intelligence and Information Systems
    • /
    • v.12 no.1
    • /
    • pp.33-41
    • /
    • 2006
  • This paper presents a method that automatically recognizes products into model type, which it flows with the conveyor belt. The specific interlaced image are occurred by moving image when we use the NTSC based camera. It is impossible to process interlaced images, so a suitable post-processing is required. For the purpose of this processing, after it remove interlaced images using de-interlacing method, it leads rectangle region of object by thresholding. And then, after rectangle region is separated into several blocks through edge detection, we calculate pixel numbers per each block, re-classify using its average, and classify products into model type. Through experiments, we know that the proposed method represent high classification ratio.

  • PDF

Application of Deep Learning Algorithm for Detecting Construction Workers Wearing Safety Helmet Using Computer Vision (건설현장 근로자의 안전모 착용 여부 검출을 위한 컴퓨터 비전 기반 딥러닝 알고리즘의 적용)

  • Kim, Myung Ho;Shin, Sung Woo;Suh, Yong Yoon
    • Journal of the Korean Society of Safety
    • /
    • v.34 no.6
    • /
    • pp.29-37
    • /
    • 2019
  • Since construction sites are exposed to outdoor environments, working conditions are significantly dangerous. Thus, wearing of the personal protective equipments such as safety helmet is very important for worker safety. However, construction workers are often wearing-off the helmet as inconvenient and uncomportable. As a result, a small mistake may lead to serious accident. For this, checking of wearing safety helmet is important task to safety managers in field. However, due to the limited time and manpower, the checking can not be executed for every individual worker spread over a large construction site. Therefore, if an automatic checking system is provided, field safety management should be performed more effectively and efficiently. In this study, applicability of deep learning based computer vision technology is investigated for automatic checking of wearing safety helmet in construction sites. Faster R-CNN deep learning algorithm for object detection and classification is employed to develop the automatic checking model. Digital camera images captured in real construction site are used to validate the proposed model. Based on the results, it is concluded that the proposed model may effectively be used for automatic checking of wearing safety helmet in construction site.

The Importance of Manpower in Major Education as an Example of Artificial Intelligence Development in Construction (건설 인공지능 개발사례로 보는 전공교육 인력의 중요성)

  • Heo, Seokjae;Lee, Sanghyun;Lee, Seungwon;Kim, Myunghun;Chung, Lan
    • Proceedings of the Korean Institute of Building Construction Conference
    • /
    • 2021.11a
    • /
    • pp.223-224
    • /
    • 2021
  • The process before the model learning stage in AI R&D can be subdivided into data collection/cleansing-data purification-data labeling. After that, according to the purpose of development, it goes through a stage of verifying the model by performing learning by using the algorithm of the artificial intelligence model. Several studies describe an important part of AI research as the learning stage, and try to increase the accuracy by changing the structure and layer of the AI model. However, if the refinement and labeling process of the learning data is tailored only to the model format and is not made for the purpose of development, the desired AI model cannot be obtained. The latest research reveals that most AI research failures are the failure of the learning data rather than the structure of the AI model. analyzed.

  • PDF

Weakly-supervised Semantic Segmentation using Exclusive Multi-Classifier Deep Learning Model (독점 멀티 분류기의 심층 학습 모델을 사용한 약지도 시맨틱 분할)

  • Choi, Hyeon-Joon;Kang, Dong-Joong
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.6
    • /
    • pp.227-233
    • /
    • 2019
  • Recently, along with the recent development of deep learning technique, neural networks are achieving success in computer vision filed. Convolutional neural network have shown outstanding performance in not only for a simple image classification task, but also for tasks with high difficulty such as object segmentation and detection. However many such deep learning models are based on supervised-learning, which requires more annotation labels than image-level label. Especially image semantic segmentation model requires pixel-level annotations for training, which is very. To solve these problems, this paper proposes a weakly-supervised semantic segmentation method which requires only image level label to train network. Existing weakly-supervised learning methods have limitations in detecting only specific area of object. In this paper, on the other hand, we use multi-classifier deep learning architecture so that our model recognizes more different parts of objects. The proposed method is evaluated using VOC 2012 validation dataset.

A Scheme on Object Tracking Techniques in Multiple CCTV IoT Environments (다중 CCTV 사물인터넷 환경에서의 객체 추적 기법)

  • Hong, Ji-Hoon;Lee, Keun-Ho
    • Journal of Internet of Things and Convergence
    • /
    • v.5 no.1
    • /
    • pp.7-11
    • /
    • 2019
  • This study suggests a methodology to track crime suspects or anomalies through CCTV in order to expand the scope of CCTV use as the number of CCTV installations continues to increase nationwide in recent years. For the abnormal behavior classification, we use the existing studies to find out suspected criminals or abnormal actors, use CNN to track objects, and connect the surrounding CCTVs to each other to predict the movement path of objectified objects CCTVs in the vicinity of the path were used to share objects' sample data to track objects and to track objects. Through this research, we will keep track of criminals who can not be traced, contribute to the national security, and continue to study them so that more diverse technologies can be applied to CCTV.

Active Vision from Image-Text Multimodal System Learning (능동 시각을 이용한 이미지-텍스트 다중 모달 체계 학습)

  • Kim, Jin-Hwa;Zhang, Byoung-Tak
    • Journal of KIISE
    • /
    • v.43 no.7
    • /
    • pp.795-800
    • /
    • 2016
  • In image classification, recent CNNs compete with human performance. However, there are limitations in more general recognition. Herein we deal with indoor images that contain too much information to be directly processed and require information reduction before recognition. To reduce the amount of data processing, typically variational inference or variational Bayesian methods are suggested for object detection. However, these methods suffer from the difficulty of marginalizing over the given space. In this study, we propose an image-text integrated recognition system using active vision based on Spatial Transformer Networks. The system attempts to efficiently sample a partial region of a given image for a given language information. Our experimental results demonstrate a significant improvement over traditional approaches. We also discuss the results of qualitative analysis of sampled images, model characteristics, and its limitations.

Frontal Face Video Analysis for Detecting Fatigue States

  • Cha, Simyeong;Ha, Jongwoo;Yoon, Soungwoong;Ahn, Chang-Won
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.6
    • /
    • pp.43-52
    • /
    • 2022
  • We can sense somebody's feeling fatigue, which means that fatigue can be detected through sensing human biometric signals. Numerous researches for assessing fatigue are mostly focused on diagnosing the edge of disease-level fatigue. In this study, we adapt quantitative analysis approaches for estimating qualitative data, and propose video analysis models for measuring fatigue state. Proposed three deep-learning based classification models selectively include stages of video analysis: object detection, feature extraction and time-series frame analysis algorithms to evaluate each stage's effect toward dividing the state of fatigue. Using frontal face videos collected from various fatigue situations, our CNN model shows 0.67 accuracy, which means that we empirically show the video analysis models can meaningfully detect fatigue state. Also we suggest the way of model adaptation when training and validating video data for classifying fatigue.

A Real-time People Counting Algorithm Using Background Modeling and CNN (배경모델링과 CNN을 이용한 실시간 피플 카운팅 알고리즘)

  • Yang, HunJun;Jang, Hyeok;Jeong, JaeHyup;Lee, Bowon;Jeong, DongSeok
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.54 no.3
    • /
    • pp.70-77
    • /
    • 2017
  • Recently, Internet of Things (IoT) and deep learning techniques have affected video surveillance systems in various ways. The surveillance features that perform detection, tracking, and classification of specific objects in Closed Circuit Television (CCTV) video are becoming more intelligent. This paper presents real-time algorithm that can run in a PC environment using only a low power CPU. Traditional tracking algorithms combine background modeling using the Gaussian Mixture Model (GMM), Hungarian algorithm, and a Kalman filter; they have relatively low complexity but high detection errors. To supplement this, deep learning technology was used, which can be trained from a large amounts of data. In particular, an SRGB(Sequential RGB)-3 Layer CNN was used on tracked objects to emphasize the features of moving people. Performance evaluation comparing the proposed algorithm with existing ones using HOG and SVM showed move-in and move-out error rate reductions by 7.6 % and 9.0 %, respectively.