• Title/Summary/Keyword: Webcam

Search Result 86, Processing Time 0.032 seconds

Real-time Human Pose Estimation using RGB-D images and Deep Learning

  • Rim, Beanbonyka;Sung, Nak-Jun;Ma, Jun;Choi, Yoo-Joo;Hong, Min
    • Journal of Internet Computing and Services
    • /
    • v.21 no.3
    • /
    • pp.113-121
    • /
    • 2020
  • Human Pose Estimation (HPE) which localizes the human body joints becomes a high potential for high-level applications in the field of computer vision. The main challenges of HPE in real-time are occlusion, illumination change and diversity of pose appearance. The single RGB image is fed into HPE framework in order to reduce the computation cost by using depth-independent device such as a common camera, webcam, or phone cam. However, HPE based on the single RGB is not able to solve the above challenges due to inherent characteristics of color or texture. On the other hand, depth information which is fed into HPE framework and detects the human body parts in 3D coordinates can be usefully used to solve the above challenges. However, the depth information-based HPE requires the depth-dependent device which has space constraint and is cost consuming. Especially, the result of depth information-based HPE is less reliable due to the requirement of pose initialization and less stabilization of frame tracking. Therefore, this paper proposes a new method of HPE which is robust in estimating self-occlusion. There are many human parts which can be occluded by other body parts. However, this paper focuses only on head self-occlusion. The new method is a combination of the RGB image-based HPE framework and the depth information-based HPE framework. We evaluated the performance of the proposed method by COCO Object Keypoint Similarity library. By taking an advantage of RGB image-based HPE method and depth information-based HPE method, our HPE method based on RGB-D achieved the mAP of 0.903 and mAR of 0.938. It proved that our method outperforms the RGB-based HPE and the depth-based HPE.

Bubble Popping Augmented Reality System Using a Vibro-Tactile Haptic Mouse (진동촉각 햅틱 마우스 기반 버블포핑 증강현실 시스템)

  • Jung, Da-Un;Lee, Woo-Keun;Jang, Seong-Eun;Kim, Man-Bae
    • Journal of Broadcast Engineering
    • /
    • v.15 no.6
    • /
    • pp.715-722
    • /
    • 2010
  • As one of applications in augmented realities, this paper presents a bubble popping system utilizing a haptic vibro-tactile mouse. In this system, virtual bubbles randomly float in the 3D space. By using the vibro-tactile mouse grabbed by a user, the bubbles are popped when they are touched by the mouse in the 3D space. Then a bubble popping effect with addition mouse vibration is delivered to the user's hand through the mouse. The proposed system is developed on ARToolkit environment. Therefore, basic components such as a camera and a marker pattern are required. The systems is composed of a vibro-haptic mouse, a webcam, a marker pattern, a graphic bubble object, and graphic mouse. Mouse vibration as well as bubble fade-out effect is delivered. Therefore, the combination of visual and tactile bubble popping effects outperforms the usage of a single effect in the experience of augmented reality.

A Study on Hand Gesture Recognition with Low-Resolution Hand Images (저해상도 손 제스처 영상 인식에 대한 연구)

  • Ahn, Jung-Ho
    • Journal of Satellite, Information and Communications
    • /
    • v.9 no.1
    • /
    • pp.57-64
    • /
    • 2014
  • Recently, many human-friendly communication methods have been studied for human-machine interface(HMI) without using any physical devices. One of them is the vision-based gesture recognition that this paper deals with. In this paper, we define some gestures for interaction with objects in a predefined virtual world, and propose an efficient method to recognize them. For preprocessing, we detect and track the both hands, and extract their silhouettes from the low-resolution hand images captured by a webcam. We modeled skin color by two Gaussian distributions in RGB color space and use blob-matching method to detect and track the hands. Applying the foodfill algorithm we extracted hand silhouettes and recognize the hand shapes of Thumb-Up, Palm and Cross by detecting and analyzing their modes. Then, with analyzing the context of hand movement, we recognized five predefined one-hand or both-hand gestures. Assuming that one main user shows up for accurate hand detection, the proposed gesture recognition method has been proved its efficiency and accuracy in many real-time demos.

Privacy Protection Technologies on IoT Environments: Case Study of Networked Cameras (사물인터넷(IoT) 환경에서 프라이버시 보호 기술: 네트워크 카메라 사례 연구)

  • Kim, Mihui
    • The Journal of the Korea Contents Association
    • /
    • v.16 no.9
    • /
    • pp.329-338
    • /
    • 2016
  • Internet of Things (IoTs) technology makes every things in physical world being digitalized and communicated with each other. The technology is emerging as a new paradigm and is expected to provide a convenient and effective life. However, for the successful realization of the IoT technologies, IoT security issues are an important prerequisite, and particularly the privacy protection is expected to become more important in view of object communication directively related with human. In this paper we describe for the security and privacy threats in IoT environment and introduce the shodan (a legitimate search engine that finds backdoor routers, switches, webcams, IoT devices connected to the Internet etc.) that can expose the security and privacy problems. Lastly, we compare the privacy threats through real-world case study of network cameras currently in use and finally derive the countermeasures for the threats.

Adaptive Background Subtraction Based on Genetic Evolution of the Global Threshold Vector (전역 임계치 벡터의 유전적 진화에 기반한 적응형 배경차분화)

  • Lim, Yang-Mi
    • Journal of Korea Multimedia Society
    • /
    • v.12 no.10
    • /
    • pp.1418-1426
    • /
    • 2009
  • There has been a lot of interest in an effective method for background subtraction in an effort to separate foreground objects from a predefined background image. Promising results on background subtraction using statistical methods have recently been reported are robust enough to operate in dynamic environments, but generally require very large computational resources and still have difficulty in obtaining clear segmentation of objects. We use a simple running-average method to model a gradually changing background, instead of using a complicated statistical technique. We employ a single global threshold vector, optimized by a genetic algorithm, instead of pixel-by-pixel thresholds. A new fitness function is defined and trained to evaluate segmentation result. The system has been implemented on a PC with a webcam, and experimental results on real images show that the new method outperforms an existing method based on a mixture of Gaussian.

  • PDF

Vehicle Displacement Estimation By GPS and Vision Sensor (영상센서/GPS에 기반한 차량의 이동변위 추정)

  • Kim, Min-Woo;Lim, Joon-Hoo;Park, Je-Doo;Kim, Hee-Sung;Lee, Hyung-Keun
    • Journal of Advanced Navigation Technology
    • /
    • v.16 no.3
    • /
    • pp.417-425
    • /
    • 2012
  • It is well known that GPS cannot provide positioning results if sufficient number of visible satellites are not available. To overcome this weak point, attentions have been recently moved to hybrid positioning methods that augments GPS with other sensors. As an extension of hybrid positiong methods, this paper proposes a new method that combines GPS and vision sensor to improve availability and accuracy of land vehicle positioning. The proposed method does not require any external map information and can provide position solutions if more than 2 navigation satellites are visible. To evaluate the performance of the proposed method, an experiment result with real measurements is provided and a result shows that accumulated error of n-axis is almost 2.5meters and that of e-axis is almost 3meters in test section.

Webcam-Based 2D Eye Gaze Estimation System By Means of Binary Deformable Eyeball Templates

  • Kim, Jin-Woo
    • Journal of information and communication convergence engineering
    • /
    • v.8 no.5
    • /
    • pp.575-580
    • /
    • 2010
  • Eye gaze as a form of input was primarily developed for users who are unable to use usual interaction devices such as keyboard and the mouse; however, with the increasing accuracy in eye gaze detection with decreasing cost of development, it tends to be a practical interaction method for able-bodied users in soon future as well. This paper explores a low-cost, robust, rotation and illumination independent eye gaze system for gaze enhanced user interfaces. We introduce two brand-new algorithms for fast and sub-pixel precise pupil center detection and 2D Eye Gaze estimation by means of deformable template matching methodology. In this paper, we propose a new algorithm based on the deformable angular integral search algorithm based on minimum intensity value to localize eyeball (iris outer boundary) in gray scale eye region images. Basically, it finds the center of the pupil in order to use it in our second proposed algorithm which is about 2D eye gaze tracking. First, we detect the eye regions by means of Intel OpenCV AdaBoost Haar cascade classifiers and assign the approximate size of eyeball depending on the eye region size. Secondly, using DAISMI (Deformable Angular Integral Search by Minimum Intensity) algorithm, pupil center is detected. Then, by using the percentage of black pixels over eyeball circle area, we convert the image into binary (Black and white color) for being used in the next part: DTBGE (Deformable Template based 2D Gaze Estimation) algorithm. Finally, using DTBGE algorithm, initial pupil center coordinates are assigned and DTBGE creates new pupil center coordinates and estimates the final gaze directions and eyeball size. We have performed extensive experiments and achieved very encouraging results. Finally, we discuss the effectiveness of the proposed method through several experimental results.

A Study on a Smart Digital Signage Using Bayesian Age Estimation Technique for the Next Generation Airport Service (차세대 공항 서비스를 위한 베이지안 연령추정기법을 이용하는 스마트 디지털 사이니지에 대한 연구)

  • Kim, Chun-Ho;Lee, Dong Woo;Baek, Gyeong Min;Moon, Seong Yeop;Heo, Chan;Na, Jong Whoa;Ohn, Seung-Yup;Choi, Woo Young
    • Journal of Advanced Navigation Technology
    • /
    • v.18 no.6
    • /
    • pp.533-540
    • /
    • 2014
  • We propose an age estimation-based smart digital signage for the next-generation airport service. The proposed system can recognize the face of the customer so that it can display the selective information. Using a webcam, the system captures the face of the customer and estimates the age of the customer by calculating the wrinkle density of the face and applying bayesian classifier. The developed age estimation method is tested with a face database for the performance evaluation. We expect the new digital signage may improve the satisfaction of customers of the airport business.

Development of Interactive Content Services through an Intelligent IoT Mirror System (지능형 IoT 미러 시스템을 활용한 인터랙티브 콘텐츠 서비스 구현)

  • Jung, Wonseok;Seo, Jeongwook
    • Journal of Advanced Navigation Technology
    • /
    • v.22 no.5
    • /
    • pp.472-477
    • /
    • 2018
  • In this paper, we develop interactive content services for preventing depression of users through an intelligent Internet of Things(IoT) mirror system. For interactive content services, an IoT mirror device measures attention and meditation data from an EEG headset device and also measures facial expression data such as "sad", "angery", "disgust", "neutral", " happy", and "surprise" classified by a multi-layer perceptron algorithm through an webcam. Then, it sends the measured data to an oneM2M-compliant IoT server. Based on the collected data in the IoT server, a machine learning model is built to classify three levels of depression (RED, YELLOW, and GREEN) given by a proposed merge labeling method. It was verified that the k-nearest neighbor (k-NN) model could achieve about 93% of accuracy by experimental results. In addition, according to the classified level, a social network service agent sent a corresponding alert message to the family, friends and social workers. Thus, we were able to provide an interactive content service between users and caregivers.

Classification of Clothing Using Googlenet Deep Learning and IoT based on Artificial Intelligence (인공지능 기반 구글넷 딥러닝과 IoT를 이용한 의류 분류)

  • Noh, Sun-Kuk
    • Smart Media Journal
    • /
    • v.9 no.3
    • /
    • pp.41-45
    • /
    • 2020
  • Recently, artificial intelligence (AI) and the Internet of things (IoT), which are represented by machine learning and deep learning among IT technologies related to the Fourth Industrial Revolution, are applied to our real life in various fields through various researches. In this paper, IoT and AI using object recognition technology are applied to classify clothing. For this purpose, the image dataset was taken using webcam and raspberry pi, and GoogLeNet, a convolutional neural network artificial intelligence network, was applied to transfer the photographed image data. The clothing image dataset was classified into two categories (shirtwaist, trousers): 900 clean images, 900 loss images, and total 1800 images. The classification measurement results showed that the accuracy of the clean clothing image was about 97.78%. In conclusion, the study confirmed the applicability of other objects using artificial intelligence networks on the Internet of Things based platform through the measurement results and the supplementation of more image data in the future.