• Title/Summary/Keyword: Computer Vision System

Search Result 1,056, Processing Time 0.026 seconds

Deep Learning-based Action Recognition using Skeleton Joints Mapping (스켈레톤 조인트 매핑을 이용한 딥 러닝 기반 행동 인식)

  • Tasnim, Nusrat;Baek, Joong-Hwan
    • Journal of Advanced Navigation Technology
    • /
    • v.24 no.2
    • /
    • pp.155-162
    • /
    • 2020
  • Recently, with the development of computer vision and deep learning technology, research on human action recognition has been actively conducted for video analysis, video surveillance, interactive multimedia, and human machine interaction applications. Diverse techniques have been introduced for human action understanding and classification by many researchers using RGB image, depth image, skeleton and inertial data. However, skeleton-based action discrimination is still a challenging research topic for human machine-interaction. In this paper, we propose an end-to-end skeleton joints mapping of action for generating spatio-temporal image so-called dynamic image. Then, an efficient deep convolution neural network is devised to perform the classification among the action classes. We use publicly accessible UTD-MHAD skeleton dataset for evaluating the performance of the proposed method. As a result of the experiment, the proposed system shows better performance than the existing methods with high accuracy of 97.45%.

Digital Image based Real-time Sea Fog Removal Technique using GPU (GPU를 이용한 영상기반 고속 해무제거 기술)

  • Choi, Woon-sik;Lee, Yoon-hyuk;Seo, Young-ho;Choi, Hyun-jun
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.20 no.12
    • /
    • pp.2355-2362
    • /
    • 2016
  • Seg fog removal is an important issue concerned by both computer vision and image processing. Sea fog or haze removal is widely used in lots of fields, such as automatic control system, CCTV, and image recognition. Color image dehazing techniques have been extensively studied, and expecially the dark channel prior(DCP) technique has been widely used. This paper propose a fast and efficient image prior - dark channel prior to remove seg-fog from a single digital image based on the GPU. We implement the basic parallel program and then optimize it to obtain performance acceleration with more than 250 times. While paralleling and the optimizing the algorithm, we improve some parts of the original serial program or basic parallel program according to the characteristics of several steps. The proposed GPU programming algorithm and implementation results may be used with advantages as pre-processing in many systems, such as safe navigation for ship, topographical survey, intelligent vehicles, etc.

Real-Time Object Recognition Using Local Features (지역 특징을 사용한 실시간 객체인식)

  • Kim, Dae-Hoon;Hwang, Een-Jun
    • Journal of IKEEE
    • /
    • v.14 no.3
    • /
    • pp.224-231
    • /
    • 2010
  • Automatic detection of objects in images has been one of core challenges in the areas such as computer vision and pattern analysis. Especially, with the recent deployment of personal mobile devices such as smart phone, such technology is required to be transported to them. Usually, these smart phone users are equipped with devices such as camera, GPS, and gyroscope and provide various services through user-friendly interface. However, the smart phones fail to give excellent performance due to limited system resources. In this paper, we propose a new scheme to improve object recognition performance based on pre-computation and simple local features. In the pre-processing, we first find several representative parts from similar type objects and classify them. In addition, we extract features from each classified part and train them using regression functions. For a given query image, we first find candidate representative parts and compare them with trained information to recognize objects. Through experiments, we have shown that our proposed scheme can achieve resonable performance.

Implementation of Digital Light Drawing System based on Stereo Vision (스테레오 비전 기반 Light Drawing 시스템 구현)

  • Park, Won-Bae;Park, Chang-Bum;Paik, Doo-Won
    • The Journal of the Korea Contents Association
    • /
    • v.10 no.2
    • /
    • pp.130-137
    • /
    • 2010
  • Light Drawing is a photographic technique which exposures are made at night or in a darkened room usually by moving a hand-held light source[1]. Due to the limitations of equipment and environment, users having difficulty in drawing a picture in 3D space. If user take a light drawing, they need a camera that have function and darkened environment. Alternative solution is that we can make a light drawing picture by using the computer drawing tool as in Photoshop. Nevertheless, this solution will let the User lose their interest in drawing because this solution cannot synchronize between the real action of human hand motion and the electronic input devices such as mouse and keyboard. This paper proposed a digital content that can make light drawing easier. We used a digital content that will facility Light Drawing easier. We can measure the light spot position by using the stereo camera. Based on the measured position of the light spot, we reproduce light drawing in virtual space by using drawing effect method.

OWC based Smart TV Remote Controller Design Using Flashlight

  • Mariappan, Vinayagam;Lee, Minwoo;Choi, Byunghoon;Kim, Jooseok;Lee, Jisung;Choi, Seongjhin
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.10 no.1
    • /
    • pp.71-76
    • /
    • 2018
  • The technology convergence of television, communication, and computing devices enables the rich social and entertaining experience through Smart TV in personal living space. The powerful smart TV computing platform allows to provide various user interaction interfaces like IR remote control, web based control, body gesture based control, etc. The presently used smart TV interaction user control methods are not efficient and user-friendly to access different type of media content and services and strongly required advanced way to control and access to the smart TV with easy user interface. This paper propose the optical wireless communication (OWC) based remote controller design for Smart TV using smart device Flashlights. In this approach, the user smart device act as a remote controller with touch based interactive smart device application and transfer the user control interface data to smart TV trough Flashlight using visible light communication method. The smart TV built-in camera follows the optical camera communication (OCC) principle to decode data and control smart TV user access functions according. This proposed method is not harmful as radio frequency (RF) radiation does it on human health and very simple to use as well user does need to any gesture moves to control the smart TV.

A Study on Auto Inspection System of Cross Coil Movement Using Machine Vision (머신비젼을 이용한 Cross Coil Movement 자동검사 시스템에 관한 연구)

  • Lee, Chul-Hun;Seol, Sung-Wook;Joo, Jae-Heum;Lee, Sang-Chan;Nam, Ki-Gon
    • Journal of the Korean Institute of Telematics and Electronics S
    • /
    • v.36S no.11
    • /
    • pp.79-88
    • /
    • 1999
  • In this paper we address the tracking method which tracks only target object in image sequence including moving object. We use a contour tracking algorithm based on intensity and motion boundaries. The motion of the moving object contour in the image is assumed to be well describable by an affine motion model with a translation, a change in scale and a rotation. The moving object contour is represented by B-spline, the position and motion of which is estimated along the image sequence. we use pattern recognition to identify target object. In order to use linear Kalman Filters we decompose the estimation process into two filters. One is estimating the affine motion parameters and the other the shape of moving object contour. In some experiments with dial plate we show that this method enables us to obtain the robust motion estimates and tracking trajectories even in case of including obstructive object.

  • PDF

Real Time Eye and Gaze Tracking (실시간 눈과 시선 위치 추적)

  • Cho, Hyeon-Seob;Kim, Hee-Sook
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.6 no.2
    • /
    • pp.195-201
    • /
    • 2005
  • This paper describes preliminary results we have obtained in developing a computer vision system based on active IR illumination for real time gaze tracking for interactive graphic display. Unlike most of the existing gaze tracking techniques, which often require assuming a static head to work well and require a cumbersome calibration process for each person, our gaze tracker can perform robust and accurate gaze estimation without calibration and under rather significant head movement. This is made possible by a new gaze calibration procedure that identifies the mapping from pupil parameters to screen coordinates using the Generalized Regression Neural Networks (GRNN). With GRNN, the mapping does not have to be an analytical function and head movement is explicitly accounted for by the gaze mapping function. Furthermore, the mapping function can generalize to other individuals not used in the training. The effectiveness of our gaze tracker is demonstrated by preliminary experiments that involve gaze-contingent interactive graphic display.

  • PDF

Analyzing and Solving GuessWhat?! (GuessWhat?! 문제에 대한 분석과 파훼)

  • Lee, Sang-Woo;Han, Cheolho;Heo, Yujung;Kang, Wooyoung;Jun, Jaehyun;Zhang, Byoung-Tak
    • Journal of KIISE
    • /
    • v.45 no.1
    • /
    • pp.30-35
    • /
    • 2018
  • GuessWhat?! is a game in which two machine players, composed of questioner and answerer, ask and answer yes-no-N/A questions about the object hidden for the answerer in the image, and the questioner chooses the correct object. GuessWhat?! has received much attention in the field of deep learning and artificial intelligence as a testbed for cutting-edge research on the interplay of computer vision and dialogue systems. In this study, we discuss the objective function and characteristics of the GuessWhat?! game. In addition, we propose a simple solver for GuessWhat?! using a simple rule-based algorithm. Although a human needs four or five questions on average to solve this problem, the proposed method outperforms state-of-the-art deep learning methods using only two questions, and exceeds human performance using five questions.

Application of Area Based Matching for the Automation of Interior Orientation (내부표정의 자동화를 위한 영역중심 영상정합기법 적용)

  • 유복모;염재홍;김원대
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.17 no.4
    • /
    • pp.321-330
    • /
    • 1999
  • Automation of observation and positioning of fiducial marks is made possible with the application of image matching technique, developed through the cooperative research effort of computer vision and digital photogrammetry. The major problem in such automation effort is to minimize the computing time and to increase the positional accuracy. Except for scanning and ground control surveying, the interior orientation process was automated in this study, through the development of an algorithm which applies the image matching and image processing techniques. The developed system was applied to close-range photogrammetry and the analysis of the results showed 54% improvement in processing time. For fiducial mark observation during interior orientation, the Laplacian of Gaussian transformation and the Hough transformation were applied to determine the accurate position of the center point, and the correlation matching and the least squares matching method were then applied to improve the accuracy of automated observation of fiducial marks. Image pyramid concept was applied to reduce the computing time of automated positioning of fiducial mark.

  • PDF

Automatic Recognition of the Front/Back Sides and Stalk States for Mushrooms(Lentinus Edodes L.) (버섯 전후면과 꼭지부 상태의 자동 인식)

  • Hwang, H.;Lee, C.H.
    • Journal of Biosystems Engineering
    • /
    • v.19 no.2
    • /
    • pp.124-137
    • /
    • 1994
  • Visual features of a mushroom(Lentinus Edodes, L.) are critical in grading and sorting as most agricultural products are. Because of its complex and various visual features, grading and sorting of mushrooms have been done manually by the human expert. To realize the automatic handling and grading of mushrooms in real time, the computer vision system should be utilized and the efficient and robust processing of the camera captured visual information be provided. Since visual features of a mushroom are distributed over the front and back sides, recognizing sides and states of the stalk including the stalk orientation from the captured image is a prime process in the automatic task processing. In this paper, the efficient and robust recognition process identifying the front and back side and the state of the stalk was developed and its performance was compared with other recognition trials. First, recognition was tried based on the rule set up with some experimental heuristics using the quantitative features such as geometry and texture extracted from the segmented mushroom image. And the neural net based learning recognition was done without extracting quantitative features. For network inputs the segmented binary image obtained from the combined type automatic thresholding was tested first. And then the gray valued raw camera image was directly utilized. The state of the stalk seriously affects the measured size of the mushroom cap. When its effect is serious, the stalk should be excluded in mushroom cap sizing. In this paper, the stalk removal process followed by the boundary regeneration of the cap image was also presented. The neural net based gray valued raw image processing showed the successful results for our recognition task. The developed technology through this research may open the new way of the quality inspection and sorting especially for the agricultural products whose visual features are fuzzy and not uniquely defined.

  • PDF