• Title/Summary/Keyword: feature points

Search Result 1,119, Processing Time 0.033 seconds

Non-Prior Training Active Feature Model-Based Object Tracking for Real-Time Surveillance Systems (실시간 감시 시스템을 위한 사전 무학습 능동 특징점 모델 기반 객체 추적)

  • 김상진;신정호;이성원;백준기
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.41 no.5
    • /
    • pp.23-34
    • /
    • 2004
  • In this paper we propose a feature point tracking algorithm using optical flow under non-prior taming active feature model (NPT-AFM). The proposed algorithm mainly focuses on analysis non-rigid objects[1], and provides real-time, robust tracking by NPT-AFM. NPT-AFM algorithm can be divided into two steps: (i) localization of an object-of-interest and (ii) prediction and correction of the object position by utilizing the inter-frame information. The localization step was realized by using a modified Shi-Tomasi's feature tracking algoriam[2] after motion-based segmentation. In the prediction-correction step, given feature points are continuously tracked by using optical flow method[3] and if a feature point cannot be properly tracked, temporal and spatial prediction schemes can be employed for that point until it becomes uncovered again. Feature points inside an object are estimated instead of its shape boundary, and are updated an element of the training set for AFH Experimental results, show that the proposed NPT-AFM-based algerian can robustly track non-rigid objects in real-time.

Hierarchical Feature Based Block Motion Estimation for Ultrasound Image Sequences (초음파 영상을 위한 계층적 특징점 기반 블록 움직임 추출)

  • Kim, Baek-Sop;Shin, Seong-Chul
    • Journal of KIISE:Software and Applications
    • /
    • v.33 no.4
    • /
    • pp.402-410
    • /
    • 2006
  • This paper presents a method for feature based block motion estimation that uses multi -resolution image sequences to obtain the panoramic images in the continuous ultrasound image sequences. In the conventional block motion estimation method, the centers of motion estimation blocks are set at the predetermined and equally spaced locations. This requires the large blocks to include at least one feature, which inevitably requires long estimation time. In this paper, we propose an adaptive method which locates the center of the motion estimation blocks at the feature points. This make it possible to reduce the block size while keeping the motion estimation accuracy The Harris-Stephen corner detector is used to get the feature points. The comer points tend to group together, which cause the error in the global motion estimation. In order to distribute the feature points as evenly as Possible, the image is firstly divided into regular subregions, and a strongest corner point is selected as a feature in each subregion. The ultrasound Images contain speckle patterns and noise. In order to reduce the noise artifact and reduce the computational time, the proposed method use the multi-resolution image sequences. The first algorithm estimates the motion in the smoothed low resolution image, and the estimated motion is prolongated to the next higher resolution image. By this way the size of search region can be reduced in the higher resolution image. Experiments were performed on three types of ultrasound image sequences. These were shown that the proposed method reduces both the computational time (from 77ms to 44ms) and the displaced frame difference (from 66.02 to 58.08).

Fast Natural Feature Tracking Using Optical Flow (광류를 사용한 빠른 자연특징 추적)

  • Bae, Byung-Jo;Park, Jong-Seung
    • The KIPS Transactions:PartB
    • /
    • v.17B no.5
    • /
    • pp.345-354
    • /
    • 2010
  • Visual tracking techniques for Augmented Reality are classified as either a marker tracking approach or a natural feature tracking approach. Marker-based tracking algorithms can be efficiently implemented sufficient to work in real-time on mobile devices. On the other hand, natural feature tracking methods require a lot of computationally expensive procedures. Most previous natural feature tracking methods include heavy feature extraction and pattern matching procedures for each of the input image frame. It is difficult to implement real-time augmented reality applications including the capability of natural feature tracking on low performance devices. The required computational time cost is also in proportion to the number of patterns to be matched. To speed up the natural feature tracking process, we propose a novel fast tracking method based on optical flow. We implemented the proposed method on mobile devices to run in real-time and be appropriately used with mobile augmented reality applications. Moreover, during tracking, we keep up the total number of feature points by inserting new feature points proportional to the number of vanished feature points. Experimental results showed that the proposed method reduces the computational cost and also stabilizes the camera pose estimation results.

A simple approach for circular Arc detection using a least squares fitting and preprocessing (최소자승법과 전 처리를 이용한 원호 검출의 간단한 접근)

  • Nkurunziza, Armel;Kim, Jong-nam
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2016.05a
    • /
    • pp.840-843
    • /
    • 2016
  • The circular arc is a very useful feature for object detection and recognition in industrial environments. In this paper, a new method to detect circular arcs is proposed. The detection of the circular arc includes the estimation of the center, the radius and the two ending points of the arc. This new method is based on determining the best part of the circular arc (part which does not contains outliers points) using 3 points designated along the arc. A least square method is applied to the best part of the arc and the center and the radius of the arc are obtained. The distance between the remaining edge's points (points which are not lying on the best part of the arc) and the radius is used to the two ending points of the arc.

  • PDF

Realizing a Mixed Reality Space Guided by a Virtual Human;Creating a Virtual Human from Incomplete 3-D Motion Data

  • Abe, Shinsuke;Yamaguti, Iku;Tan, Joo Kooi;Ishikawa, Seiji
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2003.10a
    • /
    • pp.1625-1628
    • /
    • 2003
  • Recently the VR technique has evolved into a mixed reality (MR) technique, in which a user can observe a real world in front of him/her as well as virtual objects displayed. This has been realized by the employment of a see-through type HMD (S-HMD). We have been developing a mixed reality space employing the MR technique. The objective of our study is to realize a virtual human that acts as a man-machine interface in the real space. It is important in the study to create a virtual human acting naturally in front of a user. In order to give natural motions to the virtual human, we employ a developed motion capture technique. We have already created various 3-D human motion models by the motion capture technique. In this paper, we present a technique for creating a virtual human using a human model provided by a computer graphics software, 3D Studio Max(C). The main difficulty of this issue is that 3D Studio Max(C) claims 28 feature points for describing a human motion, but the used motion capture system assumes less number of feature points. Therefore a technique is proposed in the paper for producing motion data of 28 feature points from the motion data of less number of feature points or from incomplete motion data. Performance of the proposed technique was examined by observing visually the demonstration of some motions of a created virtual human and overall natural motions were realized.

  • PDF

A Vision-based Approach for Facial Expression Cloning by Facial Motion Tracking

  • Chun, Jun-Chul;Kwon, Oryun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.2 no.2
    • /
    • pp.120-133
    • /
    • 2008
  • This paper presents a novel approach for facial motion tracking and facial expression cloning to create a realistic facial animation of a 3D avatar. The exact head pose estimation and facial expression tracking are critical issues that must be solved when developing vision-based computer animation. In this paper, we deal with these two problems. The proposed approach consists of two phases: dynamic head pose estimation and facial expression cloning. The dynamic head pose estimation can robustly estimate a 3D head pose from input video images. Given an initial reference template of a face image and the corresponding 3D head pose, the full head motion is recovered by projecting a cylindrical head model onto the face image. It is possible to recover the head pose regardless of light variations and self-occlusion by updating the template dynamically. In the phase of synthesizing the facial expression, the variations of the major facial feature points of the face images are tracked by using optical flow and the variations are retargeted to the 3D face model. At the same time, we exploit the RBF (Radial Basis Function) to deform the local area of the face model around the major feature points. Consequently, facial expression synthesis is done by directly tracking the variations of the major feature points and indirectly estimating the variations of the regional feature points. From the experiments, we can prove that the proposed vision-based facial expression cloning method automatically estimates the 3D head pose and produces realistic 3D facial expressions in real time.

Real-time Recognition System of Facial Expressions Using Principal Component of Gabor-wavelet Features (표정별 가버 웨이블릿 주성분특징을 이용한 실시간 표정 인식 시스템)

  • Yoon, Hyun-Sup;Han, Young-Joon;Hahn, Hern-Soo
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.19 no.6
    • /
    • pp.821-827
    • /
    • 2009
  • Human emotion can be reflected by their facial expressions. So, it is one of good ways to understand people's emotions by recognizing their facial expressions. General recognition system of facial expressions had selected interesting points, and then only extracted features without analyzing physical meanings. They takes a long time to find interesting points, and it is hard to estimate accurate positions of these feature points. And in order to implement a recognition system of facial expressions on real-time embedded system, it is needed to simplify the algorithm and reduce the using resources. In this paper, we propose a real-time recognition algorithm of facial expressions that project the grid points on an expression space based on Gabor wavelet feature. Facial expression is simply described by feature vectors on the expression space, and is classified by an neural network with its resources dramatically reduced. The proposed system deals 5 expressions: anger, happiness, neutral, sadness, and surprise. In experiment, average execution time is 10.251 ms and recognition rate is measured as 87~93%.

A neural network model for recognizing facial expressions based on perceptual hierarchy of facial feature points (얼굴 특징점의 지각적 위계구조에 기초한 표정인식 신경망 모형)

  • 반세범;정찬섭
    • Korean Journal of Cognitive Science
    • /
    • v.12 no.1_2
    • /
    • pp.77-89
    • /
    • 2001
  • Applying perceptual hierarchy of facial feature points, a neural network model for recognizing facial expressions was designed. Input data were convolution values of 150 facial expression pictures by Gabor-filters of 5 different sizes and 8 different orientations for each of 39 mesh points defined by MPEG-4 SNHC (Synthetic/Natural Hybrid Coding). A set of multiple regression analyses was performed with the rating value of the affective states for each facial expression and the Gabor-filtered values of 39 feature points. The results show that the pleasure-displeasure dimension of affective states is mainly related to the feature points around the mouth and the eyebrows, while a arousal-sleep dimension is closely related to the feature points around eyes. For the filter sizes. the affective states were found to be mostly related to the low spatial frequency. and for the filter orientations. the oblique orientations. An optimized neural network model was designed on the basis of these results by reducing original 1560(39x5x8) input elements to 400(25x2x8) The optimized model could predict human affective rating values. up to the correlation value of 0.886 for the pleasure-displeasure, and 0.631 for the arousal-sleep. Mapping the results of the optimized model to the six basic emotional categories (happy, sad, fear, angry, surprised, disgusted) fit 74% of human responses. Results of this study imply that, using human principles of recognizing facial expressions, a system for recognizing facial expressions can be optimized even with a a relatively little amount of information.

  • PDF

Methods for Extracting Feature Points from Ultrasound Images (초음파 영상에서의 특징점 추출 방법)

  • Kim, Sung-Jung;Yoo, JaeChern
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2020.01a
    • /
    • pp.59-60
    • /
    • 2020
  • 본 논문에서는 특징점 추출 알고리즘 중 SIFT(Scale Invariant Feature Transform)알고리즘을 사용하여 유의미한 특징점을 추출하기 위한 방법을 제안하고자한다. 추출된 특징점을 실제 이미지에 display 해봄으로써 성능을 확인해본다.

  • PDF

Development of Automatic Feature Recognition System for CAD/CAPP Interface (CAD/CAPP 인터페이스를 위한 형상특징의 자동인식시스템 개발)

  • 오수철;조규갑
    • Transactions of the Korean Society of Mechanical Engineers
    • /
    • v.16 no.1
    • /
    • pp.31-40
    • /
    • 1992
  • This paper presents an automatic feature recognition system for recognizing and extracting feature information needed for the process planning input from a 3D CAD system. A given part is modeled by using the AutoCAD and feature information is automatically extracted from the AutoCAD database. The type of parts considered in this study is prismatic parts composed of faces perpendicular to the X, Y, Z axes and the types of features recognized by the proposed system are through steps, blind steps, through slots, blind slots, and pockets. Features are recognized by using the concept of convex points and concave points. Case studies are implemented to evaluate feasibilities of the function of the proposed system. The developed system is programmed by using Turbo Pascal on the IBM PC/AT on which the AutoCAD and the proposed system are implemented.