• Title/Summary/Keyword: feature-based tracking

Search Result 315, Processing Time 0.025 seconds

AdaBoost-based Gesture Recognition Using Time Interval Window Applied Global and Local Feature Vectors with Mono Camera (모노 카메라 영상기반 시간 간격 윈도우를 이용한 광역 및 지역 특징 벡터 적용 AdaBoost기반 제스처 인식)

  • Hwang, Seung-Jun;Ko, Ha-Yoon;Baek, Joong-Hwan
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.22 no.3
    • /
    • pp.471-479
    • /
    • 2018
  • Recently, the spread of smart TV based Android iOS Set Top box has become common. This paper propose a new approach to control the TV using gestures away from the era of controlling the TV using remote control. In this paper, the AdaBoost algorithm is applied to gesture recognition by using a mono camera. First, we use Camshift-based Body tracking and estimation algorithm based on Gaussian background removal for body coordinate extraction. Using global and local feature vectors, we recognized gestures with speed change. By tracking the time interval trajectories of hand and wrist, the AdaBoost algorithm with CART algorithm is used to train and classify gestures. The principal component feature vector with high classification success rate is searched using CART algorithm. As a result, 24 optimal feature vectors were found, which showed lower error rate (3.73%) and higher accuracy rate (95.17%) than the existing algorithm.

Video Augmentation of Virtual Object by Uncalibrated 3D Reconstruction from Video Frames (비디오 영상에서의 비보정 3차원 좌표 복원을 통한 가상 객체의 비디오 합성)

  • Park Jong-Seung;Sung Mee-Young
    • Journal of Korea Multimedia Society
    • /
    • v.9 no.4
    • /
    • pp.421-433
    • /
    • 2006
  • This paper proposes a method to insert virtual objects into a real video stream based on feature tracking and camera pose estimation from a set of single-camera video frames. To insert or modify 3D shapes to target video frames, the transformation from the 3D objects to the projection of the objects onto the video frames should be revealed. It is shown that, without a camera calibration process, the 3D reconstruction is possible using multiple images from a single camera under the fixed internal camera parameters. The proposed approach is based on the simplification of the camera matrix of intrinsic parameters and the use of projective geometry. The method is particularly useful for augmented reality applications to insert or modify models to a real video stream. The proposed method is based on a linear parameter estimation approach for the auto-calibration step and it enhances the stability and reduces the execution time. Several experimental results are presented on real-world video streams, demonstrating the usefulness of our method for the augmented reality applications.

  • PDF

Far Distance Face Detection from The Interest Areas Expansion based on User Eye-tracking Information (시선 응시 점 기반의 관심영역 확장을 통한 원 거리 얼굴 검출)

  • Park, Heesun;Hong, Jangpyo;Kim, Sangyeol;Jang, Young-Min;Kim, Cheol-Su;Lee, Minho
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.49 no.9
    • /
    • pp.113-127
    • /
    • 2012
  • Face detection methods using image processing have been proposed in many different ways. Generally, the most widely used method for face detection is an Adaboost that is proposed by Viola and Jones. This method uses Haar-like feature for image learning, and the detection performance depends on the learned images. It is well performed to detect face images within a certain distance range, but if the image is far away from the camera, face images become so small that may not detect them with the pre-learned Haar-like feature of the face image. In this paper, we propose the far distance face detection method that combine the Aadaboost of Viola-Jones with a saliency map and user's attention information. Saliency Map is used to select the candidate face images in the input image, face images are finally detected among the candidated regions using the Adaboost with Haar-like feature learned in advance. And the user's eye-tracking information is used to select the interest regions. When a subject is so far away from the camera that it is difficult to detect the face image, we expand the small eye gaze spot region using linear interpolation method and reuse that as input image and can increase the face image detection performance. We confirmed the proposed model has better results than the conventional Adaboost in terms of face image detection performance and computational time.

Nonlinear observer for flexible joint robots (유연한 관절 로보트에 대한 비선형 관측기)

  • 김윤재;임규만;함철주;함운철
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1993.10a
    • /
    • pp.648-653
    • /
    • 1993
  • This paper presents an nonlinear observer scheme for flexible joint robot manipulators. This nonlinear observer scheme is based on the sliding mode method. Sliding controllers have recently been shown to feature excellent robustness and performance properties for specific classes of nonlinear tracking problems. Dynamic equations of flexible joint robot manipulators are derived from the Euler-Lagrange equations by forming the corresponding Lagrangian. Simulation results are presented to show the validness of the proposed nonlinear observer scheme.

  • PDF

Facial Point Classifier using Convolution Neural Network and Cascade Facial Point Detector (컨볼루셔널 신경망과 케스케이드 안면 특징점 검출기를 이용한 얼굴의 특징점 분류)

  • Yu, Je-Hun;Ko, Kwang-Eun;Sim, Kwee-Bo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.22 no.3
    • /
    • pp.241-246
    • /
    • 2016
  • Nowadays many people have an interest in facial expression and the behavior of people. These are human-robot interaction (HRI) researchers utilize digital image processing, pattern recognition and machine learning for their studies. Facial feature point detector algorithms are very important for face recognition, gaze tracking, expression, and emotion recognition. In this paper, a cascade facial feature point detector is used for finding facial feature points such as the eyes, nose and mouth. However, the detector has difficulty extracting the feature points from several images, because images have different conditions such as size, color, brightness, etc. Therefore, in this paper, we propose an algorithm using a modified cascade facial feature point detector using a convolutional neural network. The structure of the convolution neural network is based on LeNet-5 of Yann LeCun. For input data of the convolutional neural network, outputs from a cascade facial feature point detector that have color and gray images were used. The images were resized to $32{\times}32$. In addition, the gray images were made into the YUV format. The gray and color images are the basis for the convolution neural network. Then, we classified about 1,200 testing images that show subjects. This research found that the proposed method is more accurate than a cascade facial feature point detector, because the algorithm provides modified results from the cascade facial feature point detector.

Context Aware Feature Selection Model for Salient Feature Detection from Mobile Video Devices (모바일 비디오기기 위에서의 중요한 객체탐색을 위한 문맥인식 특성벡터 선택 모델)

  • Lee, Jaeho;Shin, Hyunkyung
    • Journal of Internet Computing and Services
    • /
    • v.15 no.6
    • /
    • pp.117-124
    • /
    • 2014
  • Cluttered background is a major obstacle in developing salient object detection and tracking system for mobile device captured natural scene video frames. In this paper we propose a context aware feature vector selection model to provide an efficient noise filtering by machine learning based classifiers. Since the context awareness for feature selection is achieved by searching nearest neighborhoods, known as NP hard problem, we apply a fast approximation method with complexity analysis in details. Separability enhancement in feature vector space by adding the context aware feature subsets is studied rigorously using principal component analysis (PCA). Overall performance enhancement is quantified by the statistical measures in terms of the various machine learning models including MLP, SVM, Naïve Bayesian, CART. Summary of computational costs and performance enhancement is also presented.

Detection of QRS Feature Based on Phase Transition Tracking for Premature Ventricular Contraction Classification (조기심실수축 분류를 위한 위상 변이 추적 기반의 QRS 특징점 검출)

  • Cho, Ik-sung;Yoon, Jeong-oh;Kwon, Hyeog-soong
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.20 no.2
    • /
    • pp.427-436
    • /
    • 2016
  • In general, QRS duration represent a distance of Q start and S end point. However, since criteria of QRS duration are vague and Q, S point is not detected accurately, arrhythmia classification performance can be reduced. In this paper, we propose extraction of Q, S start and end point RS feature based on phase transition tracking method after we detected R wave that is large peak of electrocardiogram(ECG) signal. For this purpose, we detected R wave, from noise-free ECG signal through the preprocessing method. Also, we classified QRS pattern through differentiation value of ECG signal and extracted Q, S start and end point by tracking direction and count of phase based on R wave. The performance of R wave detection is evaluated by using 48 record of MIT-BIH arrhythmia database. The achieved scores indicate the average detection rate of 99.60%. PVC classification is evaluated by using 9 record of MIT-BIH arrhythmia database that included over 30 premature ventricular contraction(PVC). The achieved scores indicate the average detection rate of 94.12% in PVC.

Head Orientation-based Gaze Tracking (얼굴의 움직임을 이용한 응시점 추적)

  • ;R.S. Ramakrishna
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 1999.10b
    • /
    • pp.401-403
    • /
    • 1999
  • 본 논문에서 우리는 제약이 없는 배경화면에서 얼굴의 움직임을 이용한 응시점 추적을 위해 얼굴의 특징점(눈, 코, 그리고 입)들을 찾고 head orientation을 구하는 효?거이고 빠른 방법을 제안한다. 얼굴을 찾는 방법이 많이 연구 되어 오고 있으나 많은 부분이 효과적이지 못하거나 제한적인 사항을 필요로 한다. 본 논문에서 제안한 방법은 이진화된 이미지에 기초하고 완전 그래프 매칭을 이용한 유사성을 구하는 방법이다. 즉, 임의의 임계치 값에 의해 이진화된 이미지를 레이블링 한 후 각 쌍의 블록에 대한 유사성을 구한다. 이때 두 눈과 가장 유사성을 갖는 두 블록을 눈으로 선택한다. 눈을 찾은 후 입과 코를 찾아간다. 360$\times$240 이미지의 평균 처리 속도는 0.2초 이내이고 다음 탐색영역을 예상하여 탐색 영역을 줄일 경우 평균 처리속도는 0.15초 이내였다. 그리고 본 논문에서는 얼굴의 움직임을 구하기 위해 각 특징점들이 이루는 각을 기준으로 한 템플릿 매칭을 이용했다. 실험은 다양한 조명환경과 여러 사용자를 대상으로 이루어졌고 속도와 정확성면에서 좋은 결과를 보였다. 도한, 명안정보만을 사용하므로 흑백가메라에서도 사용가능하여 경제적 효과도 기대할 수 있다.

  • PDF

Fish's Activity Analysis through Frequency Analysis of Angle Information (움직임 각도의 주파수 분석을 통한 활동성 분석)

  • Kim, Cheol-Ki
    • The Journal of the Korea Contents Association
    • /
    • v.7 no.5
    • /
    • pp.10-15
    • /
    • 2007
  • This paper proposes the method that detects abnormal trajectory of fish with tracking data. And it is obtained by automatic tracking system based on conventional computer vision. Also, we analyze the trajectory using subband frequency features through DWT(Discrete Wavelet Transform). Through experimental results, we confirm that our results have some statistical means. The proposed method demonstrates that DWT is useful method for detecting presence of toxicoid features in environment as for an alternative of bio-monitoring tool.

A neural network based real-time robot tracking controller using position sensitive detectors (신경회로망과 위치 검출장치를 사용한 로보트 추적 제어기의 구현)

  • 박형권;오세영;김성권
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1993.10a
    • /
    • pp.660-665
    • /
    • 1993
  • Neural networks are used in the framework of sensorbased tracking control of robot manipulators. They learn by practice movements the relationship between PSD ( an analog Position Sensitive Detector) sensor readings for target positions and the joint commands to reach them. Using this configuration, the system can track or follow a moving or stationary object in real time. Furthermore, an efficient neural network architecture has been developed for real time learning. This network uses multiple sets of simple backpropagation networks one of which is selected according to which division (corresponding to a cluster of the self-organizing feature map) in data space the current input data belongs to. This lends itself to a very fast training and processing implementation required for real time control.

  • PDF