• Title/Summary/Keyword: Feature Tracking

Search Result 567, Processing Time 0.024 seconds

Localizing Head and Shoulder Line Using Statistical Learning (통계학적 학습을 이용한 머리와 어깨선의 위치 찾기)

  • Kwon, Mu-Sik
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.32 no.2C
    • /
    • pp.141-149
    • /
    • 2007
  • Associating the shoulder line with head location of the human body is useful in verifying, localizing and tracking persons in an image. Since the head line and the shoulder line, what we call ${\Omega}$-shape, move together in a consistent way within a limited range of deformation, we can build a statistical shape model using Active Shape Model (ASM). However, when the conventional ASM is applied to ${\Omega}$-shape fitting, it is very sensitive to background edges and clutter because it relies only on the local edge or gradient. Even though appearance is a good alternative feature for matching the target object to image, it is difficult to learn the appearance of the ${\Omega}$-shape because of the significant difference between people's skin, hair and clothes, and because appearance does not remain the same throughout the entire video. Therefore, instead of teaming appearance or updating appearance as it changes, we model the discriminative appearance where each pixel is classified into head, torso and background classes, and update the classifier to obtain the appropriate discriminative appearance in the current frame. Accordingly, we make use of two features in fitting ${\Omega}$-shape, edge gradient which is used for localization, and discriminative appearance which contributes to stability of the tracker. The simulation results show that the proposed method is very robust to pose change, occlusion, and illumination change in tracking the head and shoulder line of people. Another advantage is that the proposed method operates in real time.

Real-time Control of Biological Animal Wastewater Treatment Process and Stability of Control Parameters (생물학적 축산폐수 처리공정의 자동제어 방법 및 제어 인자의 안정성)

  • Kim, W.Y.;Jung, J.H.;Ra, C.S.
    • Journal of Animal Science and Technology
    • /
    • v.46 no.2
    • /
    • pp.251-260
    • /
    • 2004
  • The feasibility and stability of ORP, pH(mV) and DO as a real-time control parameter for SBR process were evaluated in this study. During operation, NBP(nitrogen break point) and NKP(nitrate knee point), which reveal the biological and chemical changes of pollutants, were clearly observed on ORP and pH(mV)-time profiles, and those control points were easily detected by tracking the moving slope changes(MSC). However, when balance of aeration rate to loading rate, or to OUR(oxygen uptake rate), was not optimally maintained, either false NBP was occurred on ORP and DO curves before the appearance of real NBP or specific NBP feature was disappeared on ORP curve. Under that condition, however, very distinct NBP was found on pH(mV)-time profile, and stable detection of that point was feasible by tracking MSC. These results might mean that pH(mV) is superior real-time control parameter for aerobic process than ORP and DO. Meanwhile, as a real-time control parameter for anoxic process, ORP was very stable and more useful parameter than others. Based on these results, a stable real-time control of process can be achieved by using the ORP and pH(mv) parameters in combination rather than using separately. A complete removal of pollutants could be always ensured with this real-time control technology, despite the variations of wastewater and operation condition, as well as an optimization of treatment time and capacity could be feasible.

Speech Activity Detection using Lip Movement Image Signals (입술 움직임 영상 선호를 이용한 음성 구간 검출)

  • Kim, Eung-Kyeu
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.11 no.4
    • /
    • pp.289-297
    • /
    • 2010
  • In this paper, A method to prevent the external acoustic noise from being misrecognized as the speech recognition object is presented in the speech activity detection process for the speech recognition. Also this paper confirmed besides the acoustic energy to the lip movement image signals. First of all, the successive images are obtained through the image camera for personal computer and the lip movement whether or not is discriminated. The next, the lip movement image signal data is stored in the shared memory and shares with the speech recognition process. In the mean time, the acoustic energy whether or not by the utterance of a speaker is verified by confirming data stored in the shared memory in the speech activity detection process which is the preprocess phase of the speech recognition. Finally, as a experimental result of linking the speech recognition processor and the image processor, it is confirmed to be normal progression to the output of the speech recognition result if face to the image camera and speak. On the other hand, it is confirmed not to the output the result of the speech recognition if does not face to the image camera and speak. Also, the initial feature values under off-line are replaced by them. Similarly, the initial template image captured while off-line is replaced with a template image captured under on-line, so the discrimination of the lip movement image tracking is raised. An image processing test bed was implemented to confirm the lip movement image tracking process visually and to analyze the related parameters on a real-time basis. As a result of linking the speech and image processing system, the interworking rate shows 99.3% in the various illumination environments.

Exploring Optimal Threshold of RGB Pixel Values to Extract Road Features from Google Earth (Google Earth에서 도로 추출을 위한 RGB 화소값 최적구간 추적)

  • Park, Jae-Young;Um, Jung-Sup
    • Journal of Korea Spatial Information System Society
    • /
    • v.12 no.1
    • /
    • pp.66-75
    • /
    • 2010
  • The authors argues that the current road updating system based on traditional aerial photograph or multi-spectral satellite image appears to be non-user friendly due to lack of the frequent cartographic representation for the new construction sites. Google Earth are currently being emerged as one of important places to extract road features since the RGB satellite image with high multi-temporal resolution can be accessed freely over large areas. This paper is primarily intended to evaluate optimal threshold of RGB pixel values to extract road features from Google Earth. An empirical study for five experimental sites was conducted to confirm how a RGB picture provided Google Earth can be used to extact the road feature. The results indicate that optimal threshold of RGB pixel values to extract road features was identified as 126, 125, 127 for manual operation which corresponds to 25%, 30%, 19%. Also, it was found that display scale difference of Google Earth was not very influential in tracking required RGB pixel value. As a result the 61cm resolution of Quickbird RGB data has shown the potential to realistically identified the major type of road feature by large scale spatial precision while the typical algorithm revealed successfully the area-wide optimal threshold of RGB pixel for road appeared in the study area.

The Multi Knowledge-based Image Retrieval Technology for An Automobile Head Lamp Retrieval (자동차 전조등 검색을 위한 다중지식기반의 영상검색 기법)

  • 이병일;손병환;홍성욱;손성건;최흥국
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.3 no.3
    • /
    • pp.27-35
    • /
    • 2002
  • A knowledge-based image retrieval technique is image searching methods using some features from the queried image. The materials in this study are automobile head lamps. The input data is composed of characters and images which have various pattern. The numbers, special symbols, and general letters are under the category of the character. The image informations are made up of the distribution of pixel data, statistical analysis, and state of pattern which are useful for the knowledge data. In this paper, we implemented a retrieval system for the scientific crime detection at traffic accident using the proposed multi knowledge-based image retrieval technique. The values for the multi knowledge-based image features were extracted from color and gray scale each. With this 22 features, we improved the retrieval efficiency about the color information and pattern information. Visual basic, crystal report and MS access DB were used for this application. We anticipate the efficient scientific detection for the traffic accident and the tracking of suspicious vehicle.

  • PDF

A Study on the Features of Visual-Information Acquirement Shown at Searching of Spatial Information - With the Experiment of Observing the Space of Hall in Subway Station - (공간정보의 탐색과정에 나타난 시각정보획득특성에 관한 연구 - 지하철 홀 공간의 주시실험을 대상으로 -)

  • Kim, Jong-Ha
    • Korean Institute of Interior Design Journal
    • /
    • v.23 no.2
    • /
    • pp.90-98
    • /
    • 2014
  • This study has analyzed the meaning of observation time in the course of acquiring the information of subjects who observed the space of hall in subway stations to figure out the process of spatial information excluded and the features of intensive searching. The followings are the results from the analysis of searching process with the interpretation of the process for information acquirement through the interpretation of observation area and time. First, based on the general definition of observation time, the reason for analyzing the features of acquiring spatial information according to the subjects' observation time has been established. The feature of decreased analysis data reflected that of observation time in the process of perceiving and recognizing spatial information, which showed that the observation was focused on the enter of the space during the time spent in the process of observing the space and the spent time with considerable exclusion of bottom end (in particular, right bottom end). Second, while the subjects were observing the space of hall in subway stations, they focused on the top of the left center and the signs on the right exit the most, which was followed by the focus on the both side horizontally and the clock on the top. Third, the analysis of consecutive observation frequency enabled the comparison of the changes to the observation concentration by area. The difference of time by area produced the data with which the change to the contents of spatial searching in the process of searching space could be known. Fourth, as the observation frequency in the area of I changed [three times -> six times -> 9 times], the observation time included in the area increased, which showed the process for the change from perception to recognition of information with the concentration of attention through visual information. It makes it possible to understand that more time was spent on the information to be acquired with the exclusion of the unnecessary information around.

Realistic Seeing Through Method and Device Through Adaptive Registration between Building Space and Telepresence Indoor Environment

  • Lee, Jun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.25 no.1
    • /
    • pp.101-107
    • /
    • 2020
  • We propose a realistic seeing through visualization methods in mixed reality environment. When a user wants to see specific location beyond a wall in indoor environment. The proposed system recognizes and registers the selected area using environment modelling and feature-based tracking. Then the selected area is diminished and the specific location is visualized in real-time. With the proposed seeing through methods, a user can understand spatial relationship of the building and can easily find the target location. We conducted a user study comparing the seeing through method to conventional indoor navigation service in order to investigate the potential of the proposed seeing through method. The proposed seeing through method was evaluated in navigation time in comparison with conventional approach. The proposed method enable users to navigate target locations 30% faster than the conventional approach.

Gaze Recognition System using Random Forests in Vehicular Environment based on Smart-Phone (스마트 폰 기반 차량 환경에서의 랜덤 포레스트를 이용한 시선 인식 시스템)

  • Oh, Byung-Hun;Chung, Kwang-Woo;Hong, Kwang-Seok
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.15 no.1
    • /
    • pp.191-197
    • /
    • 2015
  • In this paper, we propose the system which recognize the gaze using Random Forests in vehicular environment based on smart-phone. Proposed system is mainly composed of the following: face detection using Adaboost, face component estimation using Histograms, and gaze recognition based on Random Forests. We detect a driver based on the image information with a smart-phone camera, and the face component of driver is estimated. Next, we extract the feature vectors from the estimated face component and recognize gaze direction using Random Forest recognition algorithm. Also, we collected gaze database including a variety gaze direction in real environments for the experiment. In the experiment result, the face detection rate and the gaze recognition rate showed 82.02% and 84.77% average accuracies, respectively.

On Motion Planning for Human-Following of Mobile Robot in a Predictable Intelligent Space

  • Jin, Tae-Seok;Hashimoto, Hideki
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.4 no.1
    • /
    • pp.101-110
    • /
    • 2004
  • The robots that will be needed in the near future are human-friendly robots that are able to coexist with humans and support humans effectively. To realize this, humans and robots need to be in close proximity to each other as much as possible. Moreover, it is necessary for their interactions to occur naturally. It is desirable for a robot to carry out human following, as one of the human-affinitive movements. The human-following robot requires several techniques: the recognition of the moving objects, the feature extraction and visual tracking, and the trajectory generation for following a human stably. In this research, a predictable intelligent space is used in order to achieve these goals. An intelligent space is a 3-D environment in which many sensors and intelligent devices are distributed. Mobile robots exist in this space as physical agents providing humans with services. A mobile robot is controlled to follow a walking human using distributed intelligent sensors as stably and precisely as possible. The moving objects is assumed to be a point-object and projected onto an image plane to form a geometrical constraint equation that provides position data of the object based on the kinematics of the intelligent space. Uncertainties in the position estimation caused by the point-object assumption are compensated using the Kalman filter. To generate the shortest time trajectory to follow the walking human, the linear and angular velocities are estimated and utilized. The computer simulation and experimental results of estimating and following of the walking human with the mobile robot are presented.

Effective Eye Detection for Face Recognition to Protect Medical Information (의료정보 보호를 위해 얼굴인식에 필요한 효과적인 시선 검출)

  • Kim, Suk-Il;Seok, Gyeong-Hyu
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.12 no.5
    • /
    • pp.923-932
    • /
    • 2017
  • In this paper, we propose a GRNN(: Generalized Regression Neural Network) algorithms for new eyes and face recognition identification system to solve the points that need corrective action in accordance with the existing problems of facial movements gaze upon it difficult to identify the user and. Using a Kalman filter structural information elements of a face feature to determine the authenticity of the face was estimated future location using the location information of the current head and the treatment time is relatively fast horizontal and vertical elements of the face using a histogram analysis the detected. And the light obtained by configuring the infrared illuminator pupil effects in real-time detection of the pupil, the pupil tracking was to extract the text print vector. The abstract is to be in fully-justified italicized text as it is here, below the author information.