• Title/Summary/Keyword: distance between frames

Search Result 94, Processing Time 0.022 seconds

Removal of Search Point using Motion Vector Correlation and Distance between Reference Frames in H.264/AVC (움직임 벡터의 상관도와 참조 화면의 거리를 이용한 H.264/AVC 움직임 탐색 지점 제거)

  • Moon, Ji-Hee;Choi, Jung-Ah;Ho, Yo-Sung
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37 no.2A
    • /
    • pp.113-118
    • /
    • 2012
  • In this paper, we propose the removal of search point using motion vector correlation and distance between reference frames in H.264/AVC. We remove the search points in full search method and predictive motion vectors in enhanced predictive zonal search method. Since the probability that the reference frame far from the current frame is selected as the best reference frame is decreased, we apply the weighted average based on distance between the current and reference frame to determine the fianl search range. In general, the size of search range is smaller than initial search range. We reduce motion estimation time using the final search range in full search method. Also, the refinement process is adaptively applied to each reference frame. The proposed methods reduce the computational throughput of full search method by 57.13% and of enhanced predictive zonal search by 14.71% without visible performance degradation.

Multiple Face Segmentation and Tracking Based on Robust Hausdorff Distance Matching

  • Park, Chang-Woo;Kim, Young-Ouk;Sung, Ha-Gyeong
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2003.09a
    • /
    • pp.632-635
    • /
    • 2003
  • This paper describes a system fur tracking multiple faces in an input video sequence using facial convex hull based facial segmentation and robust hausdorff distance. The algorithm adapts skin color reference map in YCbCr color space and hair color reference map in RGB color space for classifying face region. Then, we obtain an initial face model with preprocessing and convex hull. For tracking, this algorithm computes displacement of the point set between frames using a robust hausdorff distance and the best possible displacement is selected. Finally, the initial face model is updated using the displacement. We provide an example to illustrate the proposed tracking algorithm, which efficiently tracks rotating and zooming faces as well as existing multiple faces in video sequences obtained from CCD camera.

  • PDF

Multiple Face Segmentation and Tracking Based on Robust Hausdorff Distance Matching

  • Park, Chang-Woo;Kim, Young-Ouk;Sung, Ha-Gyeong;Park, Mignon
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.3 no.1
    • /
    • pp.87-92
    • /
    • 2003
  • This paper describes a system for tracking multiple faces in an input video sequence using facial convex hull based facial segmentation and robust hausdorff distance. The algorithm adapts skin color reference map in YCbCr color space and hair color reference map in RGB color space for classifying face region. Then, we obtain an initial face model with preprocessing and convex hull. For tracking, this algorithm computes displacement of the point set between frames using a robust hausdorff distance and the best possible displacement is selected. Finally, the initial face model is updated using the displacement. We provide an example to illustrate the proposed tracking algorithm, which efficiently tracks rotating and zooming faces as well as existing multiple faces in video sequences obtained from CCD camera.

Interactive Facial Expression Animation of Motion Data using CCA (CCA 투영기법을 사용한 모션 데이터의 대화식 얼굴 표정 애니메이션)

  • Kim Sung-Ho
    • Journal of Internet Computing and Services
    • /
    • v.6 no.1
    • /
    • pp.85-93
    • /
    • 2005
  • This paper describes how to distribute high multi-dimensional facial expression data of vast quantity over a suitable space and produce facial expression animations by selecting expressions while animator navigates this space in real-time. We have constructed facial spaces by using about 2400 facial expression frames on this paper. These facial spaces are created by calculating of the shortest distance between two random expressions. The distance between two points In the space of expression, which is manifold space, is described approximately as following; When the linear distance of them is shorter than a decided value, if the two expressions are adjacent after defining the expression state vector of facial status using distance matrix expressing distance between two markers, this will be considered as the shortest distance (manifold distance) of the two expressions. Once the distance of those adjacent expressions was decided, We have taken a Floyd algorithm connecting these adjacent distances to yield the shortest distance of the two expressions. We have used CCA(Curvilinear Component Analysis) technique to visualize multi-dimensional spaces, the form of expressing space, into two dimensions. While the animators navigate this two dimensional spaces, they produce a facial animation by using user interface in real-time.

  • PDF

Graph-based High-level Motion Segmentation using Normalized Cuts (Normalized Cuts을 이용한 그래프 기반의 하이레벨 모션 분할)

  • Yun, Sung-Ju;Park, An-Jin;Jung, Kee-Chul
    • Journal of KIISE:Software and Applications
    • /
    • v.35 no.11
    • /
    • pp.671-680
    • /
    • 2008
  • Motion capture devices have been utilized in producing several contents, such as movies and video games. However, since motion capture devices are expensive and inconvenient to use, motions segmented from captured data was recycled and synthesized to utilize it in another contents, but the motions were generally segmented by contents producers in manual. Therefore, automatic motion segmentation is recently getting a lot of attentions. Previous approaches are divided into on-line and off-line, where ow line approaches segment motions based on similarities between neighboring frames and off-line approaches segment motions by capturing the global characteristics in feature space. In this paper, we propose a graph-based high-level motion segmentation method. Since high-level motions consist of repeated frames within temporal distances, we consider similarities between neighboring frames as well as all similarities among all frames within the temporal distance. This is achieved by constructing a graph, where each vertex represents a frame and the edges between the frames are weighted by their similarity. Then, normalized cuts algorithm is used to partition the constructed graph into several sub-graphs by globally finding minimum cuts. In the experiments, the results using the proposed method showed better performance than PCA-based method in on-line and GMM-based method in off-line, as the proposed method globally segment motions from the graph constructed based similarities between neighboring frames as well as similarities among all frames within temporal distances.

An Efficient Video Sequence Matching Algorithm (효율적인 비디오 시퀀스 정합 알고리즘)

  • 김상현;박래홍
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.41 no.5
    • /
    • pp.45-52
    • /
    • 2004
  • According tothe development of digital media technologies various algorithms for video sequence matching have been proposed to match the video sequences efficiently. A large number of video sequence matching methods have focused on frame-wise query, whereas a relatively few algorithms have been presented for video sequence matching or video shot matching. In this paper, we propose an efficientalgorithm to index the video sequences and to retrieve the sequences for video sequence query. To improve the accuracy and performance of video sequence matching, we employ the Cauchy function as a similarity measure between histograms of consecutive frames, which yields a high performance compared with conventional measures. The key frames extracted from segmented video shots can be used not only for video shot clustering but also for video sequence matching or browsing, where the key frame is defined by the frame that is significantly different from the previous fames. Several key frame extraction algorithms have been proposed, in which similar methods used for shot boundary detection were employed with proper similarity measures. In this paper, we propose the efficient algorithm to extract key frames using the cumulative Cauchy function measure and. compare its performance with that of conventional algorithms. Video sequence matching can be performed by evaluating the similarity between data sets of key frames. To improve the matching efficiency with the set of extracted key frames we employ the Cauchy function and the modified Hausdorff distance. Experimental results with several color video sequences show that the proposed method yields the high matching performance and accuracy with a low computational load compared with conventional algorithms.

Speech/Music Discrimination Using Multi-dimensional MMCD (다차원 MMCD를 이용한 음성/음악 판별)

  • Choi, Mu-Yeol;Song, Hwa-Jeon;Park, Seul-Han;Kim, Hyung-Soon
    • Proceedings of the KSPS conference
    • /
    • 2006.11a
    • /
    • pp.142-145
    • /
    • 2006
  • Discrimination between speech and music is important in many multimedia applications. Previously we proposed a new parameter for speech/music discrimination, the mean of minimum cepstral distances (MMCD), and it outperformed the conventional parameters. One weakness of it is that its performance depends on range of candidate frames to compute the minimum cepstral distance, which requires the optimal selection of the range experimentally. In this paper, to alleviate the problem, we propose a multi-dimensional MMCD parameter which consists of multiple MMCDs with different ranges of candidate frames. Experimental results show that the multi-dimensional MMCD parameter yields an error rate reduction of 22.5% compared with the optimally chosen one-dimensional MMCD parameter.

  • PDF

Auto Setup Method of Best Expression Transfer Path at the Space of Facial Expressions (얼굴 표정공간에서 최적의 표정전이경로 자동 설정 방법)

  • Kim, Sung-Ho
    • The KIPS Transactions:PartA
    • /
    • v.14A no.2
    • /
    • pp.85-90
    • /
    • 2007
  • This paper presents a facial animation and expression control method that enables the animator to select any facial frames from the facial expression space, whose expression transfer paths the system can setup automatically. Our system creates the facial expression space from approximately 2500 captured facial frames. To create the facial expression space, we get distance between pairs of feature points on the face and visualize the space of expressions in 2D space by using the Multidimensional scaling(MDS). To setup most suitable expression transfer paths, we classify the facial expression space into four field on the basis of any facial expression state. And the system determine the state of expression in the shortest distance from every field, then the system transfer from the state of any expression to the nearest state of expression among thats. To complete setup, our system continue transfer by find second, third, or fourth near state of expression until finish. If the animator selects any key frames from facial expression space, our system setup expression transfer paths automatically. We let animators use the system to create example animations or to control facial expression, and evaluate the system based on the results.

Experimental Study on Spray Etching Process In Micro Fabrication of Lead Frame

  • Jung, Ji-Won;Choi, Gyung-Min;Kim, Duck-Jool
    • Journal of Mechanical Science and Technology
    • /
    • v.18 no.12
    • /
    • pp.2294-2302
    • /
    • 2004
  • The objective of this study is to obtain detailed information for the micro fabrication of lead frames by applying spray technology to wet etching process. Wet etching experiments were performed with different etching parameters such as injection pressure, distance from nozzle tip to etched substrate, nozzle pitch and etchant temperature. The characteristics of single and twin spray were measured to investigate the correlation between the spray characteristics and the etching characteristics. Drop size and velocity were measured by Phase-Doppler Anemometer (PDA). Four liquids of different viscosity were used to reveal the effects of viscosity on the spray characteristics. The results indicated that the shorter the distance from nozzle tip and the nozzle pitch, the larger etching factor became. The average etching factor had good positive correlation with average axial velocity and impact force. It was found that the etching characteristics depended strongly on the spray characteristics.

Variable Time-Scale Modification of Speech Using Transient Information based on LPC Cepstral Distance (LPC 켑스트럼 거리 기반의 천이구간 정보를 이용한 음성의 가변적인 시간축 변환)

  • Lee, Sung-Joo;Kim, Hee-Dong;Kim, Hyung-Soon
    • Speech Sciences
    • /
    • v.3
    • /
    • pp.167-176
    • /
    • 1998
  • Conventional time-scale modification methods have the problem that as the modification rate gets higher the time-scale modified speech signal becomes less intelligible, because they ignore the effect of articulation rate on speech characteristics. Results of research on speech perception show that the timing information of transient portions of a speech signal plays an important role in discriminating among different speech sounds. Inspired by this fact, we propose a novel scheme for modifying the time-scale of speech. In the proposed scheme, the timing information of the transient portions of speech is preserved, while the steady portions of speech are compressed or expanded somewhat excessively for maintaining overall time-scale change. In order to identify the transient and steady portions of a speech signal, we employ a simple method using LPC cepstral distance between neighboring frames. The result of the subjective preference test indicates that the proposed method produces performance superior to that of the conventional SOLA method, especially for very fast playback case.

  • PDF