• 제목/요약/키워드: 비디오 색인

Search Result 188, Processing Time 0.028 seconds

Robust Human Silhouette Extraction Using Graph Cuts (그래프 컷을 이용한 강인한 인체 실루엣 추출)

  • Ahn, Jung-Ho;Kim, Kil-Cheon;Byun, Hye-Ran
    • Journal of KIISE:Software and Applications
    • /
    • v.34 no.1
    • /
    • pp.52-58
    • /
    • 2007
  • In this paper we propose a new robust method to extract accurate human silhouettes indoors with active stereo camera. A prime application is for gesture recognition of mobile robots. The segmentation of distant moving objects includes many problems such as low resolution, shadows, poor stereo matching information and instabilities of the object and background color distributions. There are many object segmentation methods based on color or stereo information but they alone are prone to failure. Here efficient color, stereo and image segmentation methods are fused to infer object and background areas of high confidence. Then the inferred areas are incorporated in graph cut to make human silhouette extraction robust and accurate. Some experimental results are presented with image sequences taken using pan-tilt stereo camera. Our proposed algorithms are evaluated with respect to ground truth data and proved to outperform some methods based on either color/stereo or color/contrast alone.

Automatic Indexing Algorithm of Golf Video Using Audio Information (오디오 정보를 이용한 골프 동영상 자동 색인 알고리즘)

  • Kim, Hyoung-Gook
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.5
    • /
    • pp.441-446
    • /
    • 2009
  • This paper proposes an automatic indexing algorithm of golf video using audio information. In the proposed algorithm, the input audio stream is demultiplexed into the stream of video and audio. By means of Adaboost-cascade classifier, the continuous audio stream is classified into announcer's speech segment recorded in studio, music segment accompanied with players' names on TV screen, reaction segment of audience according to the play, reporter's speech segment with field background, filed noise segment like wind or waves. And golf swing sound including drive shot, iron shot, and putting shot is detected by the method of impulse onset detection and modulation spectrum verification. The detected swing and applause are used effectively to index action or highlight unit. Compared with video based semantic analysis, main advantage of the proposed system is its small computation requirement so that it facilitates to apply the technology to embedded consumer electronic devices for fast browsing.

Multiple Object Tracking for Surveillance System (감시 시스템을 위한 다중 객체 추적)

  • Cho, Yong-Il;Choi, Jin;Yang, Hyun-Seung
    • 한국HCI학회:학술대회논문집
    • /
    • 2006.02a
    • /
    • pp.653-659
    • /
    • 2006
  • 다중 객체 추적이란 컴퓨터 비전의 한 분야로, 주어진 비디오 시퀀스 내에서 관심 있는 객체들을 추적하는 것을 말한다. 다중 객체 추적 시스템은 감시 시스템, 사용자 행동 인식, 스포츠 중계, 비디오 회의와 같은 다양한 응용 분야에 핵심 기반 기술로 쓰이고 있어 그 중요성이 매우 크다. 본 논문은 감시 목적의 다중 객체를 추적하는 방법에 대하여 다룬다. 감시 시스템의 특성상, 객체의 외관이나 움직임 등에 대한 가정을 하기가 어렵다. 따라서 본 논문에서는 크기, 색, 형태 같은 객체의 단순하고 직관적인 외관 특성을 이용하면서도, 객체들끼리 부분적으로 혹은 완전히 겹쳐졌을 때에도 객체들의 위치를 적절히 추적할 수 있는 방법을 제안한다. 본 논문에서 제안하는 방법은 객체들의 경로에 대한 정보를 유지하는데 그래프 구조를 이용한다. 그래프를 확장하고, 제거하여 영상에 대한 정보를 추론한다. 크게 보면 객체들을 영역 레벨, 객체 레벨 두 단계에 걸쳐 추적한다. 영역 레벨에서는 각 객체들이 있을 수 있을만한 영역에 대한 가설을 세우고, 객체 레벨에서는 각 가설에 대한 검증을 한다. 제안된 방법은 직관적인 정보만을 이용하여 서로 다른 형태의 객체를 빠르게 추적할 수 있음을 보여준다. 다만 객체의 외관 정보만을 이용하였기 추적하기 때문에, 객체가 다른 객체에 의해 완전히 가려진 채 또다시 다른 객체와 겹쳐지면, 정확한 추적이 되지 않는다. 이를 해결하기 위해서는 객체가 겹쳐졌을 때, 그 관계에 대한 정보를 모아야 하는데 이는 향후 연구를 통해 해결하고자 한다.

  • PDF

Multi-modal Detection of Anchor Shot in News Video (다중모드 특징을 사용한 뉴스 동영상의 앵커 장면 검출 기법)

  • Yoo, Sung-Yul;Kang, Dong-Wook;Kim, Ki-Doo;Jung, Kyeong-Hoon
    • Journal of Broadcast Engineering
    • /
    • v.12 no.4
    • /
    • pp.311-320
    • /
    • 2007
  • In this paper, an efficient detection algorithm of an anchor shot in news video is presented. We observed the audio visual characteristics of news video and proposed several low level features which are appropriate for detecting an anchor shot in news video. The overall structure of the proposed algorithm is composed of 3 stages: the pause detection, the audio cluster classification, and the matching with motion activity stage. We used the audio features as well as the motion feature in order to improve the indexing accuracy and the simulation results show that the performance of the proposed algorithm is quite satisfactory.

Adaptive Residual DPCM using Weighted Linear Combination of Adjacent Residues in Screen Content Video Coding (스크린 콘텐츠 비디오의 압축을 위한 인접 화소의 가중 합을 이용한 적응적 Residual DPCM 기법)

  • Kang, Je-Won
    • Journal of Broadcast Engineering
    • /
    • v.20 no.5
    • /
    • pp.782-785
    • /
    • 2015
  • In this paper, we propose a novel residual differential pulse-code modulation (RDPCM) coding technique to improve coding efficiency of screen content videos. The proposed method uses a weighted combination of adjacent residues to provide an accurate estimate in RDPCM. The weights are trained in previously coded samples by using an L1 optimization problem with the least absolute shrinkage and selection operation (LASSO). The proposed method achieves BD-rate saving about 3.1% in all-intra coding.

A Design of LED Video Processor Board using Embedded System (임베디드 시스템을 이용한 LED 비디오 프로세서 설계)

  • Lee, Jong-Ha;Ko, Duck-Young
    • 전자공학회논문지 IE
    • /
    • v.47 no.3
    • /
    • pp.1-6
    • /
    • 2010
  • In this paper, it is designed a processor using embedded system so that moving picture can be expressed on LED electric sign board which has been expressed a simple message only like as a character or graphic. It has been fabricated a moving picture LED electric sign board which is composed to a video processor and LED display panel, in order to be able to express a digital moving picture of 24 bits that is transmitted from embedded system. It includes gamma adjustment, brightness, color contrast control, a schedule function, expression image conversion by the Internet and memory device. Also, an application program based Windows CE is designed so that a character, graphic, and moving picture can be expressed on a small LED electric sign board.

Shot boundary Frame Detection and Key Frame Detection for Multimedia Retrieval (멀티미디어 검색을 위한 shot 경계 및 대표 프레임 추출)

  • 강대성;김영호
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.2 no.1
    • /
    • pp.38-43
    • /
    • 2001
  • This Paper suggests a new feature for shot detection, using the proposed robust feature from the DC image constructed by DCT DC coefficients in the MPEG video stream, and proposes the characterizing value that reflects the characteristic of kind of video (movie, drama, news, music video etc.). The key frames are pulled out from many frames by using the local minima and maxima of differential of the value. After original frame(not do image) are reconstructed for key frame, indexing process is performed through computing parameters. Key frames that are similar to user's query image are retrieved through computing parameters. It is proved that the proposed methods are better than conventional method from experiments. The retrieval accuracy rate is so high in experiments.

  • PDF

Context Extraction and Analysis of Video Life Log Using Bayesian Network (베이지안 네트워크를 이용한 동영상 기반 라이프 로그의 분석 및 의미정보 추출)

  • Jung, Tae-Min;Cho, Sung-Bae
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2010.06c
    • /
    • pp.414-418
    • /
    • 2010
  • 최근 라이프 로그의 수집과 관리에 관련된 연구가 많이 진행 중에 있다. 또 핸드폰 카메라, 디지털 카메라, 캠코더 등의 발전으로 자신의 일상생활을 비디오로 저장하고, 인터넷을 통해 공유하는 사람도 증가하고 있다. 비디오 데이터는 많은 정보를 포함하고 있는 라이프 로그의 한 예로. 동영상의 촬영 및 수집이 활발해짐에 따라 동영상의 메타정보를 생성하고, 이를 이용해 동영상 검색과 관리에 이용하려는 연구들이 진행 중이다. 본 논문에서는 라이프 로그를 수집하고 수집된 동영상과 라이프 로그를 이용하여 의미정보를 추출하는 시스템을 제안한다. 의미정보란 사용자의 행동을 나타내는 정보로써 컴퓨터 사용, 식사, 집안일, 이동, 외출, 독서, 휴식, 일, 기타로 9가지의 의미정보를 추출한다. 제안하는 방법은 사용자로부터 GPS, 가속도센서, 캠코더를 이용해 실제 데이터를 수집하고, 전처리 과정을 통하여 특징을 추출한다. 이때 추출될 특징은 위치정보와 사용자의 상태정보 그리고 영상처리릍 통한 RGB와 HSL 색공간의 요소와 MPEG-7의 EHD(Edge Histogram Descriptor). CLD(Color Layout Descriptor)이다. 추출된 특징으로부터 사람 행동과 같은 불안정한 상황에서 강점을 보이는 확률모델 네트워크인 베이지안 네트워크를 이용하여 의미정보를 추출한다. 제안하는 방법의 유용성을 보이기 위해 실제 데이터를 수집하고 추론하고 10-Fold Cross-validation을 이용하여 데이터를 검증한다.

  • PDF

The Interesting Moving Objects Tracking Algorithm using Color Informations on Multi-Video Camera (다중 비디오카메라에서 색 정보를 이용한 특정 이동물체 추적 알고리듬)

  • Shin, Chang-Hoon;Lee, Joo-Shin
    • The KIPS Transactions:PartB
    • /
    • v.11B no.3
    • /
    • pp.267-274
    • /
    • 2004
  • In this paper, the interesting moving objects tracking algorithm using color information on Multi-Video camera is proposed Moving objects are detected by using difference image method and integral projection method to background image and objects image only with hue area, after converting RGB color coordination of image which is input from multi-video camera into HSI color coordination. Hue information of the detected moving area are normalized by 24 steps from 0$^{\circ}$ to 360$^{\circ}$ It is used for the feature parameters of the moving objects that three normalization levels with the highest distribution and distance among three normalization levels after obtaining a hue distribution chart of the normalized moving objects. Moving objects identity among four cameras is distinguished with distribution of three normalization levels and distance among three normalization levels, and then the moving objects are tracked and surveilled. To examine propriety of the proposed method, four cameras are set up indoor difference places, humans are targeted for moving objects. As surveillance results of the interesting human, hue distribution chart variation of the detected Interesting human at each camera in under 10%, and it is confirmed that the interesting human is tracked and surveilled by using feature parameters at four cameras, automatically.

Generation of Video Clips Utilizing Shot Boundary Detection (샷 경계 검출을 이용한 영상 클립 생성)

  • Kim, Hyeok-Man;Cho, Seong-Kil
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.7 no.6
    • /
    • pp.582-592
    • /
    • 2001
  • Video indexing plays an important role in the applications such as digital video libraries or web VOD which archive large volume of digital videos. Video indexing is usually based on video segmentation. In this paper, we propose a software tool called V2Web Studio which can generate video clips utilizing shot boundary detection algorithm. With the V2Web Studio, the process of clip generation consists of the following four steps: 1) Automatic detection of shot boundaries by parsing the video, 2) Elimination of errors by manually verifying the results of the detection, 3) Building a modeling structure of logical hierarchy using the verified shots, and 4) Generating multiple video clips corresponding to each logically modeled segment. The aforementioned steps are performed by shot detector, shot verifier, video modeler and clip generator in the V2Web Studio respectively.

  • PDF