• Title/Summary/Keyword: real-time foreground segmentation

Search Result 13, Processing Time 0.025 seconds

Real-Time Foreground Segmentation and Background Substitution for Protecting Privacy on Visual Communication (화상 통신에서의 사생활 보호를 위한 실시간 전경 분리 및 배경 대체)

  • Bae, Gun-Tae;Kwak, Soo-Yeong;Byun, Hye-Ran
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.5C
    • /
    • pp.505-513
    • /
    • 2009
  • This paper proposes a real-time foreground segmentation and background substitution method for protecting the privacy on visual communication. Previous works on this topic have some problems with the color and shape of foreground and the capture device such as stereo camera. we provide a solution which can segment the foreground in real-time using fixed mono camera. For improving the performance of a foreground extraction, we propose the Temporal Foreground Probability Model (TFPM) by modeling temporal information of a video. Also we provide an boundary processing method for natural and smooth synthesizing that using alpha matte and simple post-processing method.

A Real-time SoC Design of Foreground Object Segmentation (Foreground 객체 추출을 위한 실시간 SoC 설계)

  • Kim Ji-Su;Lee Tae-Ho;Lee Hyuk-Jae
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.43 no.9 s.351
    • /
    • pp.44-52
    • /
    • 2006
  • Recently developed MPEG-4 Part 2 compression standard provides a novel capability to handle arbitrary video objects. To support this capability, an efficient object segmentation technique is required. This paper proposes a real-time algorithm for foreground object segmentation in video sequences. The proposed algorithm consists of two steps: the first step that segments a video frame into multiple sub-regions using Spatio-Temporal Watershed Transform and the second step in which a foreground object segment is extracted from the sub-regions generated in the first step. For real-time processing, the algorithm is partitioned into hardware and software parts so that computationally expensive parts are off-loaded from a processor and executed by hardware accelerators. Simulation results show that the proposed implementation can handle QCIF-size video at 15 fps and extracts an accurate foreground object.

Video Segmentation Using DCT and Guided Filter in real time (DCT와 Guided 필터를 이용한 실시간 영상 분류)

  • Shin, Hyunhak;Lee, Zucheul;Kim, Wonha
    • Journal of Broadcast Engineering
    • /
    • v.20 no.5
    • /
    • pp.718-727
    • /
    • 2015
  • In this paper, we present a novel segmentation method that can extract new foreground objects from a current frame in real-time. It is performed by detecting differences between the current frame and reference frame taken from a fixed camera. We minimize computing complexity for real-time video processing. First DCT (Discrete Cosine Transform) is utilized to generate rough binary segmentation maps where foreground and background regions are separated. DCT shows better result of texture analysis than previous methods where texture analysis is performed in spatial domain. It is because texture analysis in frequency domain is easier than that in special domain and intensity and texture in DCT are taken into account at the same time. We maximize run-time efficiency of DCT by considering color information to analyze object region prior to DCT process. Last we use Guided filter for natural matting of the generated binary segmentation map. In general, Guided filter can enhance quality of intermediate result by incorporating guidance information. However, it shows some limitations in homogeneous area. Therefore, we present an additional method which can overcome them.

Maritime Object Segmentation and Tracking by using Radar and Visual Camera Integration

  • Hwang, Jae-Jeong;Cho, Sang-Gyu;Lee, Jung-Sik;Park, Sang-Hyon
    • Journal of information and communication convergence engineering
    • /
    • v.8 no.4
    • /
    • pp.466-471
    • /
    • 2010
  • We have proposed a method to detect and track moving ships using position from Radar and image processor. Real-time segmentation of moving regions in image sequences is a fundamental step in the radar-camera integrated system. Algorithms for segmentation of objects are implemented by composing of background subtraction, morphologic operation, connected components labeling, region growing, and minimum enclosing rectangle. Once the moving objects are detected, tracking is only performed upon pixels labeled as foreground with reduced additional computational burdens.

A Robust Object Extraction Method for Immersive Video Conferencing (몰입형 화상 회의를 위한 강건한 객체 추출 방법)

  • Ahn, Il-Koo;Oh, Dae-Young;Kim, Jae-Kwang;Kim, Chang-Ick
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.48 no.2
    • /
    • pp.11-23
    • /
    • 2011
  • In this paper, an accurate and fully automatic video object segmentation method is proposed for video conferencing systems in which the real-time performance is required. The proposed method consists of two steps: 1) accurate object extraction on the initial frame, 2) real-time object extraction from the next frame using the result of the first step. Object extraction on the initial frame starts with generating a cumulative edge map obtained from frame differences in the beginning. This is because we can estimate the initial shape of the foreground object from the cumulative motion. This estimated shape is used to assign the seeds for both object and background, which are needed for Graph-Cut segmentation. Once the foreground object is extracted by Graph-Cut segmentation, real-time object extraction is conducted using the extracted object and the double edge map obtained from the difference between two successive frames. Experimental results show that the proposed method is suitable for real-time processing even in VGA resolution videos contrary to previous methods, being a useful tool for immersive video conferencing systems.

Foreground Segmentation and High-Resolution Depth Map Generation Using a Time-of-Flight Depth Camera (깊이 카메라를 이용한 객체 분리 및 고해상도 깊이 맵 생성 방법)

  • Kang, Yun-Suk;Ho, Yo-Sung
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37C no.9
    • /
    • pp.751-756
    • /
    • 2012
  • In this paper, we propose a foreground extraction and depth map generation method using a time-of-flight (TOF) depth camera. Although, the TOF depth camera captures the scene's depth information in real-time, it has a built-in noise and distortion. Therefore, we perform several preprocessing steps such as image enhancement, segmentation, and 3D warping, and then use the TOF depth data to generate the depth-discontinuity regions. Then, we extract the foreground object and generate the depth map as of the color image. The experimental results show that the proposed method efficiently generates the depth map even for the object boundary and textureless regions.

An effective background subtraction in dynamic scene. (동적 환경에서의 효과적인 움직이는 객체 추출)

  • Han, Jae-Hyek;Kim, Yong-Jin;Ryu, Sae-Woon;Lee, Sang-Hwa;Park, Jong-Il
    • 한국HCI학회:학술대회논문집
    • /
    • 2009.02a
    • /
    • pp.631-636
    • /
    • 2009
  • Foreground segmentation methods have steadily been researched in the field of computer vision. Especially, background subtraction which extracts a foreground image from the difference between the current frame and a reference image, called as "background image" have been widely used for a variety of real-time applications because of low computation and high-quality. However, if the background scene was dynamically changed, the background subtraction causes lots of errors. In this paper, we propose an efficient background subtraction method in dynamic environment with both static and dynamic scene. The proposed method is a hybrid method that uses the conventional background subtraction for static scene and depth information for dynamic scene. Its validity and efficiency are verified by demonstration in dynamic environment, where a video projector projects various images in the background.

  • PDF

A Study on HMM-Based Segmentation Method for Traffic Monitoring (HMM 분할에 기반한 교통모니터링에 관한 연구)

  • Hwang, Suen-Ki;Kang, Yong-Seok;Kim, Tae-Woo;Kim, Hyun-Yul;Park, Young-Cheol;Bae, Cheol-Soo
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.5 no.1
    • /
    • pp.1-6
    • /
    • 2012
  • In this paper, we propose a HMM(Hidden Markov Model)-based segmentation method to model shadows as well as foreground and background regions. The shadow of moving objects often keeps from visual tracking. We propose an HMM-based segmentation method which classifies each object in real time. In the case of traffic monitoring movies, the effectiveness of the proposed method was proved by experiments.

An HMM-Based Segmentation Method for Traffic Monitoring (HMM 분할에 기반한 교통모니터링)

  • 남기환;배철수;정주병;나상동
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2004.05b
    • /
    • pp.587-590
    • /
    • 2004
  • In this paper proposed a HMM(Hidden Martov Model)-based segmentation method which is able to model shadows as well as foreground and background regions. Shadow of moving objects often obstruct visual tracking. We propose an HMM-based segmentation method which classifies in real time oath objects. In the case of traffic monitoring movies, the effectiveness of the proposed method has been proven through experimental results

  • PDF

Optical Recognition of Credit Card Numbers (신용카드 번호의 광학적 인식)

  • Jung, Min Chul
    • Journal of the Semiconductor & Display Technology
    • /
    • v.13 no.1
    • /
    • pp.57-62
    • /
    • 2014
  • This paper proposes a new optical recognition method of credit card numbers. Firstly, the proposed method segments numbers from the input image of a credit card. It uses the significant differences of standard deviations between the foreground numbers and the background. Secondly, the method extracts gradient features from the segmented numbers. The gradient features are defined as four directions of grayscale pixels for 16 regions of an input number. Finally, it utilizes an artificial neural network classifier that uses an error back-propagation algorithm. The proposed method is implemented using C language in an embedded Linux system for a high-speed real-time image processing. Experiments were conducted by using real credit card images. The results show that the proposed algorithm is quite successful for most credit cards. However, the method fails in some credit cards with strong background patterns.