• Title/Summary/Keyword: 다시점 영상 집합

Search Result 6, Processing Time 0.021 seconds

Joint Segmentation of Multi-View Images by Region Correspondence (영역 대응을 이용한 다시점 영상 집합의 통합 영역화)

  • Lee, Soo-Chahn;Kwon, Dong-Jin;Yun, Il-Dong;Lee, Sang-Uk
    • Journal of Broadcast Engineering
    • /
    • v.13 no.5
    • /
    • pp.685-695
    • /
    • 2008
  • This paper presents a method to segment the object of interest from a set of multi-view images with minimal user interaction. Specifically, after the user segments an initial image, we first estimate the transformations between foreground and background of the segmented image and the neighboring image, respectively. From these transformations, we obtain regions in the neighboring image that respectively correspond to the foreground and the background of the segmented image. We are then able to segment the neighboring image based on these regions, and iterate this process to segment the whole image set. Transformation of foregrounds are estimated by feature-based registration with free-form deformation, while transformation of backgrounds are estimated by homography constrained to affine transformation. Here, both are based on correspondence point pairs. Segmentation is done by estimating pixel color distributions and defining a shape prior based on the obtained foreground and background regions and applying them to a Markov random field (MRF) energy minimization framework for image segmentation. Experimental results demonstrate the effectiveness of the proposed method.

Multiview Tracking using Active Shape Model (능동형태모델 기반 다시점 영상 추적)

  • Im, Jae-Hyun;Kim, Dae-Hee;Choi, Jong-Ho;Paik, Joon-Ki
    • KSCI Review
    • /
    • v.15 no.1
    • /
    • pp.179-183
    • /
    • 2007
  • 다시점에서의 다중 객체 추적은 여러 분야에서 연구되고 있다. 다시점 영상 추적은 두 객체가 서로 근접하면 하나로 인식하는 문제점을 가지고 있다. 이러한 문제를 해결하기 위한 하나의 방법으로 능동형태모델(active shape mode: ASM)을 들 수 있다 ASM은 훈련집합을 이용하여 다른 객체에 가려진 목표 객체를 추적할 수 있다. 본 논문에서는 겹쳐진 객체를 추적하기 위해 ASM 기반의 다시점 추적 알고리듬(Multi-view tracking using ASM: MVTA)에 대해서 제안한다. 제안된 추적 방법은 (i) 영상 획득, (ii) 객체 추출, (iii) 객체 추적, 그리고 (iv) 현재 형태의 업데이트, 4가지 단계로 나눌 수 있다. 첫 번째 단계에서는 여러 대의 카메라를 사용해서 다시점 영상을 획득하며, 두 번째 단계에서는 객체를 배경으로부터 분리하며, 겹쳐진 객체로부터 목표 객체를 분리해낸다. 세 번째 단계에서는 추적을 위해 ASM을 사용하며, 마지막 단계인 네 번째 단계는 현재 입력 영상의 업데이트이다. 실험결과 제안한 MVTA는 겹쳐진 객체를 추적 시에 생기는 문제에 대해서 향상 된 결과를 보여준다.

  • PDF

Comparison Study of the Performance of CNN Models with Multi-view Image Set on the Classification of Ship Hull Blocks (다시점 영상 집합을 활용한 선체 블록 분류를 위한 CNN 모델 성능 비교 연구)

  • Chon, Haemyung;Noh, Jackyou
    • Journal of the Society of Naval Architects of Korea
    • /
    • v.57 no.3
    • /
    • pp.140-151
    • /
    • 2020
  • It is important to identify the location of ship hull blocks with exact block identification number when scheduling the shipbuilding process. The wrong information on the location and identification number of some hull block can cause low productivity by spending time to find where the exact hull block is. In order to solve this problem, it is necessary to equip the system to track the location of the blocks and to identify the identification numbers of the blocks automatically. There were a lot of researches of location tracking system for the hull blocks on the stockyard. However there has been no research to identify the hull blocks on the stockyard. This study compares the performance of 5 Convolutional Neural Network (CNN) models with multi-view image set on the classification of the hull blocks to identify the blocks on the stockyard. The CNN models are open algorithms of ImageNet Large-Scale Visual Recognition Competition (ILSVRC). Four scaled hull block models are used to acquire the images of ship hull blocks. Learning and transfer learning of the CNN models with original training data and augmented data of the original training data were done. 20 tests and predictions in consideration of five CNN models and four cases of training conditions are performed. In order to compare the classification performance of the CNN models, accuracy and average F1-Score from confusion matrix are adopted as the performance measures. As a result of the comparison, Resnet-152v2 model shows the highest accuracy and average F1-Score with full block prediction image set and with cropped block prediction image set.

Motion Activity Estimation for Mobile Interface Control (모바일 인터페이스 제어를 위한 움직임 추정 기법)

  • Lee, Chul-Woo;Kim, Chang-Su
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2008.11a
    • /
    • pp.135-138
    • /
    • 2008
  • 본 논문에서는 휴대폰이나 UMPC 등의 모바일 기기에 내장된 카메라를 이용하여 입력 영상을 통해 전역적인 움직임 벡터를 취득하고 이를 이용해서 모바일 인터페이스를 제어하는 기법을 제안한다. 카메라로부터 입력되는 영상에서 특징점을 추출하고 광흐름을 기반으로 각각의 특징점에 대한 움직임을 추정한다. 그 과정을 통해서 생성된 움직임 벡터의 집합으로부터 affine 행렬을 계산하여 전체 화상의 움직임을 표현하는 파라미터를 도출할 수 있다. 움직임 파라미터 값은 다시 인터페이스를 제어하는 신호를 생성하며 이 움직임 신호는 메뉴 네비게이션, 슬라이드 쇼 및 문서 스크롤과 같은 모바일 인터페이스의 제어에 이용될 수 있다. 모의 실험을 통하여 인터페이스 제어를 위한 화상의 움직임 정보가 적절히 획득됨을 확인한다.

  • PDF

Segment-based Foreground Extraction Dedicated to 3D Reconstruction (3차원 복원을 위한 세그멘트 기반의 전경물체 추출)

  • Kim, Jeong-Hwan;Park, An-Jin;Jeong, Gi-Cheol
    • 한국HCI학회:학술대회논문집
    • /
    • 2009.02a
    • /
    • pp.625-630
    • /
    • 2009
  • Researches of image-based 3D reconstruction have recently produced a number of good results, but they assumed that the accurate foreground to be reconstructed is already extracted from each input image. This paper proposes a novel approach to extract more accurate foregrounds by iteratively performing foreground extraction and 3D reconstruction in a manner similar to an EM algorithm on regions segmented in an initial stage, called segments. Here, the segments should preserve foreground boundaries to compensate for the boundary errors generated by visual hull, simple 3D reconstruction to minimize the computational time, and should also be composed of the small number of sets to minimize the user input. Therefore, we utilize image segmentation using the graph-cuts method, which minimizes energy function composed of data and smoothness terms, and the two methods are iteratively performed until the energy function is optimized. In the experiments, more accurate results of the foreground, especially in boundaries, were obtained, although the proposed method used a simple 3D reconstruction method.

  • PDF

VILODE : A Real-Time Visual Loop Closure Detector Using Key Frames and Bag of Words (VILODE : 키 프레임 영상과 시각 단어들을 이용한 실시간 시각 루프 결합 탐지기)

  • Kim, Hyesuk;Kim, Incheol
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.4 no.5
    • /
    • pp.225-230
    • /
    • 2015
  • In this paper, we propose an effective real-time visual loop closure detector, VILODE, which makes use of key frames and bag of visual words (BoW) based on SURF feature points. In order to determine whether the camera has re-visited one of the previously visited places, a loop closure detector has to compare an incoming new image with all previous images collected at every visited place. As the camera passes through new places or locations, the amount of images to be compared continues growing. For this reason, it is difficult for a visual loop closure detector to meet both real-time constraint and high detection accuracy. To address the problem, the proposed system adopts an effective key frame selection strategy which selects and compares only distinct meaningful ones from continuously incoming images during navigation, and so it can reduce greatly image comparisons for loop detection. Moreover, in order to improve detection accuracy and efficiency, the system represents each key frame image as a bag of visual words, and maintains indexes for them using DBoW database system. The experiments with TUM benchmark datasets demonstrates high performance of the proposed visual loop closure detector.