• 제목/요약/키워드: Consistent Scene Images

검색결과 10건 처리시간 0.027초

PROPAGATION OF MULTI-LEVEL CUES WITH ADAPTIVE CONFIDENCE FOR BILAYER SEGMENTATION OF CONSISTENT SCENE IMAGES

  • Lee, Soo-Chahn;Yun, Il-Dong;Lee, Sang-Uk
    • 한국방송∙미디어공학회:학술대회논문집
    • /
    • 한국방송공학회 2009년도 IWAIT
    • /
    • pp.148-153
    • /
    • 2009
  • Few methods have dealt with segmenting multiple images with analogous content. Concurrent images of a scene and gathered images of a similar foreground are examples of these images, which we term consistent scene images. In this paper, we present a method to segment these images based on manual segmentation of one image, by iteratively propagating information via multi-level cues with adaptive confidence. The cues are classified as low-, mid-, and high- levels based on whether they pertain to pixels, patches, and shapes. Propagated cues are used to compute potentials in an MRF framework, and segmentation is done by energy minimization. Through this process, the proposed method attempts to maximize the amount of extracted information and maximize the consistency of segmentation. We demonstrate the effectiveness of the proposed method on several sets of consistent scene images and provide a comparison with results based only on mid-level cues [1].

  • PDF

다중 단계 신호의 적응적 전파를 통한 동일 장면 영상의 이원 영역화 (Bilayer Segmentation of Consistent Scene Images by Propagation of Multi-level Cues with Adaptive Confidence)

  • 이수찬;윤일동;이상욱
    • 방송공학회논문지
    • /
    • 제14권4호
    • /
    • pp.450-462
    • /
    • 2009
  • 최근까지 단일 영상이나 동영상을 영역화하는 기법들은 다양하게 제시되어 왔으나, 유사한 장면에 대한 여러 장의 영상을 동시에 영역화하는 기법은 많지 않았다. 본 논문에서는 한 장소에서 연속적으로 촬영하였거나 전경 물체가 유사한 여러 영상들을 동일 장면 영상으로 정의하고, 이런 동일 장면 영상들을 적은 양의 사용자 입력을 통해 효과적으로 영역화하는 기법을 제안한다. 구체적으로, 사용자가 최초의 영상 한 장을 직접 영역화한 후, 그 영상의 영역화 결과와 영상의 특성을 토대로 다중 단계 신호를 적응적 가중치를 주어서 인접 영상으로 전파하고, 이를 통해 제안하는 기법은 인접 영상을 반복적으로 영역화한다. 영역화는 마르코프 랜덤 장에서의 에너지 최소화를 통해 이루어지는데, 전파되는 신호는 각 픽셀에 대한 에너지를 정의하는 바탕이 되며, 픽셀, 픽셀 패치, 그리고 영상 전체로부터 비롯되었는가에 따라 낮은 단계, 중간 단계, 그리고 높은 단계의 신호로 지칭된다. 또한 에너지 최소화 틀 안에서 전파된 신호를 통해 정의되는 에너지 역시 낮은 단계, 중간 단계, 그리고 높은 단계의 세 단계로 정의한다. 이런 과정을 통해 전파된 신호를 최대한 다양하게 활용하고, 이를 통해 다양한 영상에 영역화 결과가 일관되게 유지된다. 다양한 동일 장면 영상들에 제안하는 기법을 적용하여 성능을 평가하고, 픽셀 패치를 바탕으로 하는 중간 단계 신호만을 이용한 결과와 제안하는 다중 신호를 적용하는 기법의 결과를 비교한다.

선형특징을 사용한 항공영상의 정합 (Aerial scene matching using linear features)

  • 정재훈;박영태
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 1998년도 하계종합학술대회논문집
    • /
    • pp.689-692
    • /
    • 1998
  • Matching two images is an essential step for many computer vision applications. A new approach to the scale and rotation invariant scene matching is presented. A set of andidate parameters are hypthesized by mapping the angular difference and a new distance measure to the hough space and by detecting maximally consistent points. The proposed method is shown to be much faster than the conventinal one where the relaxation process is repeated until convergence, while providing robust matching performance, without a priori information on the geometrical transformation parameters.

  • PDF

동적 환경에 강인한 장면 인식 기반의 로봇 자율 주행 (Scene Recognition based Autonomous Robot Navigation robust to Dynamic Environments)

  • 김정호;권인소
    • 로봇학회논문지
    • /
    • 제3권3호
    • /
    • pp.245-254
    • /
    • 2008
  • Recently, many vision-based navigation methods have been introduced as an intelligent robot application. However, many of these methods mainly focus on finding an image in the database corresponding to a query image. Thus, if the environment changes, for example, objects moving in the environment, a robot is unlikely to find consistent corresponding points with one of the database images. To solve these problems, we propose a novel navigation strategy which uses fast motion estimation and a practical scene recognition scheme preparing the kidnapping problem, which is defined as the problem of re-localizing a mobile robot after it is undergone an unknown motion or visual occlusion. This algorithm is based on motion estimation by a camera to plan the next movement of a robot and an efficient outlier rejection algorithm for scene recognition. Experimental results demonstrate the capability of the vision-based autonomous navigation against dynamic environments.

  • PDF

선형특징을 사용한 불변 영상정합 기법 (Invariant Image Matching using Linear Features)

  • 박세제;박영태
    • 전자공학회논문지S
    • /
    • 제35S권12호
    • /
    • pp.55-62
    • /
    • 1998
  • 두개의 영상을 정합 하는 것은 많은 컴퓨터 시각장치의 응용과정 중 기본적인 과정이다. 본 논문에서는 선형특징을 사용한 정합기법으로서 회전각도와 크기비율에 불변한 영상정합 기법을 제안한다. 영상은 edge 검출, 세선화, 선형화 과정에 의해 선형 세그먼트의 집합으로 묘사된다. 세그먼트 사이의 각도차이와 새로운 거리척도에 의한 크기비율을 사용해 Hough 공간에서 최대로 일치하는 변환 파라메터를 추정한다. 추정된 파라메터는 1단계 relaxation과 Hough 기법으로 이루어진 고속 선형특징 정합과정에 의해 검증된다. 제안한 기법은 변환 파라메터에 대한 사전정보를 필요로 하지 않으며 추출된 선형 세그먼트 크기의 변화에 민감하지 않은 특성과 기존의 relaxation 기법에 비해 빠른 처리속도를 가진다.

  • PDF

Terrain Geometry from Monocular Image Sequences

  • McKenzie, Alexander;Vendrovsky, Eugene;Noh, Jun-Yong
    • Journal of Computing Science and Engineering
    • /
    • 제2권1호
    • /
    • pp.98-108
    • /
    • 2008
  • Terrain reconstruction from images is an ill-posed, yet commonly desired Structure from Motion task when compositing visual effects into live-action photography. These surfaces are required for choreography of a scene, casting physically accurate shadows of CG elements, and occlusions. We present a novel framework for generating the geometry of landscapes from extremely noisy point cloud datasets obtained via limited resolution techniques, particularly optical flow based vision algorithms applied to live-action video plates. Our contribution is a new statistical approach to remove erroneous tracks ('outliers') by employing a unique combination of well established techniques-including Gaussian Mixture Models (GMMs) for robust parameter estimation and Radial Basis Functions (REFs) for scattered data interpolation-to exploit the natural constraints of this problem. Our algorithm offsets the tremendously laborious task of modeling these landscapes by hand, automatically generating a visually consistent, camera position dependent, thin-shell surface mesh within seconds for a typical tracking shot.

Instant NGP를 활용한 CNC Tool의 장면 생성 및 렌더링 성능 평가 (Scene Generation of CNC Tools Utilizing Instant NGP and Rendering Performance Evaluation)

  • 정태영;유영준
    • 대한임베디드공학회논문지
    • /
    • 제19권2호
    • /
    • pp.83-90
    • /
    • 2024
  • CNC tools contribute to the production of high-precision and consistent results. However, employing damaged CNC tools or utilizing compromised numerical control can lead to significant issues, including equipment damage, overheating, and system-wide errors. Typically, the assessment of external damage to CNC tools involves capturing a single viewpoint through a camera to evaluate tool wear. This study aims to enhance existing methods by using only a single manually focused Microscope camera to enable comprehensive external analysis from multiple perspectives. Applying the NeRF (Neural Radiance Fields) algorithm to images captured with a single manual focus microscope camera, we construct a 3D rendering system. Through this system, it is possible to generate scenes of areas that cannot be captured even with a fixed camera setup, thereby assisting in the analysis of exterior features. However, the NeRF model requires considerable training time, ranging from several hours to over two days. To overcome these limitations of NeRF, various subsequent models have been developed. Therefore, this study aims to compare and apply the performance of Instant NGP, Mip-NeRF, and DS-NeRF, which have garnered attention following NeRF.

Hierarchical Clustering Approach of Multisensor Data Fusion: Application of SAR and SPOT-7 Data on Korean Peninsula

  • Lee, Sang-Hoon;Hong, Hyun-Gi
    • 대한원격탐사학회:학술대회논문집
    • /
    • 대한원격탐사학회 2002년도 Proceedings of International Symposium on Remote Sensing
    • /
    • pp.65-65
    • /
    • 2002
  • In remote sensing, images are acquired over the same area by sensors of different spectral ranges (from the visible to the microwave) and/or with different number, position, and width of spectral bands. These images are generally partially redundant, as they represent the same scene, and partially complementary. For many applications of image classification, the information provided by a single sensor is often incomplete or imprecise resulting in misclassification. Fusion with redundant data can draw more consistent inferences for the interpretation of the scene, and can then improve classification accuracy. The common approach to the classification of multisensor data as a data fusion scheme at pixel level is to concatenate the data into one vector as if they were measurements from a single sensor. The multiband data acquired by a single multispectral sensor or by two or more different sensors are not completely independent, and a certain degree of informative overlap may exist between the observation spaces of the different bands. This dependence may make the data less informative and should be properly modeled in the analysis so that its effect can be eliminated. For modeling and eliminating the effect of such dependence, this study employs a strategy using self and conditional information variation measures. The self information variation reflects the self certainty of the individual bands, while the conditional information variation reflects the degree of dependence of the different bands. One data set might be very less reliable than others in the analysis and even exacerbate the classification results. The unreliable data set should be excluded in the analysis. To account for this, the self information variation is utilized to measure the degrees of reliability. The team of positively dependent bands can gather more information jointly than the team of independent ones. But, when bands are negatively dependent, the combined analysis of these bands may give worse information. Using the conditional information variation measure, the multiband data are split into two or more subsets according the dependence between the bands. Each subsets are classified separately, and a data fusion scheme at decision level is applied to integrate the individual classification results. In this study. a two-level algorithm using hierarchical clustering procedure is used for unsupervised image classification. Hierarchical clustering algorithm is based on similarity measures between all pairs of candidates being considered for merging. In the first level, the image is partitioned as any number of regions which are sets of spatially contiguous pixels so that no union of adjacent regions is statistically uniform. The regions resulted from the low level are clustered into a parsimonious number of groups according to their statistical characteristics. The algorithm has been applied to satellite multispectral data and airbone SAR data.

  • PDF

화재대응 취약지역에서의 소방특수차량 이동제약요인 분석 : 서울시의 진입곤란지역을 대상으로 (Analysis of Mobility Constraint Factors of Fire Engines in Vulnerable Areas : A Case Study of Difficult-to-access Areas in Seoul)

  • 윤여름;김태은;최민지;황성주
    • 한국안전학회지
    • /
    • 제39권1호
    • /
    • pp.62-69
    • /
    • 2024
  • Ensuring swift on-site access to fire engines is crucial in preserving the golden time and minimizing damage. However, various mobility constraints in alleyways hinder the timely entry of fire engines to the fire scene, significantly impairing their initial response capabilities. Therefore, this study analyzed the significant mobility constraints of fire engines, focusing on Seoul, which has many old town areas. By leveraging survey responses from firefighting experts and actual observations, this study quantitatively assessed the frequency and severity of mobility constraint factors affecting the disaster responses of fire engines. Survey results revealed a consistent set of top five factors regarding the frequency and disturbance level, including illegally parked cars, narrow paths, motorcycles, poles, and awnings/banners. A comparison with actual road-view images showed notable consistency between the survey and observational results regarding the appearance frequency of mobility constraint factors in vulnerable areas in Seoul. Furthermore, the study emphasized the importance of tailored management strategies for each mobility constraint factor, considering its characteristics, such as dynamic or static. The findings of this study can serve as foundational data for creating more detailed fire safety maps and advancing technologies that monitor the mobility of fire engines through efficient vision-based inference using CCTVs in the future.

증강현실 캐릭터 구현을 위한 AI기반 객체인식 연구 (AI-Based Object Recognition Research for Augmented Reality Character Implementation)

  • 이석환;이정금;심현
    • 한국전자통신학회논문지
    • /
    • 제18권6호
    • /
    • pp.1321-1330
    • /
    • 2023
  • 본 연구는 증강현실에서 적용할 캐릭터 생성에서 단일 이미지를 통해 여러 객체에 대한 3D 자세 추정 문제를 연구한다. 기존 top-down 방식에서는 이미지 내의 모든 객체를 먼저 감지하고, 그 후에 각각의 객체를 독립적으로 재구성한다. 문제는 이렇게 재구성된 객체들 사이의 중첩이나 깊이 순서가 불일치 하는 일관성 없는 결과가 발생할 수 있다. 본 연구의 목적은 이러한 문제점을 해결하고, 장면 내의 모든 객체에 대한 일관된 3D 재구성을 제공하는 단일 네트워크를 개발하는 것이다. SMPL 매개변수체를 기반으로 한 인체 모델을 top-down 프레임워크에 통합이 중요한 선택이 되었으며, 이를 통해 거리 필드 기반의 충돌 손실과 깊이 순서를 고려하는 손실 두 가지를 도입하였다. 첫 번째 손실은 재구성된 사람들 사이의 중첩을 방지하며, 두 번째 손실은 가림막 추론과 주석이 달린 인스턴스 분할을 일관되게 렌더링하기 위해 객체들의 깊이 순서를 조정한다. 이러한 방법은 네트워크에 이미지의 명시적인 3D 주석 없이도 깊이 정보를 제공하게 한다. 실험 결과, 기존의 Interpenetration loss 방법은 MuPoTS-3D가 114, PoseTrack이 654에 비해서 본 연구의 방법론인 Lp 손실로 네트워크를 훈련시킬 때 MuPoTS-3D가 34, PoseTrack이 202로 충돌수가 크게 감소하는 것으로 나타났다. 본 연구 방법은 표준 3D 자세벤치마크에서 기존 방법보다 더 나은 성능을 보여주었고, 제안된 손실들은 자연 이미지에서 더욱 일관된 재구성을 실현하게 하였다.