• Title/Summary/Keyword: 다중 시점 카메라

Search Result 28, Processing Time 0.032 seconds

A 3D Modeling System Using Multiple Stereo Cameras (다중 스테레오 카메라를 이용한 3차원 모델링 시스템)

  • Kim, Han-Sung;Sohn, Kwang-Hoon
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.44 no.1
    • /
    • pp.1-9
    • /
    • 2007
  • In this paper, we propose a new 3D modeling and rendering system using multiple stereo cameras. When target objects are captured by cameras, each capturing PC segments the objects and estimates disparity fields, then they transmit the segmented masks, disparity fields, and color textures of objects to a 3D modeling server. The modeling server generates 3D models of the objects from the gathered masks and disparity fields. Finally, the server generates a video at the designated point of view with the 3D model and texture information from cameras.

Improving Spatial Resolution in Real-time for Ultra-thin Light Field Cameras (초박형 라이트필드 카메라의 실시간 분해능 향상 알고리즘 개발)

  • Kim, Donggun;Ryu, Jaekwan;Jo, Yongjin;Kim, Min H.
    • Journal of the Korea Computer Graphics Society
    • /
    • v.27 no.3
    • /
    • pp.25-29
    • /
    • 2021
  • 초박형 라이트필드 카메라 시스템은 이미지 센서 위에 렌즈 어레이를 부착하는 방식으로 만들어진다. 이러한 초박형 라이트필드 카메라는 하나의 이미지 센서를 여러 개의 sub-aperture가 나눠쓰는 방식으로 되어있어 개별 이미지의 분해능이 낮으며, sub-aperture 이미지들을 융합해 추가적인 분해능 향상이 수행되어야 한다. 본 연구에서는 초박형 라이트필드 카메라 시스템을 개발했으며, 개발된 카메라 시스템을 위한 실시간 분해능 향상 알고리즘을 개발, 실험을 통해 검증했다. 개발된 초박형 라이트필드 카메라는 두께 2mm, 24개(6×4)의 551×551 해상도의 sub-aperture로 구성되어 있으며, 임베디드 컴퓨팅 보드를 사용해 휴대가 가능하도록 제작되었다. 실시간 분해능 향상 알고리즘은 임베디드 컴퓨팅 보드의 GPU에서 병렬처리를 통해 라플라시안 피라미드 기반의 이미지 융합 알고리즘을 수행한다. 실험을 통해 검증한 결과로, 개발 시스템은 MTF50값이 평균 35% 정도 개선되었으며, 10.65fps의 처리속도로 실시간 처리가 가능함을 확인했다.

Multi-View Image Deblurring for 3D Shape Reconstruction (3차원 형상 복원을 위한 다중시점 영상 디블러링)

  • Choi, Ho Yeol;Park, In Kyu
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.49 no.11
    • /
    • pp.47-55
    • /
    • 2012
  • In this paper, we propose a method to reconstruct accurate 3D shape object by using multi-view images which are disturbed by motion blur. In multi-view deblurring, more precise PSF estimation can be done by using the geometric relationship between multi-view images. The proposed method first estimates initial 2D PSFs from individual input images. Then 3D PSF candidates are projected on the input images one by one to find the best one which are mostly consistent with the initial 2D PSFs. 3D PSF consists with direction and density and it represents the 3D trajectory of object motion. 야to restore 3D shape by using multi-view images computes the similarity map and estimates the position of 3D point. The estimated 3D PSF is again projected to input images and they replaces the intial 2D PSFs which are finally used in image deblurring. Experimental result shows that the quality of image deblurring and 3D reconstruction improves significantly compared with the result when the input images are independently deblurred.

High-resolution 3D Object Reconstruction using Multiple Cameras (다수의 카메라를 활용한 고해상도 3차원 객체 복원 시스템)

  • Hwang, Sung Soo;Yoo, Jisung;Kim, Hee-Dong;Kim, Sujung;Paeng, Kyunghyun;Kim, Seong Dae
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.10
    • /
    • pp.150-161
    • /
    • 2013
  • This paper presents a new system which produces high resolution 3D contents by capturing multiview images of an object using multiple cameras, and estimating geometric and texture information of the object from the captured images. Even though a variety of multiview image-based 3D reconstruction systems have been proposed, it was difficult to generate high resolution 3D contents because multiview image-based 3D reconstruction requires a large amount of memory and computation. In order to reduce computational complexity and memory size for 3D reconstruction, the proposed system predetermines the regions in input images where an object can exist to extract object boundaries fast. And for fast computation of a visual hull, the system represents silhouettes and 3D-2D projection/back-projection relations by chain codes and 1D homographies, respectively. The geometric data of the reconstructed object is compactly represented by a 3D segment-based data format which is called DoCube, and the 3D object is finally reconstructed after 3D mesh generation and texture mapping are performed. Experimental results show that the proposed system produces 3D object contents of $800{\times}800{\times}800$ resolution with a rate of 2.2 seconds per frame.

The Determination of Multiplexing-Diversity Transition Mode in the Visual-MIMO System (Visual-MIMO 시스템에서의 다중화-다이버시티 모드 전환 결정)

  • Kim, Ji-won;Kim, Ki-doo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.41 no.1
    • /
    • pp.42-50
    • /
    • 2016
  • Recently, researches about the communication between LED array and the camera (alias 'Visual-MIMO'), have been actively conducted, and the application to the vehicle and a smart phone is expected to be maximized. Since the bit error may occur if the ISI is severe in a LED array, it is necessary to switch from the multiplexing to the diversity mode. In this paper, according to the use or not of a reference array pattern, a method for determining the transition time to the diversity mode is suggested and verified. When using the reference pattern, it determines the transition time to the diversity mode from the multiplexing using the brightness information of the received image. If the reference array pattern is not used, the size of LED array compared to the entire image according to the distance is used and the size of the LED array at the distance of a severe ISI may be used for the determination of transition time to the diversity. Finally, the proposed method is verified through the simulation and hardware experiments as well as by analyzing the performance in accordance with the ISI level and the distance.

Automatic Mosaicing of Airborne Multispectral Images using GPS/INS Data and Unsupervised Classification (GPS/INS자료와 무감독 분류를 이용한 항공영상 자동 모자이킹)

  • Jang, Jae-Dong
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.9 no.1
    • /
    • pp.46-55
    • /
    • 2006
  • The purpose of this study is a development of an automatic mosaicing for applying to large number of airborne multispectral images, which reduces manual operation by human. 2436 airborne multispectral images were acquired from DuncanTech MS4100 camera with three bands; green, red and near infrared. LIDAR(LIght Detection And Ranging) data and GPS/INS(global positioning system/inertial navigation system) data were collected with the multispectral images. First, the multispectral images were converted to image patterns by unsupervised classification. Their patterns were compared with those of adjacent images to derive relative spatial position between images. Relative spatial positions were derived for 80% of the whole images. Second, it accomplished an automatic mosaicing using GPS/INS data and unsupervised classification. Since the time of GPS/INS data did not synchronized the time of readout images, synchronized GPS/INS data with the time of readout image were selected in consecutive data by comparing unsupervised classified images. This method realized mosaicing automatically for 96% images and RMSE (root mean square error) for the spatial precision of mosaiced images was only 1.44 m by validation with LIDAR data.

  • PDF

Metadata Structure for Object-oriented Personalized Media Service System based on Multiple Cameras (다중 카메라 기반의 객체중심 맞춤형 영상 미디어 시스템을 위한 메타데이터 구조)

  • Byun, Sung-Woo;Lee, So-Min;Lee, Soek-Pil;Kim, Kwang-Yong
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2014.06a
    • /
    • pp.185-187
    • /
    • 2014
  • 디지털 방송 서비스가 널리 보급되고 확대 되면서 사용자는 프로그램 제작자가 만든 순서대로 시청하는 방송이 아닌 개인이 원하는 장면을 원하는 시간대에 시청하는 맞춤형 방송 서비스에 대한 요구가 생겨나게 되었다. 이러한 맞춤형 방송 서비스에서는 프로그램의 정보를 미리 저장하고 사용자가 입력한 데이터를 바탕으로 검색하기 위하기 위한 메타데이터 정보가 중요하다. 기존의 맞춤형 방송 서비스는 사용자가 원하는 장면 별로 시청할 수 있는 반면에 본 연구는 원하는 장면뿐만 아니라 사용자가 보고 싶은 객체를 원하는 카메라 시점에서 시청할 수 있는 객체중심의 맞춤형 영상 미디어 서비스를 위한 메타데이터 구조에 대한 연구이다. 본 연구를 통하여 사용자에게 기존의 맞춤형 방송 서비스에 없었던 객체에 대한 세그먼트 정보를 제공해 줌으로써 시청의 폭을 넓혀 사용자의 방송 시청 만족도를 높일 수 있다.

  • PDF

인공지능 기반 3차원 공간 복원 최신 기술 동향

  • Im, Seong-Hun
    • Broadcasting and Media Magazine
    • /
    • v.25 no.2
    • /
    • pp.17-26
    • /
    • 2020
  • 최근 스마트폰에서의 증강현실, 미적 효과의 증대(예, 라이브 포커싱) 등의 어플리케이션을 제공하기 위해 모바일 기기에서의 3차원 공간 복원 기술에 대한 관심이 증가하고 있다. 소비자들의 요구에 발 맞춰 최근 스마트폰 제조사는 모든 플래그십 모델에 다중 카메라 및 뎁스 센서(거리 측정 센서)를 탑재하는 추세이다. 본 고에서는 모바일 폰에 탑재되고 있는 대표적인 세 축의 뎁스 추정(공간 복원) 방식에 대해 간단히 살펴보고, 최근 심층학습(Deep learning)의 등장으로 기술 발전의 새로운 국면에 접어 든 다중 시점 매칭(Multi-view stereo) 방법에 대해 소개하고자 한다. 심층 신경망이 재조명 받은 2012년 전까지 주류 연구 방향이었던 전통 기하학 기반의 방법에 대한 소개를 시작으로 심층 신경망기반의 방법론으로의 발전된 형태를 살펴본다. 또한, 신경망기반의 방법론은 크게 3 세대로 나누어 각 세대별 특징에 대해 자세히 살펴보고, 다양한 데이터에 대한 실험 결과를 통해 세대별 공간 복원 결과를 비교 분석한다.

View Synthesis Using OpenGL for Multi-viewpoint 3D TV (다시점 3차원 방송을 위한 OpenGL을 이용하는 중간영상 생성)

  • Lee, Hyun-Jung;Hur, Nam-Ho;Seo, Yong-Duek
    • Journal of Broadcast Engineering
    • /
    • v.11 no.4 s.33
    • /
    • pp.507-520
    • /
    • 2006
  • In this paper, we propose an application of OpenGL functions for novel view synthesis from multi-view images and depth maps. While image based rendering has been meant to generate synthetic images by processing the camera view with a graphic engine, little has been known about how to apply the given images and depth information to the graphic engine and render the scene. This paper presents an efficient way of constructing a 3D space with camera parameters, reconstructing the 3D scene with color and depth images, and synthesizing virtual views in real-time as well as their depth images.

Land Cover Mapping and Availability Evaluation Based on Drone Images with Multi-Spectral Camera (다중분광 카메라 탑재 드론 영상 기반 토지피복도 제작 및 활용성 평가)

  • Xu, Chun Xu;Lim, Jae Hyoung;Jin, Xin Mei;Yun, Hee Cheon
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.36 no.6
    • /
    • pp.589-599
    • /
    • 2018
  • The land cover map has been produced by using satellite and aerial images. However, these two images have the limitations in spatial resolution, and it is difficult to acquire images of a area at desired time because of the influence of clouds. In addition, it is costly and time-consuming that mapping land cover map of a small area used by satellite and aerial images. This study used multispectral camera-based drone to acquire multi-temporal images for orthoimages generation. The efficiency of produced land cover map was evaluated using time series analysis. The results indicated that the proposed method can generated RGB orthoimage and multispectral orthoimage with RMSE (Root Mean Square Error) of ${\pm}10mm$, ${\pm}11mm$, ${\pm}26mm$ and ${\pm}28mm$, ${\pm}27mm$, ${\pm}47mm$ on X, Y, H respectively. The accuracy of the pixel-based and object-based land cover map was analyzed and the results showed that the accuracy and Kappa coefficient of object-based classification were higher than that of pixel-based classification, which were 93.75%, 92.42% on July, 92.50%, 91.20% on October, 92.92%, 91.77% on February, respectively. Moreover, the proposed method can accurately capture the quantitative area change of the object. In summary, the suggest study demonstrated the possibility and efficiency of using multispectral camera-based drone in production of land cover map.