• Title/Summary/Keyword: Multiple 2D Images

Search Result 160, Processing Time 0.031 seconds

2D Barcode Detection Algorithm with Multiple Features Combination for a Long Distance Search

  • Pak, Myeongsuk;Kim, Sanghoon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2015.10a
    • /
    • pp.1506-1508
    • /
    • 2015
  • A 2D barcode region localization system for the automatic inspection of a long distance logistics objects has been developed. For the successful 2D barcode localization, variance and frequency of the pixel distribution within average 2D barcodes is modeled and the average model of 2D barcode is combined with the corner features detection to localize the final 2D barcode candidates. An automatic 2D barcode localization software was developed with the multiple features mixture method and we tested our system on real camera images of several popular 2D barcode symbologies.

Registration of 3D CT Data to 2D Endoscopic Image using a Gradient Mutual Information based Viewpoint Matching for Image-Guided Medialization Laryngoplasty

  • Yim, Yeny;Wakid, Mike;Kirmizibayrak, Can;Bielamowicz, Steven;Hahn, James
    • Journal of Computing Science and Engineering
    • /
    • v.4 no.4
    • /
    • pp.368-387
    • /
    • 2010
  • We propose a novel method for the registration of 3D CT scans to 2D endoscopic images during the image-guided medialization laryngoplasty. This study aims to allow the surgeon to find the precise configuration of the implant and place it into the desired location by employing accurate registration methods of the 3D CT data to intra-operative patient and interactive visualization tools for the registered images. In this study, the proposed registration methods enable the surgeon to compare the outcome of the procedure to the pre-planned shape by matching the vocal folds in the CT rendered images to the endoscopic images. The 3D image fusion provides an interactive and intuitive guidance for surgeon by visualizing a combined and correlated relationship of the multiple imaging modalities. The 3D Magic Lens helps to effectively visualize laryngeal anatomical structures by applying different transparencies and transfer functions to the region of interest. The preliminary results of the study demonstrated that the proposed method can be readily extended for image-guided surgery of real patients.

Deep Learning Machine Vision System with High Object Recognition Rate using Multiple-Exposure Image Sensing Method

  • Park, Min-Jun;Kim, Hyeon-June
    • Journal of Sensor Science and Technology
    • /
    • v.30 no.2
    • /
    • pp.76-81
    • /
    • 2021
  • In this study, we propose a machine vision system with a high object recognition rate. By utilizing a multiple-exposure image sensing technique, the proposed deep learning-based machine vision system can cover a wide light intensity range without further learning processes on the various light intensity range. If the proposed machine vision system fails to recognize object features, the system operates in a multiple-exposure sensing mode and detects the target object that is blocked in the near dark or bright region. Furthermore, short- and long-exposure images from the multiple-exposure sensing mode are synthesized to obtain accurate object feature information. That results in the generation of a wide dynamic range of image information. Even with the object recognition resources for the deep learning process with a light intensity range of only 23 dB, the prototype machine vision system with the multiple-exposure imaging method demonstrated an object recognition performance with a light intensity range of up to 96 dB.

Effects of Depth Map Quantization for Computer-Generated Multiview Images using Depth Image-Based Rendering

  • Kim, Min-Young;Cho, Yong-Joo;Choo, Hyon-Gon;Kim, Jin-Woong;Park, Kyoung-Shin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.5 no.11
    • /
    • pp.2175-2190
    • /
    • 2011
  • This paper presents the effects of depth map quantization for multiview intermediate image generation using depth image-based rendering (DIBR). DIBR synthesizes multiple virtual views of a 3D scene from a 2D image and its associated depth map. However, it needs precise depth information in order to generate reliable and accurate intermediate view images for use in multiview 3D display systems. Previous work has extensively studied the pre-processing of the depth map, but little is known about depth map quantization. In this paper, we conduct an experiment to estimate the depth map quantization that affords acceptable image quality to generate DIBR-based multiview intermediate images. The experiment uses computer-generated 3D scenes, in which the multiview images captured directly from the scene are compared to the multiview intermediate images constructed by DIBR with a number of quantized depth maps. The results showed that there was no significant effect on depth map quantization from 16-bit to 7-bit (and more specifically 96-scale) on DIBR. Hence, a depth map above 7-bit is needed to maintain sufficient image quality for a DIBR-based multiview 3D system.

A Study On Three-dimensional Optimized Face Recognition Model : Comparative Studies and Analysis of Model Architectures (3차원 얼굴인식 모델에 관한 연구: 모델 구조 비교연구 및 해석)

  • Park, Chan-Jun;Oh, Sung-Kwun;Kim, Jin-Yul
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.64 no.6
    • /
    • pp.900-911
    • /
    • 2015
  • In this paper, 3D face recognition model is designed by using Polynomial based RBFNN(Radial Basis Function Neural Network) and PNN(Polynomial Neural Network). Also recognition rate is performed by this model. In existing 2D face recognition model, the degradation of recognition rate may occur in external environments such as face features using a brightness of the video. So 3D face recognition is performed by using 3D scanner for improving disadvantage of 2D face recognition. In the preprocessing part, obtained 3D face images for the variation of each pose are changed as front image by using pose compensation. The depth data of face image shape is extracted by using Multiple point signature. And whole area of face depth information is obtained by using the tip of a nose as a reference point. Parameter optimization is carried out with the aid of both ABC(Artificial Bee Colony) and PSO(Particle Swarm Optimization) for effective training and recognition. Experimental data for face recognition is built up by the face images of students and researchers in IC&CI Lab of Suwon University. By using the images of 3D face extracted in IC&CI Lab. the performance of 3D face recognition is evaluated and compared according to two types of models as well as point signature method based on two kinds of depth data information.

Convenient View Calibration of Multiple RGB-D Cameras Using a Spherical Object (구형 물체를 이용한 다중 RGB-D 카메라의 간편한 시점보정)

  • Park, Soon-Yong;Choi, Sung-In
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.3 no.8
    • /
    • pp.309-314
    • /
    • 2014
  • To generate a complete 3D model from depth images of multiple RGB-D cameras, it is necessary to find 3D transformations between RGB-D cameras. This paper proposes a convenient view calibration technique using a spherical object. Conventional view calibration methods use either planar checkerboards or 3D objects with coded-pattern. In these conventional methods, detection and matching of pattern features and codes takes a significant time. In this paper, we propose a convenient view calibration method using both 3D depth and 2D texture images of a spherical object simultaneously. First, while moving the spherical object freely in the modeling space, depth and texture images of the object are acquired from all RGB-D camera simultaneously. Then, the external parameters of each RGB-D camera is calibrated so that the coordinates of the sphere center coincide in the world coordinate system.

Implementation of 3D Structure Reconstruction System Using Geometric Primitives (원시기하도형을 이용한 3차원구조 복원시스템의 구현)

  • 남현석;구본기;진성일
    • Proceedings of the IEEK Conference
    • /
    • 2003.11a
    • /
    • pp.237-240
    • /
    • 2003
  • We implement a system for 3D structure reconstruction from multiple 2D images. It uses geometric primitives such as box, wedge, pyramid, etc, each having translation, rotation, and scale parameters. Primitives are marked on input images with GUI (Graphic User Interface). Lines made by projection of primitives onto an image correspond to marked line segments of the image. Error function is defined by disparity between them and is minimized by downhill simplex method. By assigning relationship between models, the number of parameters to solve can be decreased and the resultant models become more accurate To share variables among other models also reduces computational complexity. Experiments using real images have shown that the proposed method successfully reconstructs 3D structure.

  • PDF

Recovering the Colors of Objects from Multiple Near-IR Images

  • Kim, Ari;Oh, In-Hoo;Kim, Hong-Suk;Park, Seung-Ok;Park, Youngsik
    • Journal of the Optical Society of Korea
    • /
    • v.19 no.1
    • /
    • pp.102-111
    • /
    • 2015
  • This paper proposes an algorithm for recovering the colors of objects from multiple near-infrared (near-IR) images. The International Commission on Illumination (CIE) color coordinates of objects are recovered from a series of gray images captured under multiple spectral near-IR illuminations using polynomial regression. The feasibility of the proposed algorithm is tested experimentally by using 24 color patches of the Color Rendition Chart. The experimental apparatus is composed of a monochrome digital camera without an IR cut-off filter and a custom-designed LED illuminator emitting multiple spectral near-IR illuminations, with peak wavelengths near the red edge of the visible band, namely at 700, 740, 780, and 860 nm. The average color difference between the original and the recovered colors for all 24 patches was found to be 11.1. However, if some particular patches with high value are disregarded, the average color difference is reduced to 4.2, and this value is within the acceptability tolerance for complex image on the display.

A Study on ISAR Imaging Algorithm for Radar Target Recognition (표적 구분을 위한 ISAR 영상 기법에 대한 연구)

  • Park, Jong-Il;Kim, Kyung-Tae
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.19 no.3
    • /
    • pp.294-303
    • /
    • 2008
  • ISAR(Inverse Synthetic Aperture Radar) images represent the 2-D(two-dimensional) spatial distribution of RCS (Radar Cross Section) of an object, and they can be applied to the problem of target identification. A traditional approach to ISAR imaging is to use a 2-D IFFT(Inverse Fast Fourier Transform). However, the 2-D IFFT results in low resolution ISAR images especially when the measured frequency bandwidth and angular region are limited. In order to improve the resolution capability of the Fourier transform, various high-resolution spectral estimation approaches have been applied to obtain ISAR images, such as AR(Auto Regressive), MUSIC(Multiple Signal Classification) or Modified MUSIC algorithms. In this study, these high-resolution spectral estimators as well as 2-D IFFT approach are combined with a recently developed ISAR image classification algorithm, and their performances are carefully analyzed and compared in the framework of radar target recognition.

On Shape Recovery of 3D Object from Multiple Range Images (시점이 다른 다수의 거리 영상으로부터 3차원 물체의 형상 복원)

  • Kim, Jun-Young;Yun, Il-Dong;Lee, Sang-Uk
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.37 no.1
    • /
    • pp.1-15
    • /
    • 2000
  • To reconstruct 3- D shape, It is a common strategy to acquire multiple range Images from different viewpoints and integrate them into a common coordinates In this paper, we particularly focus on the registration and integration processes for combining all range Images into one surface model. For the registration, we propose the 2-step registration algorithm, which consists of 2 steps the rough registration step using all data points and the fine registration step using the high-curved data points For the integration, we propose a new algorithm, referred to as ‘multi-registration’ technique, to alleviate the error accumulation problem, which occurs during applying the pair-wise registration to each range image sequentially, in order to transform them into a common reference frame Intensive experiments are performed on the various real range data In experiments, all range images were registered within 1 minutes on Pentium 150MHz PC The results show that the proposed algorithms registrate and integrate multiple range Images within a tolerable error bound in a reasonable computation time, and the total error between all range Images are equalized with our proposed algorithms.

  • PDF