• Title/Summary/Keyword: depth image-based

Search Result 822, Processing Time 0.026 seconds

Near-lossless Coding of Multiview Texture and Depth Information for Graphics Applications (그래픽스 응용을 위한 다시점 텍스처 및 깊이 정보의 근접 무손실 부호화)

  • Yoon, Seung-Uk;Ho, Yo-Sung
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.46 no.1
    • /
    • pp.41-48
    • /
    • 2009
  • This Paper introduces representation and coding schemes of multiview texture and depth data for complex three-dimensional scenes. We represent input color and depth images using compressed texture and depth map pairs. The proposed X-codec encodes them further to increase compression ratio in a near-lossless way. Our system resolves two problems. First, rendering time and output visual quality depend on input image resolutions rather than scene complexity since a depth image-based rendering techniques is used. Second, the random access problem of conventional image-based rendering could be effectively solved using our image block-based compression schemes. From experimental results, the proposed approach is useful to graphics applications because it provides multiview rendering, selective decoding, and scene manipulation functionalities.

Improvement of 3D Stereoscopic Perception Using Depth Map Transformation (깊이맵 변환을 이용한 3D 입체감 개선 방법)

  • Jang, Seong-Eun;Jung, Da-Un;Seo, Joo-Ha;Kim, Man-Bae
    • Journal of Broadcast Engineering
    • /
    • v.16 no.6
    • /
    • pp.916-926
    • /
    • 2011
  • It is well known that high-resolution 3D movie contents frequently do not deliver the identical 3D perception to low-resolution 3D images. For solving this problem, we propose a novel method that produces a new stereoscopic image based on depth map transformation using the spatial complexity of an image. After analyzing the depth map histogram, the depth map is decomposed into multiple depth planes that are transformed based upon the spatial complexity. The transformed depth planes are composited into a new depth map. Experimental results demonstrate that the lower the spatial complexity is, the higher the perceived video quality and depth perception are. As well, visual fatigue test showed that the stereoscopic images deliver less visual fatigue.

Real-time Eye Contact System Using a Kinect Depth Camera for Realistic Telepresence (Kinect 깊이 카메라를 이용한 실감 원격 영상회의의 시선 맞춤 시스템)

  • Lee, Sang-Beom;Ho, Yo-Sung
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37 no.4C
    • /
    • pp.277-282
    • /
    • 2012
  • In this paper, we present a real-time eye contact system for realistic telepresence using a Kinect depth camera. In order to generate the eye contact image, we capture a pair of color and depth video. Then, the foreground single user is separated from the background. Since the raw depth data includes several types of noises, we perform a joint bilateral filtering method. We apply the discontinuity-adaptive depth filter to the filtered depth map to reduce the disocclusion area. From the color image and the preprocessed depth map, we construct a user mesh model at the virtual viewpoint. The entire system is implemented through GPU-based parallel programming for real-time processing. Experimental results have shown that the proposed eye contact system is efficient in realizing eye contact, providing the realistic telepresence.

Simplified Integral Imaging Pickup Method for Real Objects Using a Depth Camera

  • Li, Gang;Kwon, Ki-Chul;Shin, Gwan-Ho;Jeong, Ji-Seong;Yoo, Kwan-Hee;Kim, Nam
    • Journal of the Optical Society of Korea
    • /
    • v.16 no.4
    • /
    • pp.381-385
    • /
    • 2012
  • In this paper, we present a novel integral imaging pickup method. We extract each pixel's actual depth data from a real object's surface using a depth camera, then generate elemental images based on the depth map. Since the proposed method generates elemental images without a lens array, it has simplified the pickup process and overcome some disadvantages caused by a conventional optical pickup process using a lens array. As a result, we can display a three-dimensional (3D) image in integral imaging. To show the usefulness of the proposed method, an experiment is presented. Though the pickup process has been simplified in the proposed method, the experimental results reveal that it can also display a full motion parallax image the same as the image reconstructed by the conventional method. In addition, if we improve calculation speed, it will be useful in a real-time integral imaging display system.

On the Study of Initializing Extended Depth of Focus Algorithm Parameters (Extended Depth of Focus 알고리듬 파라메타 초기설정에 관한 연구)

  • Yoo, Kyung-Moo;Joo, Hyo-Nam;Kim, Joon-Seek;Park, Duck-Chun;Choi, In-Ho
    • Journal of Broadcast Engineering
    • /
    • v.17 no.4
    • /
    • pp.625-633
    • /
    • 2012
  • Extended Depth of Focus (EDF) algorithms for extracting three-dimensional (3D) information from a set of optical image slices are studied by many researches recently. Due to the limited depth of focus of the microscope, only a small portion of the image slices are in focus. Most of the EDF algorithms try to find the in-focus area to generate a single focused image and a 3D depth image. Inherent to most image processing algorithms, the EDF algorithms need parameters to be properly initialized to perform successfully. In this paper, we select three popular transform-based EDF algorithms which are each based on pyramid, wavelet transform, and complex wavelet transform, and study the performance of the algorithms according to the initialization of its parameters. The parameters we considered consist of the number of levels used in the transform, the selection of the lowest level image, the window size used in high frequency filter, the noise reduction method, etc. Through extended simulation, we find a good relationship between the initialization of the parameters and the properties of both the texture and 3D ground truth images. Typically, we find that a proper initialization of the parameters improve the algorithm performance 3dB ~ 19dB over a default initialization in recovering the 3D information.

Smoke Detection Based on RGB-Depth Camera in Interior (RGB-Depth 카메라 기반의 실내 연기검출)

  • Park, Jang-Sik
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.9 no.2
    • /
    • pp.155-160
    • /
    • 2014
  • In this paper, an algorithm using RGB-depth camera is proposed to detect smoke in interrior. RGB-depth camera, the Kinect provides RGB color image and depth information. The Kinect sensor consists of an infra-red laser emitter, infra-red camera and an RGB camera. A specific pattern of speckles radiated from the laser source is projected onto the scene. This pattern is captured by the infra-red camera and is analyzed to get depth information. The distance of each speckle of the specific pattern is measured and the depth of object is estimated. As the depth of object is highly changed, the depth of object plain can not be determined by the Kinect. The depth of smoke can not be determined too because the density of smoke is changed with constant frequency and intensity of infra-red image is varied between each pixels. In this paper, a smoke detection algorithm using characteristics of the Kinect is proposed. The region that the depth information is not determined sets the candidate region of smoke. If the intensity of the candidate region of color image is larger than a threshold, the region is confirmed as smoke region. As results of simulations, it is shown that the proposed method is effective to detect smoke in interior.

View Synthesis and Coding of Multi-view Data in Arbitrary Camera Arrangements Using Multiple Layered Depth Images

  • Yoon, Seung-Uk;Ho, Yo-Sung
    • Journal of Multimedia Information System
    • /
    • v.1 no.1
    • /
    • pp.1-10
    • /
    • 2014
  • In this paper, we propose a new view synthesis technique for coding of multi-view color and depth data in arbitrary camera arrangements. We treat each camera position as a 3-D point in world coordinates and build clusters of those vertices. Color and depth data within a cluster are gathered into one camera position using a hierarchical representation based on the concept of layered depth image (LDI). Since one camera can cover only a limited viewing range, we set multiple reference cameras so that multiple LDIs are generated to cover the whole viewing range. Therefore, we can enhance the visual quality of the reconstructed views from multiple LDIs comparing with that from a single LDI. From experimental results, the proposed scheme shows better coding performance under arbitrary camera configurations in terms of PSNR and subjective visual quality.

  • PDF

Direct Depth and Color-based Environment Modeling and Mobile Robot Navigation (스테레오 비전 센서의 깊이 및 색상 정보를 이용한 환경 모델링 기반의 이동로봇 주행기술)

  • Park, Soon-Yong;Park, Mignon;Park, Sung-Kee
    • The Journal of Korea Robotics Society
    • /
    • v.3 no.3
    • /
    • pp.194-202
    • /
    • 2008
  • This paper describes a new method for indoor environment mapping and localization with stereo camera. For environmental modeling, we directly use the depth and color information in image pixels as visual features. Furthermore, only the depth and color information at horizontal centerline in image is used, where optical axis passes through. The usefulness of this method is that we can easily build a measure between modeling and sensing data only on the horizontal centerline. That is because vertical working volume between model and sensing data can be changed according to robot motion. Therefore, we can build a map about indoor environment as compact and efficient representation. Also, based on such nodes and sensing data, we suggest a method for estimating mobile robot positioning with random sampling stochastic algorithm. With basic real experiments, we show that the proposed method can be an effective visual navigation algorithm.

  • PDF

A Novel Segment Extraction and Stereo Matching Technique using Color, Motion and Initial Depth from Depth Camera (컬러, 움직임 정보 및 깊이 카메라 초기 깊이를 이용한 분할 영역 추출 및 스테레오 정합 기법)

  • Um, Gi-Mun;Park, Ji-Min;Bang, Gun;Cheong, Won-Sik;Hur, Nam-Ho;Kim, Jin-Woong
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.12C
    • /
    • pp.1147-1153
    • /
    • 2009
  • We propose a novel image segmentation and segment-based stereo matching technique using color, depth, and motion information. Proposed technique firstly splits reference images into foreground region or background region using depth information from depth camera. Then each region is segmented into small segments with color information. Moreover, extracted segments in current frame are tracked in the next frame in order to maintain depth consistency between frames. The initial depth from the depth camera is also used to set the depth search range for stereo matching. Proposed segment-based stereo matching technique was compared with conventional one without foreground and background separation and other conventional one without motion tracking of segments. Simulation results showed that the improvement of segment extraction and depth estimation consistencies by proposed technique compared to conventional ones especially at the static background region.

A Study on AI Softwear [Stable Diffusion] ControlNet plug-in Usabilities

  • Chenghao Wang;Jeanhun Chung
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.15 no.4
    • /
    • pp.166-171
    • /
    • 2023
  • With significant advancements in the field of artificial intelligence, many novel algorithms and technologies have emerged. Currently, AI painting can generate high-quality images based on textual descriptions. However, it is often challenging to control details when generating images, even with complex textual inputs. Therefore, there is a need to implement additional control mechanisms beyond textual descriptions. Based on ControlNet, this passage describes a combined utilization of various local controls (such as edge maps and depth maps) and global control within a single model. It provides a comprehensive exposition of the fundamental concepts of ControlNet, elucidating its theoretical foundation and relevant technological features. Furthermore, combining methods and applications, understanding the technical characteristics involves analyzing distinct advantages and image differences. This further explores insights into the development of image generation patterns.