• Title/Summary/Keyword: depth-map estimation

Search Result 100, Processing Time 0.023 seconds

Fractal Depth Map Sequence Coding Algorithm with Motion-vector-field-based Motion Estimation

  • Zhu, Shiping;Zhao, Dongyu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.1
    • /
    • pp.242-259
    • /
    • 2015
  • Three-dimensional video coding is one of the main challenges restricting the widespread applications of 3D video and free viewpoint video. In this paper, a novel fractal coding algorithm with motion-vector-field-based motion estimation for depth map sequence is proposed. We firstly add pre-search restriction to rule the improper domain blocks out of the matching search process so that the number of blocks involved in the search process can be restricted to a smaller size. Some improvements for motion estimation including initial search point prediction, threshold transition condition and early termination condition are made based on the feature of fractal coding. The motion-vector-field-based adaptive hexagon search algorithm on the basis of center-biased distribution characteristics of depth motion vector is proposed to accelerate the search. Experimental results show that the proposed algorithm can reach optimum levels of quality and save the coding time. The PSNR of synthesized view is increased by 0.56 dB with 36.97% bit rate decrease on average compared with H.264 Full Search. And the depth encoding time is saved by up to 66.47%. Moreover, the proposed fractal depth map sequence codec outperforms the recent alternative codecs by improving the H.264/AVC, especially in much bitrate saving and encoding time reduction.

Depth Map Refinement using Segment Plane Estimation (세그멘트 평면 추정을 이용한 깊이 지도 개선)

  • Jung, Woo-Kyung;Han, Jong-Ki
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2020.07a
    • /
    • pp.286-287
    • /
    • 2020
  • Depth map is the most common way of expressing 3D space in immersive media. In this paper, we propose a post-processing method to improve the quality of depth map. In proposed method, a depth map is divided into segments, and the plane of each segment estimated using RANSAC. In order to increase the accuracy of the RANSAC process, we apply matching reliability of each pixel in depth map as a weighting factor.

  • PDF

Confidence Measure of Depth Map for Outdoor RGB+D Database (야외 RGB+D 데이터베이스 구축을 위한 깊이 영상 신뢰도 측정 기법)

  • Park, Jaekwang;Kim, Sunok;Sohn, Kwanghoon;Min, Dongbo
    • Journal of Korea Multimedia Society
    • /
    • v.19 no.9
    • /
    • pp.1647-1658
    • /
    • 2016
  • RGB+D database has been widely used in object recognition, object tracking, robot control, to name a few. While rapid advance of active depth sensing technologies allows for the widespread of indoor RGB+D databases, there are only few outdoor RGB+D databases largely due to an inherent limitation of active depth cameras. In this paper, we propose a novel method used to build outdoor RGB+D databases. Instead of using active depth cameras such as Kinect or LIDAR, we acquire a pair of stereo image using high-resolution stereo camera and then obtain a depth map by applying stereo matching algorithm. To deal with estimation errors that inevitably exist in the depth map obtained from stereo matching methods, we develop an approach that estimates confidence of depth maps based on unsupervised learning. Unlike existing confidence estimation approaches, we explicitly consider a spatial correlation that may exist in the confidence map. Specifically, we focus on refining confidence feature with the assumption that the confidence feature and resultant confidence map are smoothly-varying in spatial domain and are highly correlated to each other. Experimental result shows that the proposed method outperforms existing confidence measure based approaches in various benchmark dataset.

Restoration of underwater images using depth and transmission map estimation, with attenuation priors

  • Jarina, Raihan A.;Abas, P.G. Emeroylariffion;De Silva, Liyanage C.
    • Ocean Systems Engineering
    • /
    • v.11 no.4
    • /
    • pp.331-351
    • /
    • 2021
  • Underwater images are very much different from images taken on land, due to the presence of a higher disturbance ratio caused by the presence of water medium between the camera and the target object. These distortions and noises result in unclear details and reduced quality of the output image. An underwater image restoration method is proposed in this paper, which uses blurriness information, background light neutralization information, and red-light intensity to estimate depth. The transmission map is then estimated using the derived depth map, by considering separate attenuation coefficients for direct and backscattered signals. The estimated transmission map and estimated background light are then used to recover the scene radiance. Qualitative and quantitative analysis have been used to compare the performance of the proposed method against other state-of-the-art restoration methods. It has been shown that the proposed method can yield good quality restored underwater images. The proposed method has also been evaluated using different qualitative metrics, and results have shown that method is highly capable of restoring underwater images with different conditions. The results are significant and show the applicability of the proposed method for underwater image restoration work.

Deep Learning-based Depth Map Estimation: A Review

  • Abdullah, Jan;Safran, Khan;Suyoung, Seo
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.1
    • /
    • pp.1-21
    • /
    • 2023
  • In this technically advanced era, we are surrounded by smartphones, computers, and cameras, which help us to store visual information in 2D image planes. However, such images lack 3D spatial information about the scene, which is very useful for scientists, surveyors, engineers, and even robots. To tackle such problems, depth maps are generated for respective image planes. Depth maps or depth images are single image metric which carries the information in three-dimensional axes, i.e., xyz coordinates, where z is the object's distance from camera axes. For many applications, including augmented reality, object tracking, segmentation, scene reconstruction, distance measurement, autonomous navigation, and autonomous driving, depth estimation is a fundamental task. Much of the work has been done to calculate depth maps. We reviewed the status of depth map estimation using different techniques from several papers, study areas, and models applied over the last 20 years. We surveyed different depth-mapping techniques based on traditional ways and newly developed deep-learning methods. The primary purpose of this study is to present a detailed review of the state-of-the-art traditional depth mapping techniques and recent deep learning methodologies. This study encompasses the critical points of each method from different perspectives, like datasets, procedures performed, types of algorithms, loss functions, and well-known evaluation metrics. Similarly, this paper also discusses the subdomains in each method, like supervised, unsupervised, and semi-supervised methods. We also elaborate on the challenges of different methods. At the conclusion of this study, we discussed new ideas for future research and studies in depth map research.

Segmentation-Based Depth Map Adjustment for Improved Grasping Pose Detection (물체 파지점 검출 향상을 위한 분할 기반 깊이 지도 조정)

  • Hyunsoo Shin;Muhammad Raheel Afzal;Sungon Lee
    • The Journal of Korea Robotics Society
    • /
    • v.19 no.1
    • /
    • pp.16-22
    • /
    • 2024
  • Robotic grasping in unstructured environments poses a significant challenge, demanding precise estimation of gripping positions for diverse and unknown objects. Generative Grasping Convolution Neural Network (GG-CNN) can estimate the position and direction that can be gripped by a robot gripper for an unknown object based on a three-dimensional depth map. Since GG-CNN uses only a depth map as an input, the precision of the depth map is the most critical factor affecting the result. To address the challenge of depth map precision, we integrate the Segment Anything Model renowned for its robust zero-shot performance across various segmentation tasks. We adjust the components corresponding to the segmented areas in the depth map aligned through external calibration. The proposed method was validated on the Cornell dataset and SurgicalKit dataset. Quantitative analysis compared to existing methods showed a 49.8% improvement with the dataset including surgical instruments. The results highlight the practical importance of our approach, especially in scenarios involving thin and metallic objects.

A Depth Estimation Using Infocused and Defocused Images (인포커스 및 디포커스 영상으로부터 깊이맵 생성)

  • Mahmoudpour, Seed;Kim, Manbae
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2013.11a
    • /
    • pp.114-115
    • /
    • 2013
  • The blur amount of an image changes proportional to scene depth. Depth from Defocus (DFD) is an approach in which a depth map can be obtained using blur amount calculation. In this paper, a novel DFD method is proposed in which depth is measured using an infocused and a defocused image. Subbaro's algorithm is used as a preliminary depth estimation method and edge blur estimation is provided to overcome drawbacks in edge.

  • PDF

Depth Map Completion using Nearest Neighbor Kernel (최근접 이웃 커널을 이용한 깊이 영상 완성 기술)

  • Taehyun, Jeong;Kutub, Uddin;Byung Tae, Oh
    • Journal of Broadcast Engineering
    • /
    • v.27 no.6
    • /
    • pp.906-913
    • /
    • 2022
  • In this paper, we propose a new deep network architecture using nearest neighbor kernel for the estimation of dense depth map from its sparse map and corresponding color information. First, we propose to decompose the depth map signal into the structure and details for easier prediction. We then propose two separate subnetworks for prediction of both structure and details using classification and regression approaches, respectively. Moreover, the nearest neighboring kernel method has been newly proposed for accurate prediction of structure signal. As a result, the proposed method showed better results than other methods quantitatively and qualitatively.

Toward Occlusion-Free Depth Estimation for Video Production

  • Park, Jong-Il;Seiki-Inoue
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1997.06a
    • /
    • pp.131-136
    • /
    • 1997
  • We present a method to estimate a dense and sharp depth map using multiple cameras for the application to flexible video production. A key issue for obtaining sharp depth map is how to overcome the harmful influence of occlusion. Thus, we first propose to selectively use the depth information from multiple cameras. With a simple sort and discard technique, we resolve the occlusion problem considerably at a slight sacrifice of noise tolerance. However, boundary overreach of more textured area to less textured area at object boundaries still remains to be solved. We observed that the amount of boundary overreach is less than half the size of the matching window and, unlike usual stereo matching, the boundary overreach with the proposed occlusion-overcoming method shows very abrupt transition. Based on these observations, we propose a hierarchical estimation scheme that attempts to reduce boundary overreach such that edges of the depth map coincide with object boundaries on the one hand, and to reduce noisy estimates due to insufficient size of matching window on the other hand. We show the hierarchical method can produce a sharp depth map for a variety of images.

  • PDF

High Resolution Depth-map Estimation in Real-time using Efficient Multi-threading (효율적인 멀티 쓰레딩을 이용한 고해상도 깊이지도의 실시간 획득)

  • Cho, Chil-Suk;Jun, Ji-In;Choo, Hyon-Gon;Park, Jong-Il
    • Journal of Broadcast Engineering
    • /
    • v.17 no.6
    • /
    • pp.945-953
    • /
    • 2012
  • A depth map can be obtained by projecting/capturing patterns of stripes using a projector-camera system and analyzing the geometric relationship between the projected patterns and the captured patterns. This is usually called structured light technique. In this paper, we propose a new multi-threading scheme for accelerating a conventional structured light technique. On CPUs and GPUs, multi-threading can be implemented by using OpenMP and CUDA, respectively. However, the problem is that their performance changes according to the computational conditions of partial processes of a structured light technique. In other words, OpenMP (using multiple CPUs) outperformed CUDA (using multiple GPUs) in partial processes such as pattern decoding and depth estimation. In contrast, CUDA outperformed OpenMP in partial processes such as rectification and pattern segmentation. Therefore, we carefully analyze the computational conditions where each outperforms the other and do use the better one in the related conditions. As a result, the proposed method can estimate a depth map in a speed of over 25 fps on $1280{\times}800$ images.