• Title/Summary/Keyword: RGBD

Search Result 19, Processing Time 0.027 seconds

RGBD Panoramic Image Generation Using Frechet Distance Loss Function (프레쳇 거리 손실함수를 이용한 RGBD 파노라마 영상 생성)

  • Kim, Soo Jie;Park, In Kyu
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2022.06a
    • /
    • pp.1229-1231
    • /
    • 2022
  • RGBD 영상은 다양한 3 차원 비전 연구에서 유용하게 사용되며 고품질 RGBD 영상을 취득하기 위한 많은 연구들이 수행되었다. 기존의 영상 생성 연구들은 주로 좁은 FoV(Field of View) 영상을 사용하여서 전체 장면 중 상당 부분이 소실된 영상에 대한 정보를 생성한다. 본 논문에서는 기존의 좁은 FoV 영상으로부터 360 도 전방향 RGBD 영상을 생성하는 기법을 제안한다. 오버랩 되지 않는 4 장의 소수 영상으로부터 전체 파노라마 영상에 대해서 상대적인 FoV 를 추정하고, 360 도 RGBD 영상을 동시에 생성하는 적대적 생성 신경망 기반의 영상 생성 네트워크이다. 360 도 영상의 특징을 반영하도록 설계하여서 개선된 성능을 보인다.

  • PDF

Transformations and Their Analysis from a RGBD Image to Elemental Image Array for 3D Integral Imaging and Coding

  • Yoo, Hoon
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.5
    • /
    • pp.2273-2286
    • /
    • 2018
  • This paper describes transformations between elemental image arrays and a RGBD image for three-dimensional integral imaging and transmitting systems. Two transformations are introduced and analyzed in the proposed method. Normally, a RGBD image is utilized in efficient 3D data transmission although 3D imaging and display is restricted. Thus, a pixel-to-pixel mapping is required to obtain an elemental image array from a RGBD image. However, transformations and their analysis have little attention in computational integral imaging and transmission. Thus, in this paper, we introduce two different mapping methods that are called as the forward and backward mapping methods. Also, two mappings are analyzed and compared in terms of complexity and visual quality. In addition, a special condition, named as the hole-free condition in this paper, is proposed to understand the methods analytically. To verify our analysis, we carry out experiments for test images and the results indicate that the proposed methods and their analysis work in terms of the computational cost and visual quality.

Indoor object detection method using a RGBD image (RGBD 카메라를 이용한 실내에서의 물체 검출 알고리즘)

  • Heo, Seon;Lee, Sang Hwa;Kim, Myung Sik;Han, Seung Beom;Cho, Nam Ik
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2015.11a
    • /
    • pp.100-103
    • /
    • 2015
  • 본 논문에서는 실내에서 RGBD 영상을 이용하여 물체를 검출하는 방법을 제안한다. 특정 물체가 아닌 일반적인 여러 가지 물체에 대한 특징을 규정하기 어려우므로 본 논문에서는 영상 정보에 의존하기 보다 물체와 픽셀의 기하학적 구조에 기반하여 물체를 검출한다. 우선 컬러 정보를 이용하여 대략적인 영상 영역분할을 하고 이를 같은 레이블로 분류하여 물체와 배경의 후보를 얻는다. 대체로 실내 환경에서 바닥은 평면이라 가정할 수 있으므로 바닥의 평면 모델을 만들어서 물체 후보에서 이를 제외시킨다. 또한, 물체에 대한 간단한 가정을 통해 바닥 이외의 배경 역시 물체와 구분하여서 물체 후보들을 가려낸다. 최종적으로 3 차원 공간에서 가까이 위치하는 레이블을 하나로 통합하는 과정을 통해 최종적인 물체 영역을 검출하고 이를 bounding box 로 표시한다. 직접 촬영한 몇몇 실내 RGBD 영상에서 실험한 결과, 제안하는 방법이 기존 방법들에 비해 물체 검출 성능이 좋은 것을 확인하였다.

  • PDF

Object Detection with LiDAR Point Cloud and RGBD Synthesis Using GNN

  • Jung, Tae-Won;Jeong, Chi-Seo;Lee, Jong-Yong;Jung, Kye-Dong
    • International journal of advanced smart convergence
    • /
    • v.9 no.3
    • /
    • pp.192-198
    • /
    • 2020
  • The 3D point cloud is a key technology of object detection for virtual reality and augmented reality. In order to apply various areas of object detection, it is necessary to obtain 3D information and even color information more easily. In general, to generate a 3D point cloud, it is acquired using an expensive scanner device. However, 3D and characteristic information such as RGB and depth can be easily obtained in a mobile device. GNN (Graph Neural Network) can be used for object detection based on these characteristics. In this paper, we have generated RGB and RGBD by detecting basic information and characteristic information from the KITTI dataset, which is often used in 3D point cloud object detection. We have generated RGB-GNN with i-GNN, which is the most widely used LiDAR characteristic information, and color information characteristics that can be obtained from mobile devices. We compared and analyzed object detection accuracy using RGBD-GNN, which characterizes color and depth information.

360 RGBD Image Synthesis from a Sparse Set of Images with Narrow Field-of-View (소수의 협소화각 RGBD 영상으로부터 360 RGBD 영상 합성)

  • Kim, Soojie;Park, In Kyu
    • Journal of Broadcast Engineering
    • /
    • v.27 no.4
    • /
    • pp.487-498
    • /
    • 2022
  • Depth map is an image that contains distance information in 3D space on a 2D plane and is used in various 3D vision tasks. Many existing depth estimation studies mainly use narrow FoV images, in which a significant portion of the entire scene is lost. In this paper, we propose a technique for generating 360° omnidirectional RGBD images from a sparse set of narrow FoV images. The proposed generative adversarial network based image generation model estimates the relative FoV for the entire panoramic image from a small number of non-overlapping images and produces a 360° RGB and depth image simultaneously. In addition, it shows improved performance by configuring a network reflecting the spherical characteristics of the 360° image.

A Study on Hand Gesture Classification Deep learning method device based on RGBD Image (RGBD 이미지 기반 핸드제스처 분류 딥러닝 기법의 연구)

  • Park, Jong-Chan;Li, Yan;Shin, Byeong-Seok
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2019.10a
    • /
    • pp.1173-1175
    • /
    • 2019
  • 소음이 심하거나 긴급한 상황 등에서 서로 다른 핸드제스처에 대한 인식을 컴퓨터의 입력으로 받고 이를 특정 명령으로 인식하는 등의 연구가 로봇 분야에서 연구되고 있다. 그러나 핸드제스처에 대한 전처리 과정에서 RGB데이터를 활용하거나 또는 스켈레톤을 활용하는 연구들이 다양하게 연구되었지만, 실생활에서의 노이즈가 많아 분류 정확도가 높지 않거나 컴퓨팅 파워의 사용이 과다한 문제가 발생했다. 본 논문에서는 RGBD 이미지를 사용하여 Hand Gesture를 트레이닝 받은 Keras 모델을 통해 입력받은 Hand Gesture을 분류하는 연구를 진행하였다. Depth Camera를 통하여 입력받은 Hand Gesture Raw-Data를 Image로 재구성하여 딥러닝을 진행하였다.

Real-time user motion generation and correction using RGBD sensor (RGBD 센서를 이용한 실시간 사용자 동작 생성 및 보정)

  • Gu, Tae Hong;Kim, Un Mi;Kim, Jong Min;Kwon, Tae Soo
    • Journal of the Korea Computer Graphics Society
    • /
    • v.23 no.5
    • /
    • pp.67-73
    • /
    • 2017
  • We propose several techniques which can be employed in a 3D fitness program for monitoring and correcting user's posture. To implement a 3D fitness program, improved reference motion generating techniques and visualizing techniques are necessary. First, in order to understand the difference between the user and the reference movement of a professional, a retargeting method between two different body shapes are studied. Second, the problem of self-occlusion, which occurs when using a low-cost depth sensor to represent complex motions, is solved by using a sample database and time consistency. The system proposed in this paper evaluates the user's posture considering the physical characteristics of the user, and then provides feedback to the user.

Fall Detection Based on 2-Stacked Bi-LSTM and Human-Skeleton Keypoints of RGBD Camera (RGBD 카메라 기반의 Human-Skeleton Keypoints와 2-Stacked Bi-LSTM 모델을 이용한 낙상 탐지)

  • Shin, Byung Geun;Kim, Uung Ho;Lee, Sang Woo;Yang, Jae Young;Kim, Wongyum
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.10 no.11
    • /
    • pp.491-500
    • /
    • 2021
  • In this study, we propose a method for detecting fall behavior using MS Kinect v2 RGBD Camera-based Human-Skeleton Keypoints and a 2-Stacked Bi-LSTM model. In previous studies, skeletal information was extracted from RGB images using a deep learning model such as OpenPose, and then recognition was performed using a recurrent neural network model such as LSTM and GRU. The proposed method receives skeletal information directly from the camera, extracts 2 time-series features of acceleration and distance, and then recognizes the fall behavior using the 2-Stacked Bi-LSTM model. The central joint was obtained for the major skeletons such as the shoulder, spine, and pelvis, and the movement acceleration and distance from the floor were proposed as features of the central joint. The extracted features were compared with models such as Stacked LSTM and Bi-LSTM, and improved detection performance compared to existing studies such as GRU and LSTM was demonstrated through experiments.

Stencil-based 3D facial relief creation from RGBD images for 3D printing

  • Jung, Soonchul;Choi, Yoon-Seok;Kim, Jin-Seo
    • ETRI Journal
    • /
    • v.42 no.2
    • /
    • pp.272-281
    • /
    • 2020
  • Three-dimensional (3D) selfie services, one of the major 3D printing services, print 3D models of an individual's face via scanning. However, most of these services require expensive full-color supporting 3D printers. The high cost of such printers poses a challenge in launching a variety of 3D printing application services. This paper presents a stencil-based 3D facial relief creation method employing a low-cost RGBD sensor and a 3D printer. Stencil-based 3D facial relief is an artwork in which some parts are holes, similar to that in a stencil, and other parts stand out, as in a relief. The proposed method creates a new type of relief by combining the existing stencil techniques and relief techniques. As a result, the 3D printed product resembles a two-colored object rather than a one-colored object even when a monochrome 3D printer is used. Unlike existing personalization-based 3D printing services, the proposed method enables the printing and delivery of products to customers in a short period of time. Experimental results reveal that, compared to existing 3D selfie products printed by monochrome 3D printers, our products have a higher degree of similarity and are more profitable.

Semantic Segmentation of Indoor Scenes Using Depth Superpixel (깊이 슈퍼 픽셀을 이용한 실내 장면의 의미론적 분할 방법)

  • Kim, Seon-Keol;Kang, Hang-Bong
    • Journal of Korea Multimedia Society
    • /
    • v.19 no.3
    • /
    • pp.531-538
    • /
    • 2016
  • In this paper, we propose a novel post-processing method of semantic segmentation from indoor scenes with RGBD inputs. For accurate segmentation, various post-processing methods such as superpixel from color edges or Conditional Random Field (CRF) method considering neighborhood connectivity have been used, but these methods are not efficient due to high complexity and computational cost. To solve this problem, we maximize the efficiency of post processing by using depth superpixel extracted from disparity image to handle object silhouette. Our experimental results show reasonable performances compared to previous methods in the post processing of semantic segmentation.