• Title/Summary/Keyword: RGB-D Images

Search Result 107, Processing Time 0.344 seconds

Acquisition of Intrinsic Image by Omnidirectional Projection of ROI and Translation of White Patch on the X-chromaticity Space (X-색도 공간에서 ROI의 전방향 프로젝션과 백색패치의 평행이동에 의한 본질 영상 획득)

  • Kim, Dal-Hyoun;Hwang, Dong-Guk;Lee, Woo-Ram;Jun, Byoung-Min
    • The KIPS Transactions:PartB
    • /
    • v.18B no.2
    • /
    • pp.51-56
    • /
    • 2011
  • Algorithms for intrinsic images reduce color differences in RGB images caused by the temperature of black-body radiators. Based on the reference light and detecting single invariant direction, these algorithms are weak in real images which can have multiple invariant directions when the scene illuminant is a colored illuminant. To solve these problems, this paper proposes a method of acquiring an intrinsic image by omnidirectional projection of an ROI and a translation of white patch in the ${\chi}$-chromaticity space. Because it is not easy to analyze an image in the three-dimensional RGB space, the ${\chi}$-chromaticity is also employed without the brightness factor in this paper. After the effect of the colored illuminant is decreased by a translation of white patch, an invariant direction is detected by omnidirectional projection of an ROI in this chromaticity space. In case the RGB image has multiple invariant directions, only one ROI is selected with the bin, which has the highest frequency in 3D histogram. And then the two operations, projection and inverse transformation, make intrinsic image acquired. In the experiments, test images were four datasets presented by Ebner and evaluation methods was the follows: standard deviation of the invariant direction, the constancy measure, the color space measure and the color constancy measure. The experimental results showed that the proposed method had lower standard deviation than the entropy, that its performance was two times higher than the compared algorithm.

Stereoscopic Perception Improvement Using Color and Depth Transformation (컬러 및 깊이 데이터 변환을 이용하는 입체감 향상)

  • Gil, Jong-In;Jang, Seung-Eun;Seo, Joo-Ha;Kim, Man-Bae
    • Journal of Broadcast Engineering
    • /
    • v.16 no.4
    • /
    • pp.584-595
    • /
    • 2011
  • Recently, RGB images and depth maps have been supplied to academic fields. The depth maps are utilized to the generation of stereoscopic images in the diverse formats according to the users' preference. A variety of methods that use depth maps have been introduced so far. One of applications is a medical field. In this area, the improvement of the perceptual quality of 2D medical images has gained much interest. In this paper, we propose a novel scheme that expands the conventional method to 3D stereoscopic image, thereby achieving the perceptual depth quality improvement as well as 3D stereoscopic perception enhancement at the same time. For this, contrast transformation as well as depth darkening are proposed and their performance is validated through the subjective test. Subjective experiments peformed for stereoscopic enhancement as well as visual fatigue validate that the proposed method achieves better 3D perception than the usage of the original stereoscopic image and suggests the limitation in terms of the visual fatigue.

A Study on Analysis of Color characteristic of On-line of Game Play Image of according to the Users (온라인 게임 플레이영상의 이용자에 따른 색채 특성 분석 연구)

  • Jang, Hyun-Joo;Jung, Hyung-Won
    • Journal of Digital Convergence
    • /
    • v.15 no.10
    • /
    • pp.399-404
    • /
    • 2017
  • If game images use colors improper for certain ages, the characteristics of games disappear, and the efficiency of image expression drops. thus a study in color is very important. Nevertheless, there is a lack of research on this subject. In this study, we investigated the difference of color of game images according to users. User group is divided into all users and users over 18 years old. Among the rank of games,10 each from games for all ages and games for non-teen-rated are selected, and the game-playing images are taken as snapshots to have 100 each images each, A total of 2000 images were extracted and the mean values of RGB and HSB of each game were obtained and independent sample t test was performed. The results showed that the RGB color values between the two users were significantly significant. In the future, it is expected that it will be helpful in color selection when developing psychotherapy games and emotional games using color psychology. Furthermore, we will use color image scales to express colors as adjectives, analyze colors in a variety of ways, and investigate the difference in color of game images in each country.

A Robust Object Detection and Tracking Method using RGB-D Model (RGB-D 모델을 이용한 강건한 객체 탐지 및 추적 방법)

  • Park, Seohee;Chun, Junchul
    • Journal of Internet Computing and Services
    • /
    • v.18 no.4
    • /
    • pp.61-67
    • /
    • 2017
  • Recently, CCTV has been combined with areas such as big data, artificial intelligence, and image analysis to detect various abnormal behaviors and to detect and analyze the overall situation of objects such as people. Image analysis research for this intelligent video surveillance function is progressing actively. However, CCTV images using 2D information generally have limitations such as object misrecognition due to lack of topological information. This problem can be solved by adding the depth information of the object created by using two cameras to the image. In this paper, we perform background modeling using Mixture of Gaussian technique and detect whether there are moving objects by segmenting the foreground from the modeled background. In order to perform the depth information-based segmentation using the RGB information-based segmentation results, stereo-based depth maps are generated using two cameras. Next, the RGB-based segmented region is set as a domain for extracting depth information, and depth-based segmentation is performed within the domain. In order to detect the center point of a robustly segmented object and to track the direction, the movement of the object is tracked by applying the CAMShift technique, which is the most basic object tracking method. From the experiments, we prove the efficiency of the proposed object detection and tracking method using the RGB-D model.

2D to 3D Anaglyph Image Conversion using Quadratic & Cubic Bézier Curve in HTML5 (HTML5에서 Quadratic & Cubic Bézier 곡선을 이용한 2D to 3D 입체 이미지 변환)

  • Park, Young Soo
    • Journal of Digital Convergence
    • /
    • v.12 no.12
    • /
    • pp.553-560
    • /
    • 2014
  • In this paper, we propose a method to convert 2D image to 3D anaglyph using quadratic & cubic B$\acute{e}$zier Curves in HTML5. In order to convert 2D image to 3D anaglyph image, we filter the original image to extract the RGB color values and create two images for the left and right eyes. Users are to set up the depth values of the image through the control point using the quadratic and cubic B$\acute{e}$zier curves. We have processed the depth values of 2D image based on this control point to create the 3D image conversion reflecting the value of the control point which the users select. All of this work has been designed and implemented in Web environment in HTML5. So we have made it for anyone who wants to create their 3D images and it is very easy and convenient to use.

An Input/Output Technology for 3-Dimensional Moving Image Processing (3차원 동영상 정보처리용 영상 입출력 기술)

  • Son, Jung-Young;Chun, You-Seek
    • Journal of the Korean Institute of Telematics and Electronics S
    • /
    • v.35S no.8
    • /
    • pp.1-11
    • /
    • 1998
  • One of the desired features for the realizations of high quality Information and Telecommunication services in future is "the Sensation of Reality". This will be achieved only with the visual communication based on the 3- dimensional (3-D) moving images. The main difficulties in realizing 3-D moving image communication are that there is no developed data transmission technology for the hugh amount of data involved in 3-D images and no established technologies for 3-D image recording and displaying in real time. The currently known stereoscopic imaging technologies can only present depth, no moving parallax, so they are not effective in creating the sensation of the reality without taking eye glasses. The more effective 3-D imaging technologies for achieving the sensation of reality are those based on the multiview 3-D images which provides the object image changes as the eyes move to different directions. In this paper, a multiview 3-D imaging system composed of 8 CCD cameras in a case, a RGB(Red, Green, Blue) beam projector, and a holographic screen is introduced. In this system, the 8 view images are recorded by the 8 CCD cameras and the images are transmitted to the beam projector in sequence by a signal converter. This signal converter converts each camera signal into 3 different color signals, i.e., RGB signals, combines each color signal from the 8 cameras into a serial signal train by multiplexing and drives the corresponding color channel of the beam projector to 480Hz frame rate. The beam projector projects images to the holographic screen through a LCD shutter. The LCD shutter consists of 8 LCD strips. The image of each LCD strip, created by the holographic screen, forms as sub-viewing zone. Since the ON period and sequence of the LCD strips are synchronized with those of the camera image sampling adn the beam projector image projection, the multiview 3-D moving images are viewed at the viewing zone.

  • PDF

Full-Color AMOLED with RGBW Pixel Pattern

  • Amold, A.D.;Hatwar, T.K.;Hettel, M.V.;Kane, P.J.;Miller, M.E.;Murdoch, M.J.;Spindler, J.P.;Slyke, S.A. Van;Mameno, K.;Nishikawa, R.;Omura, T.;Matsumoto, S.
    • 한국정보디스플레이학회:학술대회논문집
    • /
    • 2004.08a
    • /
    • pp.808-811
    • /
    • 2004
  • A full-color AMOLED display with an RGBW color filter pattern has been fabricated. Displays with this format require about $^1/_2$ the power of analogous RGB displays. RGBW and RGB 2.16inch diagonal displays with average power consumptions of 180 mW and 340 mW, respectively, are demonstrated for a set of standard digital still camera images at a luminance of 100 cd/$m^2$. In both cases, a white-emitting AMOLED is used as the light source. The higher efficiency of the RGBW format results because a large fraction of a typical image can be represented as white, and the white sub-pixel in an RGBW AMOLED display is highly efficient because of the absence of any color filter. RGBW and RGB AMOLED displays have the same color gamut and, aside from the power consumption difference, are indistinguishable.

  • PDF

An Efficient Hand Gesture Recognition Method using Two-Stream 3D Convolutional Neural Network Structure (이중흐름 3차원 합성곱 신경망 구조를 이용한 효율적인 손 제스처 인식 방법)

  • Choi, Hyeon-Jong;Noh, Dae-Cheol;Kim, Tae-Young
    • The Journal of Korean Institute of Next Generation Computing
    • /
    • v.14 no.6
    • /
    • pp.66-74
    • /
    • 2018
  • Recently, there has been active studies on hand gesture recognition to increase immersion and provide user-friendly interaction in a virtual reality environment. However, most studies require specialized sensors or equipment, or show low recognition rates. This paper proposes a hand gesture recognition method using Deep Learning technology without separate sensors or equipment other than camera to recognize static and dynamic hand gestures. First, a series of hand gesture input images are converted into high-frequency images, then each of the hand gestures RGB images and their high-frequency images is learned through the DenseNet three-dimensional Convolutional Neural Network. Experimental results on 6 static hand gestures and 9 dynamic hand gestures showed an average of 92.6% recognition rate and increased 4.6% compared to previous DenseNet. The 3D defense game was implemented to verify the results of our study, and an average speed of 30 ms of gesture recognition was found to be available as a real-time user interface for virtual reality applications.

Feature based Pre-processing Method to compensate color mismatching for Multi-view Video (다시점 비디오의 색상 성분 보정을 위한 특징점 기반의 전처리 방법)

  • Park, Sung-Hee;Yoo, Ji-Sang
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.15 no.12
    • /
    • pp.2527-2533
    • /
    • 2011
  • In this paper we propose a new pre-processing algorithm applied to multi-view video coding using color compensation algorithm based on image features. Multi-view images have a difference between neighboring frames according to illumination and different camera characteristics. To compensate this color difference, first we model the characteristics of cameras based on frame's feature from each camera and then correct the color difference. To extract corresponding features from each frame, we use Harris corner detection algorithm and characteristic coefficients used in the model is estimated by using Gauss-Newton algorithm. In this algorithm, we compensate RGB components of target images, separately from the reference image. The experimental results with many test images show that the proposed algorithm peformed better than the histogram based algorithm as much as 14 % of bit reduction and 0.5 dB ~ 0.8dB of PSNR enhancement.

Estimation of a Gaze Point in 3D Coordinates using Human Head Pose (휴먼 헤드포즈 정보를 이용한 3차원 공간 내 응시점 추정)

  • Shin, Chae-Rim;Yun, Sang-Seok
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.177-179
    • /
    • 2021
  • This paper proposes a method of estimating location of a target point at which an interactive robot gazes in an indoor space. RGB images are extracted from low-cost web-cams, user head pose is obtained from the face detection (Openface) module, and geometric configurations are applied to estimate the user's gaze direction in the 3D space. The coordinates of the target point at which the user stares are finally measured through the correlation between the estimated gaze direction and the plane on the table plane.

  • PDF