• Title/Summary/Keyword: 3D video

Search Result 1,152, Processing Time 0.027 seconds

3D-Based Monitoring System and Cloud Computing for Panoramic Video Service (3차원 기반의 모니터링 시스템과 클라우드 컴퓨팅을 이용한 파노라믹 비디오 서비스)

  • Cho, Yongwoo;Seok, Joo Myoung;Suh, Doug Young
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.39B no.9
    • /
    • pp.590-597
    • /
    • 2014
  • This paper proposes multi-camera system that relies on 3D views for panoramic video and distribution method about panoramic video generation algorithm by using cloud computing. The proposed monitoring system monitors the projected 3D model view, instead of individual 2D views, to detect image distortions. This can minimize compensation errors caused by parallax, thereby improving the quality of the resulting panoramic video. Panoramic video generation algorithm can be divided into registration part and compositing part. Therefore we propose off-loading method of these parts with cloud computing for panoramic video service.

Fast Stereoscopic 3D Broadcasting System using x264 and GPU (x264와 GPU를 이용한 고속 양안식 3차원 방송 시스템)

  • Choi, Jung-Ah;Shin, In-Yong;Ho, Yo-Sung
    • Journal of Broadcast Engineering
    • /
    • v.15 no.4
    • /
    • pp.540-546
    • /
    • 2010
  • Since the stereoscopic 3-dimensional (3D) video that provides users with a realistic multimedia service requires twice as much data as 2-dimensional (2D) video, it is difficult to construct the fast system. In this paper, we propose a fast stereoscopic 3D broadcasting system based on the depth information. Before the transmission, we encode the input 2D+depth video using x264, an open source H.264/AVC fast encoder to reduce the size of the data. At the receiver, we decode the transmitted bitstream in real time using a compute unified device architecture (CUDA) video decoder API on NVIDIA graphics processing unit (GPU). Then, we apply a fast view synthesis method that generates the virtual view using GPU. The proposed system can display the output video in both 2DTV and 3DTV. From the experiment, we verified that the proposed system can service the stereoscopic 3D contents in 24 frames per second at most.

View Synthesis Error Removal for Comfortable 3D Video Systems (편안한 3차원 비디오 시스템을 위한 영상 합성 오류 제거)

  • Lee, Cheon;Ho, Yo-Sung
    • Smart Media Journal
    • /
    • v.1 no.3
    • /
    • pp.36-42
    • /
    • 2012
  • Recently, the smart applications, such as smart phone and smart TV, become a hot issue in IT consumer markets. In particular, the smart TV provides 3D video services, hence efficient coding methods for 3D video data are required. Three-dimensional (3D) video involves stereoscopic or multi-view images to provide depth experience through 3D display systems. Binocular cues are perceived by rendering proper viewpoint images obtained at slightly different view angles. Since the number of viewpoints of the multi-view video is limited, 3D display devices should generate arbitrary viewpoint images using available adjacent view images. In this paper, after we explain a view synthesis method briefly, we propose a new algorithm to compensate view synthesis errors around object boundaries. We describe a 3D warping technique exploiting the depth map for viewpoint shifting and a hole filling method using multi-view images. Then, we propose an algorithm to remove boundary noises that are generated due to mismatches of object edges in the color and depth images. The proposed method reduces annoying boundary noises near object edges by replacing erroneous textures with alternative textures from the other reference image. Using the proposed method, we can generate perceptually inproved images for 3D video systems.

  • PDF

Effective Hand Gesture Recognition by Key Frame Selection and 3D Neural Network

  • Hoang, Nguyen Ngoc;Lee, Guee-Sang;Kim, Soo-Hyung;Yang, Hyung-Jeong
    • Smart Media Journal
    • /
    • v.9 no.1
    • /
    • pp.23-29
    • /
    • 2020
  • This paper presents an approach for dynamic hand gesture recognition by using algorithm based on 3D Convolutional Neural Network (3D_CNN), which is later extended to 3D Residual Networks (3D_ResNet), and the neural network based key frame selection. Typically, 3D deep neural network is used to classify gestures from the input of image frames, randomly sampled from a video data. In this work, to improve the classification performance, we employ key frames which represent the overall video, as the input of the classification network. The key frames are extracted by SegNet instead of conventional clustering algorithms for video summarization (VSUMM) which require heavy computation. By using a deep neural network, key frame selection can be performed in a real-time system. Experiments are conducted using 3D convolutional kernels such as 3D_CNN, Inflated 3D_CNN (I3D) and 3D_ResNet for gesture classification. Our algorithm achieved up to 97.8% of classification accuracy on the Cambridge gesture dataset. The experimental results show that the proposed approach is efficient and outperforms existing methods.

Performance Comparisons of Depth Map Post Processing for 3D Video System (3 차원 영상 시스템의 깊이영상 후처리 필터 성능 비교)

  • Lee, Do Hoon;Yoon, Eun Ji;Oh, Byung Tae
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2014.06a
    • /
    • pp.81-83
    • /
    • 2014
  • This paper provides the performance comparison of selected post filters for depth map in 3D video system. In the performance comparison, the dilation filter which is currently adopted in 3D-ATM reference S/W, the bilateral filter, and the depth-oriented depth boundary reconstruction filter. In the paper, we first introduce those filters in details, and show the experimental results as post filter in 3D video system.

  • PDF

The Implementation of Information Providing Method System for Indoor Area by using the Immersive Media's Video Information (실감미디어 동영상정보를 이용한 실내 공간 정보 제공 시스템 구현)

  • Lee, Sangyoon;Ahn, Heuihak
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.12 no.3
    • /
    • pp.157-166
    • /
    • 2016
  • This paper presents the interior space information using 6D-360 degree immersive media video information. And we implement the augmented reality, which includes a variety of information such as position information, movement information of the specific location in the interior space GPS signal does not reach the position information. Augmented reality containing the 6D-360 degree immersive media video information provides the position information and the three dimensional space image information to identify the exact location of a user in an interior space of a moving object as well as a fixed interior space. This paper constitutes a three dimensional image database based on the 6D-360 degree immersive media video information and provides augmented reality service. Therefore, to map the various information to 6D-360 degree immersive media video information, the user can check the plant in the same environment as the actual. It suggests the augmented reality service for the emergency escape and repair to the passengers and employees.

HTTP Adaptive Streaming Method for Service-compatible 3D Contents Based on MPEG DASH (MPEG DASH 기반 service-compatible 3D 콘텐츠 대상 HTTP adaptive streaming 적용방안)

  • Park, Gi-Jun;Lee, Gil-Bok;Lee, Jang-Won;Kim, Kyu-Heon
    • Journal of Broadcast Engineering
    • /
    • v.17 no.2
    • /
    • pp.207-222
    • /
    • 2012
  • Recently, many consumer electronics manufacturers have produced 3D devices such as 3DTVs and 3D monitors as interests in a stereoscopic video service are getting increased. However, most of 3D services are focused on local storage or bandwidth guaranteed service since 3D stereoscopic video service require bandwidth more stable and larger. This property causes difficulties in seamless stereoscopic video streaming services under IP based open network environment that cannot guarantee quality of services. In order to achieve a seamless video streaming service the international standard organization MPEG (Moving Pictures Experts Group) has developed the adaptive HTTP streaming technology called as DASH (Dynamic Adaptive Streaming over HTTP). However, the DASH doesn't have obvious scheme which can express the two elementary video streams based service-compatible stereoscopic contents in one segment. Therefore, this paper proposes a scheme of efficient 3D adaptive streaming service based on the DASH, which covers not only frame-packing stereoscopic contents but also service-compatible ones. The 3D adaptive HTTP streaming scheme introduced in this paper is able to provide 3D contents with various qualities to user and also has benefit that single 3D content can be applied to a variety of devices.

Style Synthesis of Speech Videos Through Generative Adversarial Neural Networks (적대적 생성 신경망을 통한 얼굴 비디오 스타일 합성 연구)

  • Choi, Hee Jo;Park, Goo Man
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.11 no.11
    • /
    • pp.465-472
    • /
    • 2022
  • In this paper, the style synthesis network is trained to generate style-synthesized video through the style synthesis through training Stylegan and the video synthesis network for video synthesis. In order to improve the point that the gaze or expression does not transfer stably, 3D face restoration technology is applied to control important features such as the pose, gaze, and expression of the head using 3D face information. In addition, by training the discriminators for the dynamics, mouth shape, image, and gaze of the Head2head network, it is possible to create a stable style synthesis video that maintains more probabilities and consistency. Using the FaceForensic dataset and the MetFace dataset, it was confirmed that the performance was increased by converting one video into another video while maintaining the consistent movement of the target face, and generating natural data through video synthesis using 3D face information from the source video's face.

3D video coding for e-AG using spatio-temporal scalability (e-AG를 위한 시공간적 계위를 이용한 3차원 비디오 압축)

  • 오세찬;이영호;우운택
    • Proceedings of the IEEK Conference
    • /
    • 2003.11a
    • /
    • pp.199-202
    • /
    • 2003
  • In this paper, we propose a new 3D coding method for heterogeneous systems over enhanced Access Grid (e-AG) with 3D display using spatio-temporal scalability. The proposed encoder produces four bit-streams: one base layer and enhancement layer l, 2 and 3. The base layer represents a video sequence for left eye with lower spatial resolution. An enhancement layer l provides additional bit-stream needed for reproduction of frames produced in base layer with full resolution. Similarly, the enhancement layer 2 represents a video sequence for right eye with lower spatial resolution and an enhancement layer 3 provides additional bit-stream needed for reproduction of its reference pictures with full resolution. In this system, temporal resolution reduction is obtained by dropping B-frames in the receiver according to network condition. The receiver system can select the spatial and temporal resolution of video sequence with its display condition by properly combining bit-streams.

  • PDF

Digital Video Watermarking Using Frame Division And 3D Wavelet Transform (프레임 분할과 3D 웨이블릿 변환을 이용한 비디오 워터마킹)

  • Kim, Kwang-Il;Cui, Jizhe;Kim, Jong-Weon;Choi, Jong-Uk
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.18 no.3
    • /
    • pp.155-162
    • /
    • 2008
  • In this paper we proposed a video watermarking algorithm based on a three dimension discrete wavelet transform (3D DWT) and direct spread spectrum (DSS). In the proposed method, the information watermark is embedded into followed frames, after sync watermark is embedded into the first frame. Input frames are divided into sub frames which are located odd row and even row. The sub frames are arranged as 3D frames, and transformed into 3D wavelet domain. In this domain the watermark is embedded using DSS. Existing video watermarking using 3D DWT is non-blind method but, proposed algorithm uses blind method. The experimental results show that the proposed algorithm is robust against frame cropping, noise addition, compression, etc. acquiring BER of 10% or below and sustains level of 40dB or above on the average.