• Title/Summary/Keyword: 3-D video generation

Search Result 93, Processing Time 0.025 seconds

3D Video Processing for 3DTV

  • Sohn, Kwang-Hoon
    • 한국정보디스플레이학회:학술대회논문집
    • /
    • 2007.08b
    • /
    • pp.1231-1234
    • /
    • 2007
  • This paper presents the overview of 3D video processing technologies for 3DTV such as 3D content generation, 3D video codec and video processing techniques for 3D displays. Some experimental results for 3D contents generation are shown in 3D mixed reality and 2D/3D conversion.

  • PDF

Fast Generation of 3-D Video Holograms using a Look-up Table and Temporal Redundancy of 3-D Video Image (룩업테이블과 3차원 동영상의 시간적 중복성을 이용한 3차원 비디오 홀로그램의 고속 생성)

  • Kim, Seung-Cheol;Kim, Eun-Soo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.10B
    • /
    • pp.1076-1085
    • /
    • 2009
  • In this paper, a new method for efficient computation of CGH patterns for 3-D video images is proposed by combined use of temporal redundancy and look-up table techniques. In the conventional N-LT method, fringe patterns for other object points on that image plane can be obtained by simply shifting these pre-calculated PFP (Principle Fringe Patterns). But there have been many practical limitations in real-time generation of 3-D video holograms because the computation time required for the generation of 3-D video holograms must be massively increased compared to that of the static holograms. On the other hand, as ordinary 3-D moving pictures have numerous similarities between video frames, called by temporal redundancy, and this redundancy is used to compress the video image. Therefore, in this paper, we proposed the efficient hologram generation method using the temporal redundancy of 3-D video image and N-LT method. To confirm the feasibility of the proposed method, some experiments with test 3-D videos are carried out, and the results are comparatively discussed with the conventional methods in terms of the number of object points and computation time.

Real-Time 2D-to-3D Conversion for 3DTV using Time-Coherent Depth-Map Generation Method

  • Nam, Seung-Woo;Kim, Hye-Sun;Ban, Yun-Ji;Chien, Sung-Il
    • International Journal of Contents
    • /
    • v.10 no.3
    • /
    • pp.9-16
    • /
    • 2014
  • Depth-image-based rendering is generally used in real-time 2D-to-3D conversion for 3DTV. However, inaccurate depth maps cause flickering issues between image frames in a video sequence, resulting in eye fatigue while viewing 3DTV. To resolve this flickering issue, we propose a new 2D-to-3D conversion scheme based on fast and robust depth-map generation from a 2D video sequence. The proposed depth-map generation algorithm divides an input video sequence into several cuts using a color histogram. The initial depth of each cut is assigned based on a hypothesized depth-gradient model. The initial depth map of the current frame is refined using color and motion information. Thereafter, the depth map of the next frame is updated using the difference image to reduce depth flickering. The experimental results confirm that the proposed scheme performs real-time 2D-to-3D conversions effectively and reduces human eye fatigue.

Pattern-based Depth Map Generation for Low-complexity 2D-to-3D Video Conversion (저복잡도 2D-to-3D 비디오 변환을 위한 패턴기반의 깊이 생성 알고리즘)

  • Han, Chan-Hee;Kang, Hyun-Soo;Lee, Si-Woong
    • The Journal of the Korea Contents Association
    • /
    • v.15 no.2
    • /
    • pp.31-39
    • /
    • 2015
  • 2D-to-3D video conversion vests 3D effects in a 2D video by generating stereoscopic views using depth cues inherent in the 2D video. This technology would be a good solution to resolve the problem of 3D content shortage during the transition period to the full ripe 3D video era. In this paper, a low-complexity depth generation method for 2D-to-3D video conversion is presented. For temporal consistency in global depth, a pattern-based depth generation method is newly introduced. A low-complexity refinement algorithm for local depth is also provided to improve 3D perception in object regions. Experimental results show that the proposed method outperforms conventional methods in terms of complexity and subjective quality.

Efficient generation of hologram news ticker using N-LUT method

  • Kim, Seung-Cheol;Kim, Eun-Soo
    • 한국정보디스플레이학회:학술대회논문집
    • /
    • 2009.10a
    • /
    • pp.1375-1378
    • /
    • 2009
  • In this paper, a new method to efficiently generate the holographic news ticker in holographic 3DTV or 3-D movies using N-LUT method is proposed. The proposed method is largely consisted of five steps: construction of the LUT for each character, extraction of characters in news ticker, generation and shift of the CGH pattern for news ticker using the LUT, composition of hologram pattern for 3-D video and news ticker and reconstruct the holographic 3D video with news ticker. From some simulation results confirmed the feasibility of the proposed method in fast generation of CGH patterns for holographic news ticker.

  • PDF

Depth Map Based Distributed Multi-view Video Coding Scheme through an Efficient Side Information Generation (효율적인 보조 정보 생성을 통한 깊이지도 기반의 분산 다시점 비디오 코딩 기법)

  • Yoo, Ji-Hwan;Lee, Dong-Seok;Kim, Tae-June;Yoo, Ji-Sang
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.10B
    • /
    • pp.1093-1103
    • /
    • 2009
  • In this paper, we propose a new depth map based distributed multi-view video coding algorithm through an efficient side information generation. A distributed video coding scheme corrects errors between an original image and side information generated at a decoder by using channel coding techniques. Therefore, the more accurate side information is generated, the better performance of distributed video coding scheme is achieved. In the proposed algorithm, a distributed video coding scheme is applied to multi-view video coding based on depth map. Side information is also generated from images of adjacent views through 3D warping by using a depth map and is also combined with MCTI(motion compensated temporal interpolation) which uses images on a temporal axis, and 3D warping. Experimental results show that side information generated by using the proposed algorithm has 0.97dB better average PSNR compared with using MCTI and 3D warping separated. In addition, 8.01% of average bit-rate has been decreased while the same PSNR in R-D curves is kept.

Efficient Generation of 3-D Video Holograms Using Temporal-Spatial Redundancy of 3-D Moving Images (3차원 동영상의 시ㆍ공간적 정보 중복성을 이용한 효과적인 3차원 비디오 홀로그램의 생성)

  • Kim, Dong-Wook;Koo, Jung-Sik;Kim, Seung-Cheol;Kim, Eun-Soo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37C no.10
    • /
    • pp.859-869
    • /
    • 2012
  • In this paper, a new method to efficiently generate the 3-D(three-dimensional) video holograms for 3-D moving scenes, which is called here the TSR-N-LUT method, is proposed by the combined use of temporal-spatial redundancy(TSR) of 3-D video images and novel look-up table(N-LUT) technique. That is, in the proposed scheme, with the differential pulse code modulation (DPCM) algorithm, temporally redundancy redundant data in the inter-frame of a 3-D video images are removed between the frames, and then inter-line redundant data in the inter-frame of 3-D video images are also removed by using the DPCM method between the lines. Experimental results show that the proposed method could reduced the number of calculated object points and the calculation time of one object point by 23.72% and 19.55%, respectively on the average compared to the conventional method. Good experimental results with 3-D test moving pictures finally confirmed the feasibility of the proposed method to the fast generation of CGH patterns of the 3-D video images.

Video Captioning with Visual and Semantic Features

  • Lee, Sujin;Kim, Incheol
    • Journal of Information Processing Systems
    • /
    • v.14 no.6
    • /
    • pp.1318-1330
    • /
    • 2018
  • Video captioning refers to the process of extracting features from a video and generating video captions using the extracted features. This paper introduces a deep neural network model and its learning method for effective video captioning. In this study, visual features as well as semantic features, which effectively express the video, are also used. The visual features of the video are extracted using convolutional neural networks, such as C3D and ResNet, while the semantic features are extracted using a semantic feature extraction network proposed in this paper. Further, an attention-based caption generation network is proposed for effective generation of video captions using the extracted features. The performance and effectiveness of the proposed model is verified through various experiments using two large-scale video benchmarks such as the Microsoft Video Description (MSVD) and the Microsoft Research Video-To-Text (MSR-VTT).

Performance Analysis of 3D-HEVC Video Coding (3D-HEVC 비디오 부호화 성능 분석)

  • Park, Daemin;Choi, Haechul
    • Journal of Broadcast Engineering
    • /
    • v.19 no.5
    • /
    • pp.713-725
    • /
    • 2014
  • Multi-view and 3D video technologies for a next generation video service are widely studied. These technologies can make users feel realistic experience as supporting various views. Because acquisition and transmission of a large number of views require a high cost, main challenges for multi-view and 3D video include view synthesis, video coding, and depth coding. Recently, JCT-3V (joint collaborative team on 3D video coding extension development) has being developed a new standard for multi-view and 3D video. In this paper, major tools adopted in this standard are introduced and evaluated in terms of coding efficiency and complexity. This performance analysis would be helpful for the development of a fast 3D video encoder as well as a new 3D video coding algorithm.

A Study on Generation of Free Stereo Mosaic Image Using Video Sequences (비디오 프레임 영상을 이용한 자유 입체 모자이크 영상 제작에 관한 연구)

  • Noh, Myoung-Jong;Cho, Woo-Sug;Park, June-Ku
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.27 no.4
    • /
    • pp.453-460
    • /
    • 2009
  • For constructing 3D information using aerial photograph or video sequences, left and right stereo images having different viewing angle should be prepared in overlapping area. In video sequences, left and right stereo images would be generated by mosaicing left and right slice images extracted in consecutive video sequences. Therefore, this paper is focused on generating left and right stereo mosaic images that are able to construct 3D information and video sequences could be made for the best use. In the stereo mosaic generation, motion parameters between video sequences should be firstly determined. In this paper, to determine motion parameters, free mosaic method using geometric relationship, such as relative orientation parameters, between consecutive frame images without GPS/INS geo-data have applied. After determining the motion parameters, the mosaic image have generated by 4 step processes: image registration, image slicing, determining on stitching line, and 3D image mosaicking. As the result of experiment, generated stereo mosaic image and analyzed result of x, y-parallax have showed.