• Title/Summary/Keyword: 2D-to-3D Video Conversion

Search Result 37, Processing Time 0.028 seconds

Stereoscopic Conversion of Monoscopic Video using Edge Direction Histogram (에지 방향성 히스토그램을 이용한 2차원 동영상의 3차원 입체변환기법)

  • Kim, Jee-Hong;Yoo, Ji-Sang
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.8C
    • /
    • pp.782-789
    • /
    • 2009
  • In this paper, we propose an algorithm for creating stereoscopic video from a monoscopic video. Parallel straight lines in a 3D space get narrower as they are farther from the perspective images on a 2D plane and finally meet at one point that is called a vanishing point. A viewer uses depth perception clues called a vanishing point which is the farthest from a viewer's viewpoint in order to perceive depth information from objects and surroundings thereof to the viewer. The viewer estimates the vanishing point with geometrical features in monoscopic images, and can perceive the depth information with the relationship between the position of the vanishing point and the viewer's viewpoint. In this paper, we propose a method to estimate a vanishing point with edge direction histogram in a general monoscopic image and to create a depth map depending on the position of the vanishing point. With the conversion method proposed through the experimental results, it is seen that stable stereoscopic conversion of a given monoscopic video is achieved.

Applying differential techniques for 2D/3D video conversion to the objects grouped by depth information (2D/3D 동영상 변환을 위한 그룹화된 객체별 깊이 정보의 차등 적용 기법)

  • Han, Sung-Ho;Hong, Yeong-Pyo;Lee, Sang-Hun
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.13 no.3
    • /
    • pp.1302-1309
    • /
    • 2012
  • In this paper, we propose applying differential techniques for 2D/3D video conversion to the objects grouped by depth information. One of the problems converting 2D images to 3D images using the technique tracking the motion of pixels is that objects not moving between adjacent frames do not give any depth information. This problem can be solved by applying relative height cue only to the objects which have no moving information between frames, after the process of splitting the background and objects and extracting depth information using motion vectors between objects. Using this technique all the background and object can have their own depth information. This proposed method is used to generate depth map to generate 3D images using DIBR(Depth Image Based Rendering) and verified that the objects which have no movement between frames also had depth information.

Motion Depth Generation Using MHI for 3D Video Conversion (3D 동영상 변환을 위한 MHI 기반 모션 깊이맵 생성)

  • Kim, Won Hoi;Gil, Jong In;Choi, Changyeol;Kim, Manbae
    • Journal of Broadcast Engineering
    • /
    • v.22 no.4
    • /
    • pp.429-437
    • /
    • 2017
  • 2D-to-3D conversion technology has been studied over past decades and integrated to commercial 3D displays and 3DTVs. Generally, depth cues extracted from a static image is used for generating a depth map followed by DIBR (Depth Image Based Rendering) for producing a stereoscopic image. Further, motion is also an important cue for depth estimation and is estimated by block-based motion estimation, optical flow and so forth. This papers proposes a new method for motion depth generation using Motion History Image (MHI) and evaluates the feasiblity of the MHI utilization. In the experiments, the proposed method was performed on eight video clips with a variety of motion classes. From a qualitative test on motion depth maps as well as the comparison of the processing time, we validated the feasibility of the proposed method.

Technical Improvement Using a Three-Dimensional Video System for Laparoscopic Partial Nephrectomy

  • Komatsuda, Akari;Matsumoto, Kazuhiro;Miyajima, Akira;Kaneko, Gou;Mizuno, Ryuichi;Kikuchi, Eiji;Oya, Mototsugu
    • Asian Pacific Journal of Cancer Prevention
    • /
    • v.17 no.5
    • /
    • pp.2475-2478
    • /
    • 2016
  • Background: Laparoscopic partial nephrectomy is one of the major surgical techniques for small renal masses. However, it is difficult to manage cutting and suturing procedures within acceptable time periods. To overcome this difficulty, we applied a three-dimensional (3D) video system with laparoscopic partial nephrectomy, and evaluated its utility. Materials and Methods: We retrospectively enrolled 31 patients who underwent laparoscopic partial nephrectomy between November 2009 and June 2014. A conventional two-dimensional (2D) video system was used in 20 patients, and a 3D video system in 11. Patient characteristics and video system type (2D or 3D) were recorded, and correlations with perioperative outcomes were analyzed. Results: Mean age of the patients was $55.8{\pm}12.4$, mean body mass index was $25.7{\pm}3.9kg/m^2$, mean tumor size was $2.0{\pm}0.8cm$, mean R.E.N.A.L nephrometry score was $6.9{\pm}1.9$, and clinical stage was T1a in all patients. There were no significant differences in operative time (p=0.348), pneumoperitoneum time (p=0.322), cutting time (p=0.493), estimated blood loss (p=0.335), and Clavien grade of >II complication rate (p=0.719) between the two groups. However, warm ischemic time was significantly shorter in the 3D group than the 2D group (16.1 min vs. 21.2min, p=0.021), which resulted from short suturing time (9.1 min vs. 15.2 min, p=0.008). No open conversion occurred in either group. Conclusions: A 3D video system allows the shortening of warm ischemic time in laparoscopic partial nephrectomy and thus may be useful in improving the procedure.

Producing Stereoscopic Video Contents Using Transformation of Character Objects (캐릭터 객체의 변환을 이용하는 입체 동영상 콘텐츠 제작)

  • Lee, Kwan-Wook;Won, Ji-Yeon;Choi, Chang-Yeol;Kim, Man-Bae
    • Journal of Broadcast Engineering
    • /
    • v.16 no.1
    • /
    • pp.33-43
    • /
    • 2011
  • Recently, 3D displays are supplied in the 3D markets so that the demand for 3D stereoscopic contents increases. In general, a simple method is to use a stereoscopic camera. As well, the production of 3D from 2D materials is regarded as an important technology. Such conversion works have gained much interest in the field of 3D converting. However, the stereoscopic image generation from a single 2D image is limited to simple 2D to 3D conversion so that the better realistic perception is difficult to deliver to the users. This paper presents a new stereoscopic content production method where foreground objects undergo alive action events. Further stereoscopic animation is viewed on 3D displays. Given a 2D image, the production is composed of background image generation, foreground object extraction, object/background depth maps and stereoscopic image generation The alive objects are made using the geometric transformation (e.g., translation, rotation, scaling, etc). The proposed method is performed on a Korean traditional painting, Danopungjung as well as Pixar's Up. The animated video showed that through the utilization of simple object transformations, more realistic perception can be delivered to the viewers.

A New Copyright Protection Scheme for Depth Map in 3D Video

  • Li, Zhaotian;Zhu, Yuesheng;Luo, Guibo;Guo, Biao
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.7
    • /
    • pp.3558-3577
    • /
    • 2017
  • In 2D-to-3D video conversion process, the virtual left and right view can be generated from 2D video and its corresponding depth map by depth image based rendering (DIBR). The depth map plays an important role in conversion system, so the copyright protection for depth map is necessary. However, the provided virtual views may be distributed illegally and the depth map does not directly expose to viewers. In previous works, the copyright information embedded into the depth map cannot be extracted from virtual views after the DIBR process. In this paper, a new copyright protection scheme for the depth map is proposed, in which the copyright information can be detected from the virtual views even without the depth map. The experimental results have shown that the proposed method has a good robustness against JPEG attacks, filtering and noise.

Stereoscopic Conversion based on Key Frames (키 프레임 기반 스테레오스코픽 변환 방법)

  • 김만배;박상훈
    • Journal of Broadcast Engineering
    • /
    • v.7 no.3
    • /
    • pp.219-228
    • /
    • 2002
  • In this paper, we propose a new method of converting 2D video into 3D stereoscopic video, called stereoscopic conversion. In general, stereoscopic images are produced using the motion informations. However unreliable motion informations obtained especially from block-based motion estimation cause the wrong generation of stereoscopic images. To solve for this problem, we propose a stereoscopic conversion method based upon the utilization of key frame that has the better accuracy of estimated motion informations. As well, as generation scheme of stereoscopic images associated with the motion type of each key frame is proposed. For the performance evaluation of our proposed method, we apply it to five test images and measure the accuracy of key frame-based stereoscopic conversion. Experimental results show that our proposed method has the accuracy more than about 90 percent in terms of the detection ratio of key frames.

Non-rigid 3D Shape Recovery from Stereo 2D Video Sequence (스테레오 2D 비디오 영상을 이용한 비정형 3D 형상 복원)

  • Koh, Sung-shik
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.20 no.2
    • /
    • pp.281-288
    • /
    • 2016
  • The natural moving objects are the most non-rigid shapes with randomly time-varying deformation, and its types also very diverse. Methods of non-rigid shape reconstruction have widely applied in field of movie or game industry in recent years. However, a realistic approach requires moving object to stick many beacon sets. To resolve this drawback, non-rigid shape reconstruction researches from input video without beacon sets are investigated in multimedia application fields. In this regard, our paper propose novel CPSRF(Chained Partial Stereo Rigid Factorization) algorithm that can reconstruct a non-rigid 3D shape. Our method is focused on the real-time reconstruction of non-rigid 3D shape and motion from stereo 2D video sequences per frame. And we do not constrain that the deformation of the time-varying non-rigid shape is limited by a Gaussian distribution. The experimental results show that the 3D reconstruction performance of the proposed CPSRF method is superior to that of the previous method which does not consider the random deformation of shape.

High-Quality Depth Map Generation of Humans in Monocular Videos (단안 영상에서 인간 오브젝트의 고품질 깊이 정보 생성 방법)

  • Lee, Jungjin;Lee, Sangwoo;Park, Jongjin;Noh, Junyong
    • Journal of the Korea Computer Graphics Society
    • /
    • v.20 no.2
    • /
    • pp.1-11
    • /
    • 2014
  • The quality of 2D-to-3D conversion depends on the accuracy of the assigned depth to scene objects. Manual depth painting for given objects is labor intensive as each frame is painted. Specifically, a human is one of the most challenging objects for a high-quality conversion, as a human body is an articulated figure and has many degrees of freedom (DOF). In addition, various styles of clothes, accessories, and hair create a very complex silhouette around the 2D human object. We propose an efficient method to estimate visually pleasing depths of a human at every frame in a monocular video. First, a 3D template model is matched to a person in a monocular video with a small number of specified user correspondences. Our pose estimation with sequential joint angular constraints reproduces a various range of human motions (i.e., spine bending) by allowing the utilization of a fully skinned 3D model with a large number of joints and DOFs. The initial depth of the 2D object in the video is assigned from the matched results, and then propagated toward areas where the depth is missing to produce a complete depth map. For the effective handling of the complex silhouettes and appearances, we introduce a partial depth propagation method based on color segmentation to ensure the detail of the results. We compared the result and depth maps painted by experienced artists. The comparison shows that our method produces viable depth maps of humans in monocular videos efficiently.

Overlay Text Graphic Region Extraction for Video Quality Enhancement Application (비디오 품질 향상 응용을 위한 오버레이 텍스트 그래픽 영역 검출)

  • Lee, Sanghee;Park, Hansung;Ahn, Jungil;On, Youngsang;Jo, Kanghyun
    • Journal of Broadcast Engineering
    • /
    • v.18 no.4
    • /
    • pp.559-571
    • /
    • 2013
  • This paper has presented a few problems when the 2D video superimposed the overlay text was converted to the 3D stereoscopic video. To resolve the problems, it proposes the scenario which the original video is divided into two parts, one is the video only with overlay text graphic region and the other is the video with holes, and then processed respectively. And this paper focuses on research only to detect and extract the overlay text graphic region, which is a first step among the processes in the proposed scenario. To decide whether the overlay text is included or not within a frame, it is used the corner density map based on the Harris corner detector. Following that, the overlay text region is extracted using the hybrid method of color and motion information of the overlay text region. The experiment shows the results of the overlay text region detection and extraction process in a few genre video sequence.