• Title/Summary/Keyword: Depth Video

Search Result 450, Processing Time 0.028 seconds

View Synthesis Error Removal for Comfortable 3D Video Systems (편안한 3차원 비디오 시스템을 위한 영상 합성 오류 제거)

  • Lee, Cheon;Ho, Yo-Sung
    • Smart Media Journal
    • /
    • v.1 no.3
    • /
    • pp.36-42
    • /
    • 2012
  • Recently, the smart applications, such as smart phone and smart TV, become a hot issue in IT consumer markets. In particular, the smart TV provides 3D video services, hence efficient coding methods for 3D video data are required. Three-dimensional (3D) video involves stereoscopic or multi-view images to provide depth experience through 3D display systems. Binocular cues are perceived by rendering proper viewpoint images obtained at slightly different view angles. Since the number of viewpoints of the multi-view video is limited, 3D display devices should generate arbitrary viewpoint images using available adjacent view images. In this paper, after we explain a view synthesis method briefly, we propose a new algorithm to compensate view synthesis errors around object boundaries. We describe a 3D warping technique exploiting the depth map for viewpoint shifting and a hole filling method using multi-view images. Then, we propose an algorithm to remove boundary noises that are generated due to mismatches of object edges in the color and depth images. The proposed method reduces annoying boundary noises near object edges by replacing erroneous textures with alternative textures from the other reference image. Using the proposed method, we can generate perceptually inproved images for 3D video systems.

  • PDF

A Fast TU Size Decision Method for HEVC RQT Coding

  • Wu, Jinfu;Guo, Baolong;Yan, Yunyi;Hou, Jie;Zhao, Dan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.6
    • /
    • pp.2271-2288
    • /
    • 2015
  • The emerging high efficiency video coding (HEVC) standard adopts the quadtree-structured transform unit (TU) in the residual quadtree (RQT) coding. Each TU allows to be split into four equal sub-TUs recursively. The RQT coding is performed for all the possible transform depth levels to achieve the highest coding efficiency, but it requires a very high computational complexity for HEVC encoders. In order to reduce the computational complexity requested by the RQT coding, in this paper, we propose a fast TU size decision method incorporating an adaptive maximum transform depth determination (AMTD) algorithm and a full check skipping - early termination (FCS-ET) algorithm. Because the optimal transform depth level is highly content-dependent, it is not necessary to perform the RQT coding at all transform depth levels. By the AMTD algorithm, the maximum transform depth level is determined for current treeblock to skip those transform depth levels rarely used by its spatially adjacent treeblocks. Additionally, the FCS-ET algorithm is introduced to exploit the correlations of transform depth level between four sub-CUs generated by one coding unit (CU) quadtree partitioning. Experimental results demonstrate that the proposed overall algorithm significantly reduces on average 21% computational complexity while maintaining almost the same rate distortion (RD) performance as the HEVC test model reference software, HM 13.0.

Depth compression method for 3D video (3차원 영상을 위한 깊이 영상 압축 방법)

  • Nam, Jung-Hak;Hwang, Neung-Joo;Cho, Gwang-Shin;Sim, Dong-Gyu;Lee, Soo-Youn;Bang, Gun;Hur, Nam-Ho
    • Journal of Broadcast Engineering
    • /
    • v.15 no.5
    • /
    • pp.703-706
    • /
    • 2010
  • Recently, a need to encode a depth image has been raising with the deployment of 3D video services. The 3DV/FTV group in the MPEG has standardized the compression method of depth map image. Because conventional depth map coding methods are independently encoded without referencing the color image, coding performance of conventional algorithms is poor. In this letter, we proposed a novel method which rearranged modes of depth blocks according to modes of corresponding color blocks by using a correlation between color and depth images. In experimental results, the proposed method achieves bits reduction of 2.2% compared with coding method based on JSVM.

Depth Map Pre-processing using Gaussian Mixture Model and Mean Shift Filter (혼합 가우시안 모델과 민쉬프트 필터를 이용한 깊이 맵 부호화 전처리 기법)

  • Park, Sung-Hee;Yoo, Ji-Sang
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.15 no.5
    • /
    • pp.1155-1163
    • /
    • 2011
  • In this paper, we propose a new pre-processing algorithm applied to depth map to improve the coding efficiency. Now, 3DV/FTV group in the MPEG is working for standard of 3DVC(3D video coding), but compression method for depth map images are not confirmed yet. In the proposed algorithm, after dividing the histogram distribution of a given depth map by EM clustering method based on GMM, we classify the depth map into several layered images. Then, we apply different mean shift filter to each classified image according to the existence of background or foreground in it. In other words, we try to maximize the coding efficiency while keeping the boundary of each object and taking average operation toward inner field of the boundary. The experiments are performed with many test images and the results show that the proposed algorithm achieves bits reduction of 19% ~ 20% and computation time is also reduced.

Multi-View Video Processing: IVR, Graphics Composition, and Viewer

  • Kwon, Jun-Sup;Hwang, Won-Young;Choi, Chang-Yeol;Chang, Eun-Young;Hur, Nam-Ho;Kim, Jin-Woong;Kim, Man-Bae
    • Journal of Broadcast Engineering
    • /
    • v.12 no.4
    • /
    • pp.333-341
    • /
    • 2007
  • Multi-view video has recently gained much attraction from academic and commercial fields because it can deliver the immersive viewing of natural scenes. This paper presents multi-view video processing being composed of intermediate view reconstruction (IVR), graphics composition, and multi-view video viewer. First we generate virtual views between multi-view cameras using depth and texture images of the input videos. Then we mix graphic objects to the generated view images. The multi-view video viewer is developed to examine the reconstructed images and composite images. As well, it can provide users with some special effects of multi-view video. We present experimental results that validate our proposed method and show that graphic objects could become the inalienable part of the multi-view video.

Extracting Graphics Information for Better Video Compression

  • Hong, Kang Woon;Ryu, Won;Choi, Jun Kyun;Lim, Choong-Gyoo
    • ETRI Journal
    • /
    • v.37 no.4
    • /
    • pp.743-751
    • /
    • 2015
  • Cloud gaming services are heavily dependent on the efficiency of real-time video streaming technology owing to the limited bandwidths of wire or wireless networks through which consecutive frame images are delivered to gamers. Video compression algorithms typically take advantage of similarities among video frame images or in a single video frame image. This paper presents a method for computing and extracting both graphics information and an object's boundary from consecutive frame images of a game application. The method will allow video compression algorithms to determine the positions and sizes of similar image blocks, which in turn, will help achieve better video compression ratios. The proposed method can be easily implemented using function call interception, a programmable graphics pipeline, and off-screen rendering. It is implemented using the most widely used Direct3D API and applied to a well-known sample application to verify its feasibility and analyze its performance. The proposed method computes various kinds of graphics information with minimal overhead.

Assessment of the proficiency and usability of direct laryngoscopy and video laryngoscopy (직접 후두경과 비디오 후두경의 숙련도 및 유용성 평가)

  • Shin, Gyo-Seok;Tak, Yang-Ju
    • The Korean Journal of Emergency Medical Services
    • /
    • v.23 no.1
    • /
    • pp.87-99
    • /
    • 2019
  • Purpose: The aim of this study was conducted to assess the proficiency of both direct laryngoscopy and video laryngoscopy and the usefulness of each laryngoscope, thereby provide basic data for further education using video laryngoscopy. Methods: Forty one paramedic subjects participated in this study. Usability was measured with the System usability scale. The Macintosh direct laryngoscope and $C-MAC^{(R)}$ video laryngoscope were two instruments evaluated in the study. Results: Training with video laryngoscopy showed significantly better results within the categories of dental injury (p=.004), esophageal intubation (p=.001), and proper depth placement of intubation tubes (p=.019). The results of the System usability scale questionnaire and the degrees of visibility based on the Cormack & Lehane classification were also found to be better achieved with the video laryngoscopy (p=.000). Conclusion: This study suggests enhancing education with video laryngoscopy, which could reduce the risk of complications and duration of intubation while increasing the success rate among students and emergency medical technicians with little experience, rather than the existing method of only using direct laryngoscope, which requires considerable experience and skills.

Designing Video-based Teacher Professional Development: Teachers' Meaning Making with a Video Annotation Tool

  • SO, Hyo-Jeong;LIM, Weiying;XIONG, Yao
    • Educational Technology International
    • /
    • v.17 no.1
    • /
    • pp.87-116
    • /
    • 2016
  • In this research, we designed a teacher professional development (PD) program where a small group of mathematics teachers could share, reflect on, and discuss their pedagogical knowledge and practices of ICT-integrated lessons, using a video annotation tool called DIVER. The main purposes of this paper are both micro and macro: to examine how the teachers were engaged in the meaning-making process in a video-based PD (micro); and to derive implications about how to design effective video-based teacher PD programs toward a teacher community of practices (macro). To examine teachers' meaning-making in the PD sessions, discourse data from a series of 10 meetings was segmented into idea units and coded to identify discourse patterns, focusing on (a) participation levels, (b) conversation topics, and (c) conversation depth. Regarding the affordance of DIVER, discourse patterns of two meetings, before and after individual annotation with DIVER were compared through qualitative vignette analysis. Overall, we found that the teacher discourse shifted the focus from surface features to deeper pedagogical issues as the PD sessions progressed. In particular, the annotation function in DIVER afforded the teachers to exercise descriptive analyses of video clips in a flexible manner, thereby helping them cognitively prepared to take interpretative and evaluative stances in face-to-face discussions with colleagues. In conclusion, deriving from our research experiences, we discuss the possibilities and challenges of designing video-based teacher PD in a school context.

An Exploratory Study on Video Information Literacy (영상정보 활용능력에 관한 탐색적 연구)

  • Min Kyung Na;Jee Yeon Lee
    • Journal of the Korean Society for information Management
    • /
    • v.41 no.2
    • /
    • pp.19-46
    • /
    • 2024
  • In this study, we conducted a literature review and exploratory research to identify the characteristics of recently popular video information and to propose the basic capabilities required for video information literacy. Through a literature review, the distinct characteristics of video information were examined from various perspectives, differentiating it from other types of information. Subsequently, we had one-on-one, in-depth, semi-structured interviews with 16 participants in their teens to 50s to collect their video usage experiences. The interview contents were categorized to create a codebook, and content analysis was performed. Based on this analysis, we derived the characteristics of video information. Finally, the characteristics of video information were identified through the literature review and interview analysis outcomes, and these characteristics were classified into properties of video and characteristics related to video information usage. Based on the identified characteristics of video information, this study proposed the basic capabilities required for video information literacy.

A Recognition Method for Moving Objects Using Depth and Color Information (깊이와 색상 정보를 이용한 움직임 영역의 인식 방법)

  • Lee, Dong-Seok;Kwon, Soon-Kak
    • Journal of Korea Multimedia Society
    • /
    • v.19 no.4
    • /
    • pp.681-688
    • /
    • 2016
  • In the intelligent video surveillance, recognizing the moving objects is important issue. However, the conventional moving object recognition methods have some problems, that is, the influence of light, the distinguishing between similar colors, and so on. The recognition methods for the moving objects using depth information have been also studied, but these methods have limit of accuracy because the depth camera cannot measure the depth value accurately. In this paper, we propose a recognition method for the moving objects by using both the depth and the color information. The depth information is used for extracting areas of moving object and then the color information for correcting the extracted areas. Through tests with typical videos including moving objects, we confirmed that the proposed method could extract areas of moving objects more accurately than a method using only one of two information. The proposed method can be not only used in CCTV field, but also used in other fields of recognizing moving objects.