• Title/Summary/Keyword: 3D video

Search Result 1,152, Processing Time 0.029 seconds

Performance Evaluation of Differentiated Services to MPEG-4 FGS Video Streaming (MPEC-4 FGS 비디오 스트리밍에 대한 네트워크 차별화 서비스의 성능분석)

  • 신지태;김종원
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.27 no.7A
    • /
    • pp.711-723
    • /
    • 2002
  • A finer granular scalable (FGS) version of ISO/IEC MPEG-4 video streaming is investigated in this work with the prioritized stream delivery over loss-rate differentiated networks. Our proposed system is focused on the seamless integration of rate adaptation, prioritized packetization, and simplified differentiation for the MPEG-4 FGS video streaming. The proposed system consists of three key components: 1) rate adaptation with scalable source encoding, 2) content-aware prioritized packetization, and 3) loss-based differential forwarding. More specifically, a constant-quality rate adaptation is first achieved by optimally truncating the over-coded FGS stream based on the embedding rate-distortion (R-D) information (obtained from a piecewise linear R-D model). The rate-controlled video stream is then packetized and prioritized according to the loss impact of each packet. Prioritized packets are transmitted over the underlying network, where packets are subject to differentiated dropping and forwarding. By focusing on the end-to-end quality, we establish an effective working conditions for the proposed video streaming and the superior performance is verified by simulated MPEG-4 FGS video streaming.

A Real-time Multiview Video Coding System using Fast Disparity Estimation

  • Bae, Kyung-Hoon;Woo, Byung-Kwang
    • Journal of the Korean Institute of Illuminating and Electrical Installation Engineers
    • /
    • v.22 no.7
    • /
    • pp.37-42
    • /
    • 2008
  • In this paper, a real-time multiview video coding system using fast disparity estimation is proposed. In the multiview encoder, adaptive disparity-motion estimation (DME) for an effective 3-dimensional (3D) processing are proposed. That is, by adaptively predicting the mutual correlation between stereo images in the key-frame using the proposed algorithm, the bandwidth of stereo input images can be compressed to the level of a conventional 2D image and a predicted image also can be effectively reconstructed using a reference image and adaptive disparity vectors. Also, in multiview decoder, intermediate view reconstruction (IVR) using adaptive disparity search algorithm (DSA) for real-time multiview video processing is proposed. The proposed IVR can reduce a processing time of disparity estimation by selecting adaptively disparity search range. Accordingly, the proposed multiview video coding system is able to increase the efficiency of the coding rate and improve the resolution.

A Robust Approach for Human Activity Recognition Using 3-D Body Joint Motion Features with Deep Belief Network

  • Uddin, Md. Zia;Kim, Jaehyoun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.2
    • /
    • pp.1118-1133
    • /
    • 2017
  • Computer vision-based human activity recognition (HAR) has become very famous these days due to its applications in various fields such as smart home healthcare for elderly people. A video-based activity recognition system basically has many goals such as to react based on people's behavior that allows the systems to proactively assist them with their tasks. A novel approach is proposed in this work for depth video based human activity recognition using joint-based motion features of depth body shapes and Deep Belief Network (DBN). From depth video, different body parts of human activities are segmented first by means of a trained random forest. The motion features representing the magnitude and direction of each joint in next frame are extracted. Finally, the features are applied for training a DBN to be used for recognition later. The proposed HAR approach showed superior performance over conventional approaches on private and public datasets, indicating a prominent approach for practical applications in smartly controlled environments.

Robust 3D Wavelet Watermarking for Video

  • Jie Yang;Kim, Young-Gon;Lee, Moon-Ho
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2002.11a
    • /
    • pp.173-176
    • /
    • 2002
  • This paper proposes a new approach for digital watermarking and secure copyright protection of video. the principal aim being to discourage illicit copying and distribution of copyrighted material. The method presented here is based on the three dimensional discrete wavelet transform of video scene. The watermark is a copyright information encoded in the form of a spread spectrum signal to ensure the system security and is embedded in the 3D DWT magnitude of video chunks. The performance of the presented technique is evaluated experimentally.

  • PDF

Video-Based Augmented Reality without Euclidean Camera Calibration (유클리드 카메라 보정을 하지 않는 비디오 기반 증강현실)

  • Seo, Yong-Deuk
    • Journal of the Korea Computer Graphics Society
    • /
    • v.9 no.3
    • /
    • pp.15-21
    • /
    • 2003
  • An algorithm is developed for augmenting a real video with virtual graphics objects without computing Euclidean information. Real motion of the camera is obtained in affine space by a direct linear method using image matches. Then, virtual camera is provided by determining the locations of four basis points in two input images as initialization process. The four pairs of 2D location and its 3D affine coordinates provide Euclidean orthographic projection camera through the whole video sequence. Our method has the capability of generating views of objects shaded by virtual light sources, because we can make use of all the functions of the graphics library written on the basis of Euclidean geometry. Our novel formulation and experimental results with real video sequences are presented.

  • PDF

Real-time video Surveillance System Design Proposal Using Abnormal Behavior Recognition Technology

  • Lee, Jiyoo;Shin, Seung-Jung
    • International journal of advanced smart convergence
    • /
    • v.9 no.4
    • /
    • pp.120-123
    • /
    • 2020
  • The surveillance system to prevent crime and accidents in advance has become a necessity, not an option in real life. Not only public institutions but also individuals are installing surveillance cameras to protect their property and privacy. However, since the installed surveillance camera cannot be monitored for 24 hours, the focus is on the technology that tracks the video after an accident occurs rather than prevention. In this paper, we propose a system model that monitors abnormal behaviors that may cause crimes through real-time video, and when a specific behavior occurs, the surveillance system automatically detects it and responds immediately through an alarm. We are a model that analyzes real-time images from surveillance cameras and uses I3D models from analysis servers to analyze abnormal behavior and deliver notifications to web servers and then to clients. If the system is implemented with the proposed model, immediate response can be expected when a crime occurs.

A Method for Improving Anchor Picture Quality of Multiview Video Coding Scheme (다시점 비디오 부호화의 기준 영상 화질 향상 방법)

  • Park, Min-Woo;Park, Jong-Tae;Park, Gwang-Hoon
    • Journal of Broadcast Engineering
    • /
    • v.13 no.3
    • /
    • pp.388-400
    • /
    • 2008
  • This paper introduces a cost-effective method fur improving anchor picture quality of a multiview video coding scheme that is mainly based on chroma compensation. Proposed method is applied to both INTER $16{\times}16$ and SKIP modes in only anchor P-pictures. By testing using JVT common test conditions, simulation results show that proposed method can obtain the average BD-PSNR gains fur U and V as 0.136 dB and 0.127 dB, respectively, while maintaining almost same performance for Y (luminance). For the range of low bit-rates, it is observed that average BD-PSNR gains of Y, U, and V are 0.141 dB, 0.494 dB and 0.525 dB, respectively. Necessary computational complexity is very marginal because the number of anchor P-pictures is only 4.18% in comparison with whole coded sequences, however it can be found that the proposed method can significantly improve the coding efficiencies of color components.

2D Adjacency Matrix Generation using DCT for UWV Contents (DCT를 통한 UWV 콘텐츠의 2D 인접도 행렬 생성)

  • Xiaorui, Li;Kim, Kyuheon
    • Journal of Broadcast Engineering
    • /
    • v.22 no.3
    • /
    • pp.366-374
    • /
    • 2017
  • Since a display device such as TV or digital signage is getting larger, the types of media is getting changed into wider view one such as UHD, panoramic and jigsaw-like media. Especially, panoramic and jigsaw-like media is realized by stitching video clips, which are captured by different camera or devices. However, a stitching process takes long time, and has difficulties in applying for a real-time process. Thus, this paper suggests to find out 2D Adjacency Matrix, which tells spatial relationships among those video clips in order to decrease a stitching processing time. Using the Discrete Cosine Transform (DCT), we convert the each frame of video source from the spatial domain (2D) into frequency domain. Based on the aforementioned features, 2D Adjacency Matrix of images could be found that we can efficiently make the spatial map of the images by using DCT. This paper proposes a new method of generating 2D adjacency matrix by using DCT for producing a panoramic and jigsaw-like media through various individual video clips.

Improved Disparity Map Computation on Stereoscopic Streaming Video with Multi-core Parallel Implementation

  • Kim, Cheong Ghil;Choi, Yong Soo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.2
    • /
    • pp.728-741
    • /
    • 2015
  • Stereo vision has become an important technical issue in the field of 3D imaging, machine vision, robotics, image analysis, and so on. The depth map extraction from stereo video is a key technology of stereoscopic 3D video requiring stereo correspondence algorithms. This is the matching process of the similarity measure for each disparity value, followed by an aggregation and optimization step. Since it requires a lot of computational power, there are significant speed-performance advantages when exploiting parallel processing available on processors. In this situation, multi-core CPU may allow many parallel programming technologies to be realized in users computing devices. This paper proposes parallel implementations for calculating disparity map using a shared memory programming and exploiting the streaming SIMD extension technology. By doing so, we can take advantage both of the hardware and software features of multi-core processor. For the performance evaluation, we implemented a parallel SAD algorithm with OpenMP and SSE2. Their processing speeds are compared with non parallel version on stereoscopic streaming video. The experimental results show that both technologies have a significant effect on the performance and achieve great improvements on processing speed.

Video Captioning with Visual and Semantic Features

  • Lee, Sujin;Kim, Incheol
    • Journal of Information Processing Systems
    • /
    • v.14 no.6
    • /
    • pp.1318-1330
    • /
    • 2018
  • Video captioning refers to the process of extracting features from a video and generating video captions using the extracted features. This paper introduces a deep neural network model and its learning method for effective video captioning. In this study, visual features as well as semantic features, which effectively express the video, are also used. The visual features of the video are extracted using convolutional neural networks, such as C3D and ResNet, while the semantic features are extracted using a semantic feature extraction network proposed in this paper. Further, an attention-based caption generation network is proposed for effective generation of video captions using the extracted features. The performance and effectiveness of the proposed model is verified through various experiments using two large-scale video benchmarks such as the Microsoft Video Description (MSVD) and the Microsoft Research Video-To-Text (MSR-VTT).