• Title/Summary/Keyword: H.264 video decoder

Search Result 94, Processing Time 0.019 seconds

Dynamic Full-Scalability-Conversion in SVC (스케일러블 비디오 코딩에서의 실시간 스케일러빌리티 변환)

  • Lee, Dong-Su;Bae, Tae-Meon;Ro, Yong-Man
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.43 no.6 s.312
    • /
    • pp.60-70
    • /
    • 2006
  • Currently, Scalable Video Coding (SVC) is being standardized. By using scalability of SVC, QoS managed video streaming service is enabled in heterogeneous networks even with only one original bitstream. But current SVC is insufficient to dynamic video conversion for the scalability, thereby the adaptation of bitrate to meet a fluctuating network condition is limited. In this paper, we propose dynamic full-scalability conversion method for QoS adaptive video streaming in H.264/AVC SVC. To accomplish full scalability dynamic conversion, we propose corresponding bitstream extraction, encoding and decoding schemes. On the encoder, we newly insert the IDR NAL to solve the problems of spatial scalability conversion. On the extractor, we analyze the SVC bitstream to get the information which enable dynamic extraction. By using this information, real time extraction is achieved. Finally, we develop the decoder so that it can manage changing bitrate to support real time full-scalability. The experimental results showed that dynamic full-scalability conversion was verified and it was necessary for time varying network condition.

Real-time Stereo Video Generation using Graphics Processing Unit (GPU를 이용한 실시간 양안식 영상 생성 방법)

  • Shin, In-Yong;Ho, Yo-Sung
    • Journal of Broadcast Engineering
    • /
    • v.16 no.4
    • /
    • pp.596-601
    • /
    • 2011
  • In this paper, we propose a fast depth-image-based rendering method to generate a virtual view image in real-time using a graphic processor unit (GPU) for a 3D broadcasting system. Before the transmission, we encode the input 2D+depth video using the H.264 coding standard. At the receiver, we decode the received bitstream and generate a stereo video using a GPU which can compute in parallel. In this paper, we apply a simple and efficient hole filling method to reduce the decoder complexity and reduce hole filling errors. Besides, we design a vertical parallel structure for a forward mapping process to take advantage of the single instruction multiple thread structure of GPU. We also utilize high speed GPU memories to boost the computation speed. As a result, we can generate virtual view images 15 times faster than the case of CPU-based processing.

Efficient Coding of Motion Vector Predictor using Phased-in Code (Phased-in 코드를 이용한 움직임 벡터 예측기의 효율적인 부호화 방법)

  • Moon, Ji-Hee;Choi, Jung-Ah;Ho, Yo-Sung
    • Journal of Broadcast Engineering
    • /
    • v.15 no.3
    • /
    • pp.426-433
    • /
    • 2010
  • The H.264/AVC video coding standard performs inter prediction using variable block sizes to improve coding efficiency. Since we predict not only the motion of homogeneous regions but also the motion of non-homogeneous regions accurately using variable block sizes, we can reduce residual information effectively. However, each motion vector should be transmitted to the decoder. In low bit rate environments, motion vector information takes approximately 40% of the total bitstream. Thus, motion vector competition was proposed to reduce the amount of motion vector information. Since the size of the motion vector difference is reduced by motion vector competition, it requires only a small number of bits for motion vector information. However, we need to send the corresponding index of the best motion vector predictor for decoding. In this paper, we propose a new codeword table based on the phased-in code to encode the index of motion vector predictor efficiently. Experimental results show that the proposed algorithm reduces the average bit rate by 7.24% for similar PSNR values, and it improves the average image quality by 0.36dB at similar bit rates.

Improved AR-FGS Coding Scheme for Scalable Video Coding (확장형 비디오 부호화(SVC)의 AR-FGS 기법에 대한 부호화 성능 개선 기법)

  • Seo, Kwang-Deok;Jung, Soon-Heung;Kim, Jin-Soo;Kim, Jae-Gon
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.31 no.12C
    • /
    • pp.1173-1183
    • /
    • 2006
  • In this paper, we propose an efficient method for improving visual quality of AR-FGS (Adaptive Reference FGS) which is adopted as a key scheme for SVC (Scalable Video Coding) or H.264 scalable extension. The standard FGS (Fine Granularity Scalability) adopts AR-FGS that introduces temporal prediction into FGS layer by using a high quality reference signal which is constructed by the weighted average between the base layer reconstructed imageand enhancement reference to improve the coding efficiency in the FGS layer. However, when the enhancement stream is truncated at certain bitstream position in transmission, the rest of the data of the FGS layer will not be available at the FGS decoder. Thus the most noticeable problem of using the enhancement layer in prediction is the degraded visual quality caused by drifting because of the mismatch between the reference frame used by the FGS encoder and that by the decoder. To solve this problem, we exploit the principle of cyclical block coding that is used to encode quantized transform coefficients in a cyclical manner in the FGS layer. Encoding block coefficients in a cyclical manner places 'higher-value' bits earlier in the bitstream. The quantized transform coefficients included in the ealry coding cycle of cyclical block coding have higher probability to be correctly received and decoded than the others included in the later cycle of the cyclical block coding. Therefore, we can minimize visual quality degradation caused by bitstream truncation by adjusting weighting factor to control the contribution of the bitstream produced in each coding cycle of cyclical block coding when constructing the enhancement layer reference frame. It is shown by simulations that the improved AR-FGS scheme outperforms the standard AR-FGS by about 1 dB in maximum in the reconstructed visual quality.