• 제목/요약/키워드: Inter-coding

검색결과 318건 처리시간 0.026초

Efficient Inter Prediction Mode Decision Method for Fast Motion Estimation in High Efficiency Video Coding

  • Lee, Alex;Jun, Dongsan;Kim, Jongho;Choi, Jin Soo;Kim, Jinwoong
    • ETRI Journal
    • /
    • 제36권4호
    • /
    • pp.528-536
    • /
    • 2014
  • High Efficiency Video Coding (HEVC) is the most recent video coding standard to achieve a higher coding performance than the previous H.264/AVC. In order to accomplish this improved coding performance, HEVC adopted several advanced coding tools; however, these cause heavy computational complexity. Similar to previous video coding standards, motion estimation (ME) of HEVC requires the most computational complexity; this is because ME is conducted for three inter prediction modes - namely, uniprediction in list 0, uniprediction in list 1, and biprediction. In this paper, we propose an efficient inter prediction mode (EIPM) decision method to reduce the complexity of ME. The proposed EIPM method computes the priority of all inter prediction modes and performs ME only on a selected inter prediction mode. Experimental results show that the proposed method reduces computational complexity arising from ME by up to 51.76% and achieves near similar coding performance compared to HEVC test model version 10.1.

비디오 압축을 위한 영상간 차분 DCT 계수의 문맥값 기반 부호화 방법 (Context-based coding of inter-frame DCT coefficients for video compression)

  • Lee, Jin-Hak;Kim, Jae-Kyoon
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2000년도 제13회 신호처리 합동 학술대회 논문집
    • /
    • pp.281-285
    • /
    • 2000
  • This paper proposes context-based coding methods for variable length coding of inter-frame DCT coefficients. The proposed methods classify run-level symbols depending on the preceding coefficients. No extra overhead needs to be transmitted, since the information of the previously transmitted coefficients is used for classification. Two entropy coding methods, arithmetic coding and Huffman coding, are used for the proposed context-based coding. For Huffman coding, there is no complexity increase from the current standards by using the existing inter/intra VLC tables. Experimental results show that the proposed methods give ~ 19% bits gain and ~ 0.8 dB PSNR improvement for adaptive inter/intra VLC table selection, and ~ 37% bits gain and ~ 2.7dB PSNR improvement for arithmetic coding over the current standards, MPEG-4 and H.263. Also, the proposed methods obtain larger gain for small quantizaton parameters and the sequences with fast and complex motion. Therefore, for high quality video coding, the proposed methods have more advantage.

  • PDF

Inter-layer Texture and Syntax Prediction for Scalable Video Coding

  • Lim, Woong;Choi, Hyomin;Nam, Junghak;Sim, Donggyu
    • IEIE Transactions on Smart Processing and Computing
    • /
    • 제4권6호
    • /
    • pp.422-433
    • /
    • 2015
  • In this paper, we demonstrate inter-layer prediction tools for scalable video coders. The proposed scalable coder is designed to support not only spatial, quality and temporal scalabilities, but also view scalability. In addition, we propose quad-tree inter-layer prediction tools to improve coding efficiency at enhancement layers. The proposed inter-layer prediction tools generate texture prediction signal with exploiting texture, syntaxes, and residual information from a reference layer. Furthermore, the tools can be used with inter and intra prediction blocks within a large coding unit. The proposed framework guarantees the rate distortion performance for a base layer because it does not have any compulsion such as constraint intra prediction. According to experiments, the framework supports the spatial scalable functionality with about 18.6%, 18.5% and 25.2% overhead bits against to the single layer coding. The proposed inter-layer prediction tool in multi-loop decoding design framework enables to achieve coding gains of 14.0%, 5.1%, and 12.1% in BD-Bitrate at the enhancement layer, compared to a single layer HEVC for all-intra, low-delay, and random access cases, respectively. For the single-loop decoding design, the proposed quad-tree inter-layer prediction can achieve 14.0%, 3.7%, and 9.8% bit saving.

Scalable Multi-view Video Coding based on HEVC

  • Lim, Woong;Nam, Junghak;Sim, Donggyu
    • IEIE Transactions on Smart Processing and Computing
    • /
    • 제4권6호
    • /
    • pp.434-442
    • /
    • 2015
  • In this paper, we propose an integrated spatial and view scalable video codec based on high efficiency video coding (HEVC). The proposed video codec is developed based on similarity and uniqueness between the scalable extension and 3D multi-view extension of HEVC. To improve compression efficiency using the proposed scalable multi-view video codec, inter-layer and inter-view predictions are jointly employed by using high-level syntaxes that are defined to identify view and layer information. For the inter-view and inter-layer predictions, a decoded picture buffer (DPB) management algorithm is also proposed. The inter-view and inter-layer motion predictions are integrated into a consolidated prediction by harmonizing with the temporal motion prediction of HEVC. We found that the proposed scalable multi-view codec achieves bitrate reduction of 36.1%, 31.6% and 15.8% on the top of ${\times}2$, ${\times}1.5$ parallel scalable codec and parallel multi-view codec, respectively.

Efficient Mode Decision Algorithm Based on Spatial, Temporal, and Inter-layer Rate-Distortion Correlation Coefficients for Scalable Video Coding

  • Wang, Po-Chun;Li, Gwo-Long;Huang, Shu-Fen;Chen, Mei-Juan;Lin, Shih-Chien
    • ETRI Journal
    • /
    • 제32권4호
    • /
    • pp.577-587
    • /
    • 2010
  • The layered coding structure of scalable video coding (SVC) with adaptive inter-layer prediction causes noticeable computational complexity increments when compared to existing video coding standards. To lighten the computational complexity of SVC, we present a fast algorithm to speed up the inter-mode decision process. The proposed algorithm terminates inter-mode decision early in the enhancement layers by estimating the rate-distortion (RD) cost from the macroblocks of the base layer and the enhancement layer in temporal, spatial, and inter-layer directions. Moreover, a search range decision algorithm is also proposed in this paper to further increase the motion estimation speed by using the motion vector information from temporal, spatial, or inter-layer domains. Simulation results show that the proposed algorithm can determine the best mode and provide more efficient total coding time saving with very slight RD performance degradation for spatial and quality scalabilities.

인접블록의 상관관계에 기반한 RGB video coding 개선 알고리즘 (Enhanced RGB Video Coding Based on Correlation in the Adjacent Block)

  • 김양수;정진우;최윤식
    • 전기학회논문지
    • /
    • 제58권12호
    • /
    • pp.2538-2541
    • /
    • 2009
  • H.264/AVC High 4:4:4 Intra/Predictive profiles supports RGB 4:4:4 sequences for high fidelity video. RGB color planes rather than YCbCr color planes are preferred by high-fidelity video applications such as digital cinema, medical imaging, and UHDTV. Several RGB coding tools have therefore been developed to improve the coding efficiency of RGB video. In this paper, we propose a new method to extract more accurate correlation parameters for inter-plane prediction. We use a searching method to determine the matched macroblock (MB) that has a similar inter-color relation to the current MB. Using this block, we can infer more accurate correlation parameters to predict chroma MB from luma MB. Our proposed inter-plane prediction mode shows an average bits saving of 15.6% and a PSNR increase of 0.99 dB compared with H.264 high4:4:4 intra-profile RGB coding. Furthermore, extensive performance evaluation revealed that our proposed algorithm has better coding efficiency than existing algorithms..

H.264 기반 스케일러블 비디오 부호화에서 인트라 블럭에 대한 적응적인 계층간 예측 연구 (Adaptive Inter-Layer Prediction for Intra Texture on H.264 Scalable Video Coding)

  • 오형석;박성호;천민수;김원하
    • 대한전기학회:학술대회논문집
    • /
    • 대한전기학회 2005년도 학술대회 논문집 정보 및 제어부문
    • /
    • pp.195-197
    • /
    • 2005
  • In the scalable extension of H.264/AVC, spatial scalability is provided residual information as encoding layered spatial resolution between layers. We use the inter-layer prediction to remove this redundancy. In the inter-layer prediction, as the prediction we can use the signal that is the upsampled signal of the lower resolution layer. In this case, coding efficiency can be different from optimal prediction by kinds of interpolation filter. This paper indicates technique to choose the interpolation filter and to enhance coding efficiency for finding more correct prediction in intra macroblock.

  • PDF

멀티채널 Spatial Audio Coding에서의 효율적인 Spatial Cues 사용과 그에 따른 Spectrum 보상방법 (An efficient method of spatial cues and compensation method of spectrums on multichannel spatial audio coding)

  • 이병화;백승권;서정일;한민수
    • 대한음성학회지:말소리
    • /
    • 제53호
    • /
    • pp.157-169
    • /
    • 2005
  • This paper proposes an efficiently representing method of spatial cues on multichannel spatial audio coding. The Binaural Cue Coding (BCC) method introduced recently represents multichannel audio signals by means of Inter Channel Level Difference (ICLD) or Source Index (SI). We tried to express more efficiently ICLD and SI information based on Inter Channel Correlation in this paper. We adopt different spatial cues according to ICC and propose a compensation method of empty spectrums created by using SI. We performed a MOS test and measuring spectral distortion. The results show that the proposed method can reduce the bitrate of side information without large degradation of the audio quality.

  • PDF

Global Disparity Compensation for Multi-view Video Coding

  • ;호요성
    • 방송공학회논문지
    • /
    • 제12권6호
    • /
    • pp.624-629
    • /
    • 2007
  • While single view video coding uses the temporal prediction scheme, multi-view video coding (MVC) applies both temporal and inter-view prediction schemes. Thus, the key problem of MVC is how to reduce the inter-view redundancy efficiently, because various existing video coding schemes have already provided solutions to reduce the temporal correlation. In this paper, we propose a global disparity compensation scheme which increases the inter-view correlation and a new inter-view prediction structure based on the global disparity compensation. By experiment, we demonstrate that the proposed global disparity compensation scheme is less sensitive to change of the search range. In addition, the new Inter-view prediction structure achieved about $0.1{\sim}0.3dB$ quality improvement compared to the reference software.

Bayesian-theory-based Fast CU Size and Mode Decision Algorithm for 3D-HEVC Depth Video Inter-coding

  • Chen, Fen;Liu, Sheng;Peng, Zongju;Hu, Qingqing;Jiang, Gangyi;Yu, Mei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제12권4호
    • /
    • pp.1730-1747
    • /
    • 2018
  • Multi-view video plus depth (MVD) is a mainstream format of 3D scene representation in free viewpoint video systems. The advanced 3D extension of the high efficiency video coding (3D-HEVC) standard introduces new prediction tools to improve the coding performance of depth video. However, the depth video in 3D-HEVC is time consuming. To reduce the complexity of the depth video inter coding, we propose a fast coding unit (CU) size and mode decision algorithm. First, an off-line trained Bayesian model is built which the feature vector contains the depth levels of the corresponding spatial, temporal, and inter-component (texture-depth) neighboring largest CUs (LCUs). Then, the model is used to predict the depth level of the current LCU, and terminate the CU recursive splitting process. Finally, the CU mode search process is early terminated by making use of the mode correlation of spatial, inter-component (texture-depth), and inter-view neighboring CUs. Compared to the 3D-HEVC reference software HTM-10.0, the proposed algorithm reduces the encoding time of depth video and the total encoding time by 65.03% and 41.04% on average, respectively, with negligible quality degradation of the synthesized virtual view.