• Title/Summary/Keyword: encoded bits

Search Result 71, Processing Time 0.026 seconds

Adaptive Rate Control Based on Statistical Modeling in Frame-layer for H.264/AVC (H.264/AVC를 위한 통계 모델 기반 프레임 단위의 비트율 제어 기법)

  • Kim, Myoung-Jin;Hong, Min-Cheol
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.35 no.11C
    • /
    • pp.917-928
    • /
    • 2010
  • In this paper, we propose an efficient and adaptive rate control in Frame-layer for H.264/AVC. For given QP, bits according to video characteristics, and current frame is close correlation between the adjacent frames. Using the statistical characteristic, we obtain change of occurrence bit about QP to apply the bit amount by QP from the video characteristic and applied in the estimated bit amount of the each unit of current frame. In addition, we use weight with QP and occurrence bit amount that is statistical information of encoded previous frames. Simulation results show that the proposed rate control scheme could not only achieves time saving of more than 99% over existing rate control algorithm, but also increase the average PSNR of reconstructed video for around 0.02~0.43 dB in all the sequences.

Initial QP Determination Algorithm for Low Bit Rate Video Coding (저전송률 비디오 압축에서 초기 QP 결정 알고리즘)

  • Park, Sang-Hyun
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.13 no.10
    • /
    • pp.2071-2078
    • /
    • 2009
  • The first frame is encoded in intra mode which generates a larger number of bits. In addition, the first frame is used for the inter mode encoding of the following frames. Thus the intial QP (Quantization Parameter) for the first frame affects the first frame as well as the following frames. Traditionally, the initial QP is determined among four constant values only depending on the bpp. In the case of low bit rate video coding, the initial QP value is fixed to 35 regardless of the output bandwidth. Although this initialization scheme is simple, yet it is not accurate enough. An accurate intial QP prediction scheme should not only depends on bpp but also on the complexity of the video sequence and the output bandwidth. In the proposed scheme, we use a linear model because there is a linear inverse proportional relationship between the output bandwidth and the optimal intial QP. Model parameters of the model are determined depending on the spatial complexity of the first frame. It is shown by experimental results that the new algorithm can predict the optimal initial QP more accurately and generate the PSNR performance better than that of the existing JM algorithm.

Audio High-Band Coding based on Autoencoder with Side Information (부가 정보를 이용하는 오토 인코더 기반의 오디오 고대역 부호화 기술)

  • Cho, Hyo-Jin;Shin, Seong-Hyeon;Beack, Seung Kwon;Lee, Taejin;Park, Hochong
    • Journal of Broadcast Engineering
    • /
    • v.24 no.3
    • /
    • pp.387-394
    • /
    • 2019
  • In this study, a new method of audio high-band coding based on autoencoder with side information is proposed. The proposed method operates in the MDCT domain, and improves the performance by using additional side information consisting of the previous and current low bands, which is different from the conventional autoencoder that only inputs information to be encoded. Moreover, the side information in a time-frequency domain enables the high-band coder to utilize temporal characteristics of the signal. In the proposed method, the encoder transmits a 4-dimensional latent vector computed by the autoencoder and a gain variable using 12 bits for each frame. The decoder reconstructs the high band by applying the decoded low bands in the previous and current frames and the transmitted information to the autoencoder. Subjective evaluation confirms that the proposed method provides equivalent performance to the SBR at approximately half the bit rate of the SBR.

Analysis of Quantization Noise in Magnetic Resonance Imaging Systems (자기공명영상 시스템의 양자화잡음 분석)

  • Ahn C.B.
    • Investigative Magnetic Resonance Imaging
    • /
    • v.8 no.1
    • /
    • pp.42-49
    • /
    • 2004
  • Purpose : The quantization noise in magnetic resonance imaging (MRI) systems is analyzed. The signal-to-quantization noise ratio (SQNR) in the reconstructed image is derived from the level of quantization in the signal in spatial frequency domain. Based on the derived formula, the SQNRs in various main magnetic fields with different receiver systems are evaluated. From the evaluation, the quantization noise could be a major noise source determining overall system signal-to-noise ratio (SNR) in high field MRI system. A few methods to reduce the quantization noise are suggested. Materials and methods : In Fourier imaging methods, spin density distribution is encoded by phase and frequency encoding gradients in such a way that it becomes a distribution in the spatial frequency domain. Thus the quantization noise in the spatial frequency domain is expressed in terms of the SQNR in the reconstructed image. The validity of the derived formula is confirmed by experiments and computer simulation. Results : Using the derived formula, the SQNRs in various main magnetic fields with various receiver systems are evaluated. Since the quantization noise is proportional to the signal amplitude, yet it cannot be reduced by simple signal averaging, it could be a serious problem in high field imaging. In many receiver systems employing analog-to-digital converters (ADC) of 16 bits/sample, the quantization noise could be a major noise source limiting overall system SNR, especially in a high field imaging. Conclusion : The field strength of MRI system keeps going higher for functional imaging and spectroscopy. In high field MRI system, signal amplitude becomes larger with more susceptibility effect and wider spectral separation. Since the quantization noise is proportional to the signal amplitude, if the conversion bits of the ADCs in the receiver system are not large enough, the increase of signal amplitude may not be fully utilized for the SNR enhancement due to the increase of the quantization noise. Evaluation of the SQNR for various systems using the formula shows that the quantization noise could be a major noise source limiting overall system SNR, especially in three dimensional imaging in a high field imaging. Oversampling and off-center sampling would be an alternative solution to reduce the quantization noise without replacement of the receiver system.

  • PDF

GIS Vector Map Compression using Spatial Energy Compaction based on Bin Classification (빈 분류기반 공간에너지집중기법을 이용한 GIS 벡터맵 압축)

  • Jang, Bong-Joo;Lee, Suk-Hwan;Kwon, Ki-Ryong
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.49 no.3
    • /
    • pp.15-26
    • /
    • 2012
  • Recently, due to applicability increase of vector data based digital map for geographic information and evolution of geographic measurement techniques, large volumed GIS(geographic information service) services having high resolution and large volumed data are flowing actively. This paper proposed an efficient vector map compression technique using the SEC(spatial energy compaction) based on classified bins for the vector map having 1cm detail and hugh range. We encoded polygon and polyline that are the main objects to express geographic information in the vector map. First, we classified 3 types of bins and allocated the number of bits for each bin using adjacencies among the objects. and then about each classified bin, energy compaction and or pre-defined VLC(variable length coding) were performed according to characteristics of classified bins. Finally, for same target map, while a vector simplification algorithm had about 13%, compression ratio in 1m resolution we confirmed our method having more than 80% encoding efficiencies about original vector map in the 1cm resolution. Also it has not only higher compression ratio but also faster computing speed than present SEC based compression algorithm through experimental results. Moreover, our algorithm presented much more high performances about accuracy and computing power than vector approximation algorithm on same data volume sizes.

Frame-Layer H.264 Rate Control for Scene-Change Video at Low Bit Rate (저 비트율 장면 전환 영상에 대한 향상된 H.264 프레임 단위 데이터율 제어 알고리즘)

  • Lee, Chang-Hyun;Jung, Yun-Ho;Kim, Jae-Seok
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.44 no.11
    • /
    • pp.127-136
    • /
    • 2007
  • An abrupt scene-change frame is one that is hardly correlated with the previous frames. In that case, because an intra-coded frame has less distortion than an inter-coded one, almost all macroblocks are encoded in intra mode. This breaks up the rate control flow and increases the number of bits used. Since the reference software for H.264 takes no special action for a scene-change frame, several studies have been conducted to solve the problem using the quadratic R-D model. However, since this model is more suitable for inter frames, the existing schemes are unsuitable for computing the QP of the scene-change intra frame. In this paper, an improved rate control scheme accounting for the characteristics of intra coding is proposed for scene-change frames. The proposed scheme was validated using 16 test sequences. The results showed that the proposed scheme performed better than the existing H.264 rate control schemes. The PSNR was improved by an average of 0.4-0.6 dB and a maximum of 1.1-1.6 dB. The PSNR fluctuation was also in proved by an average of 18.6 %.

Joint video coding for multiple video program transmission based on rate-distortion estimation (다중 비디오 프로그램 전송을 위한 비트율-왜곡 추정 기반의 결합 비디오 부호화)

  • 홍성훈;김성대
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.23 no.5
    • /
    • pp.1325-1341
    • /
    • 1998
  • A conventional CBR channel is now capable of delivering several digitally compressed video programs due to recent advances in video compression, such as MPEG-2, and digital transmission technology. This paper presents a joint video coding scheme that is to maintain a constant sum of bit rates for all the programs but to allow the variable bit rate for individual program in the transimission environment mentioned above. Thus advantages of VBR video compression can be obtained. This paper contributes in two aspects. First, a rate-distortion estimation method for MPEG-2 video is proposed, which enavle us predict the amount of bits and the distortion generated from an encoded picture at a given quantization step size and vice versa. The most attractive features of the proposed rate-distortion estimation method are its accuracy and a computational complexity low enough to be applied to real-time video coding applications. Second, this paper presents an efficient and accurate joint rate control scheme using the rate-distortion estimation results and verifies its performance with experiments. The experimental results show that our coding scheme gives a significant gain even though a small number of video programs are coded jointly. For example, a stable picture quality is maintained among the video programs as well as within a program, and additional extra programs can be transmitted over the same channel bandwidth if the proposed joint video coding scheme is used.

  • PDF

A Technical Analysis on Deep Learning based Image and Video Compression (딥 러닝 기반의 이미지와 비디오 압축 기술 분석)

  • Cho, Seunghyun;Kim, Younhee;Lim, Woong;Kim, Hui Yong;Choi, Jin Soo
    • Journal of Broadcast Engineering
    • /
    • v.23 no.3
    • /
    • pp.383-394
    • /
    • 2018
  • In this paper, we investigate image and video compression techniques based on deep learning which are actively studied recently. The deep learning based image compression technique inputs an image to be compressed in the deep neural network and extracts the latent vector recurrently or all at once and encodes it. In order to increase the image compression efficiency, the neural network is learned so that the encoded latent vector can be expressed with fewer bits while the quality of the reconstructed image is enhanced. These techniques can produce images of superior quality, especially at low bit rates compared to conventional image compression techniques. On the other hand, deep learning based video compression technology takes an approach to improve performance of the coding tools employed for existing video codecs rather than directly input and process the video to be compressed. The deep neural network technologies introduced in this paper replace the in-loop filter of the latest video codec or are used as an additional post-processing filter to improve the compression efficiency by improving the quality of the reconstructed image. Likewise, deep neural network techniques applied to intra prediction and encoding are used together with the existing intra prediction tool to improve the compression efficiency by increasing the prediction accuracy or adding a new intra coding process.

Bit-Rate Control Using Histogram Based Rate-Distortion Characteristics (히스토그램 기반의 비트율-왜곡 특성을 이용한 비트율 제어)

  • 홍성훈;유상조;박수열;김성대
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.9B
    • /
    • pp.1742-1754
    • /
    • 1999
  • In this paper, we propose a rate control scheme, using histogram based rate-distortion (R-D) estimation, which produces a consistent picture quality between consecutive frames. The histogram based R-D estimation used in our rate control scheme offers a closed-form mathematical model that enable us to predict the bits and the distortion generated from an encoded frame at a given quantization parameter (QP) and vice versa. The most attractive feature of the R-D estimation is low complexity of computing the R-D data because its major operation is just to obtain a histogram or weighted histogram of DCT coefficients from an input picture. Furthermore, it is accurate enough to be applied to the practical video coding. Therefore, the proposed rate control scheme using this R-D estimation model is appropriate for the applications requiring low delay and low complexity, and controls the output bit-rate ad quality accurately. Our rate control scheme ensures that the video buffer do not underflow and overflow by satisfying the buffer constraint and, additionally, prevents quality difference between consecutive frames from exceeding certain level by adopting the distortion constraint. In addition, a consistent considering the maximum tolerance BER of the voice service. Also in Rician fading channel of K=6 and K=10, considering CLP=$10^{-3}$ as a criterion, it is observed that the performance improment of about 3.5 dB and 1.5 dB is obtained, respectively, in terms of $E_b$/$N_o$ by employing the concatenated FEC code with pilot symbols.

  • PDF

Efficient Pipeline Architecture of CABAC in H.264/AVC (H.264/AVC의 효율적인 파이프라인 구조를 적용한 CABAC 하드웨어 설계)

  • Choi, Jin-Ha;Oh, Myung-Seok;Kim, Jae-Seok
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.45 no.7
    • /
    • pp.61-68
    • /
    • 2008
  • In this paper, we propose an efficient hardware architecture and algorithm to increase an encoding process rate and implement a hardware for CABAC (Context Adaptive Binary Arithmetic Coding) which is used with one of the entropy coding ways for the latest video compression technique, H.264/AVC (Advanced Video Coding). CABAC typically provides a better high compression performance maximum 15% compared with CAVLC. However, the complexity of operation of CABAC is significantly higher than the CAVLC. Because of complicated data dependency during the encoding process, the complexity of operation is higher. Therefore, various architectures were proposed to reduce an amount of operation. However, they have still latency on account of complicated data dependency. The proposed architecture has two techniques to implement efficient pipeline architecture. The one is quick calculation of 7, 8th bits used to calculate a probability is the first step in Binary arithmetic coding. The other is one step reduced pipeline arcbitecture when the type of the encoded symbols is MPS. By adopting these two techniques, the required processing time was reduced about 27-29% compared with previous architectures. It is designed in a hardware description language and total logic gate count is 19K using 0.18um standard cell library.