• Title/Summary/Keyword: encoding complexity

Search Result 329, Processing Time 0.022 seconds

Performance Analysis of Double Binary Turbo Coded PPM-TH UWB Systems (이중 이진 터보 부호화된 펄스 위치변조-시간도약 초광대역 무선 통신 시스템의 성능 분석)

  • Kim, Eun-Cheol;Kwak, Do-Young;Park, Jae-Sung;Kim, Jin-Young
    • 한국정보통신설비학회:학술대회논문집
    • /
    • 2008.08a
    • /
    • pp.429-432
    • /
    • 2008
  • In this paper, performance of a double binary turbo coded ultra wide band (UWB) system is analyzed and simulated in an indoor wireless channel. Binary pulse position modulation-time hopping (BPPM-TH) signals are considered. The indoor wireless channel is modeled as a modified Saleh and Valenzuela (SV) channel. The performance is evaluated in terms of bit error probability (BER). From the simulation results, it is seen that double binary turbo coding offers considerable coding gain with reasonable encoding complexity. It is also demonstrated that the performance of the UWB system can be substantially improved by increasing the number of iterations.

  • PDF

Efficient distributed estimation based on non-regular quantized data

  • Kim, Yoon Hak
    • Journal of IKEEE
    • /
    • v.23 no.2
    • /
    • pp.710-715
    • /
    • 2019
  • We consider parameter estimation in distributed systems in which measurements at local nodes are quantized in a non-regular manner, where multiple codewords are mapped into a single local measurement. For the system with non-regular quantization, to ensure a perfect independent encoding at local nodes, a local measurement can be encoded into a set of a great number of codewords which are transmitted to a fusion node where estimation is conducted with enormous computational cost due to the large cardinality of the sets. In this paper, we propose an efficient estimation technique that can handle the non-regular quantized data by efficiently finding the feasible combination of codewords without searching all of the possible combinations. We conduct experiments to show that the proposed estimation performs well with respect to previous novel techniques with a reasonable complexity.

Limiting Motion Search Range for the Pseudo Video Sequence-based Light Field Image Coding (유사 비디오 시퀀스 기반의 라이트필드 영상 부호화를 위한 움직임 탐색 영역 제한)

  • Yim, Jonghoon;Duong, Vinh Van;Huu, Thuc Nguyen;Jeon, Byeungwoo
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2022.06a
    • /
    • pp.182-183
    • /
    • 2022
  • The large data volume of light field (LF) image has motivated much research on how to compress the data volume more efficiently. One of the approaches is to compress LF images after representing them in the form of pseudo video sequence. In this way, the pseudo temporal redundancy between views can be exploited by motion estimation and compensation. Based on our observation that images obtained by LF cameras have small range of disparity values between adjacent views, we propose to limit the motion search range to reduce the time complexity of motion estimation. Our experimental results show that a smaller motion search range reduces the encoding time while not affecting the bitrate of H.266/VVC much.

  • PDF

Low Complexity Motion Estimation Based on Spatio - Temporal Correlations (시간적-공간적 상관성을 이용한 저 복잡도 움직임 추정)

  • Yoon Hyo-Sun;Kim Mi-Young;Lee Guee-Sang
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.9
    • /
    • pp.1142-1149
    • /
    • 2004
  • Motion Estimation(ME) has been developed to reduce temporal redundancy in digital video signals and increase data compression ratio. ME is an Important part of video encoding systems, since it can significantly affect the output quality of encoded sequences. However, ME requires high computational complexity, it is difficult to apply to real time video transmission. for this reason, motion estimation algorithms with low computational complexity are viable solutions. In this paper, we present an efficient method with low computational complexity based on spatial and temporal correlations of motion vectors. The proposed method uses temporally and spatially correlated motion information, the motion vector of the block with the same coordinate in the reference frame and the motion vectors of neighboring blocks around the current block in the current frame, to decide the search pattern and the location of search starting point adaptively. Experiments show that the image quality improvement of the proposed method over MVFAST (Motion Vector Field Adaptive Search Technique) and PMVFAST (Predictive Motion Vector Field Adaptive Search Technique) is 0.01~0.3(dB) better and the speedup improvement is about 1.12~l.33 times faster which resulted from lower computational complexity.

Tile, Slice, and Deblocking Filter Parallelization Method in HEVC (HEVC 복호기에서의 타일, 슬라이스, 디블록킹 필터 병렬화 방법)

  • Son, Sohee;Baek, Aram;Choi, Haechul
    • Journal of Broadcast Engineering
    • /
    • v.22 no.4
    • /
    • pp.484-495
    • /
    • 2017
  • The development of display devices and the increase of network transmission bandwidth bring demands for over 2K high resolution video such as panorama video, 4K ultra-high definition commercial broadcasting, and ultra-wide viewing video. To compress these image sequences with significant amount of data, High Efficiency Video Coding (HEVC) standard with the highest coding efficiency is a promising solution. HEVC, the latest video coding standard, provides high encoding efficiency using various advanced encoding tools, but it also requires significant amounts of computation complexity compared to previous coding standards. In particular, the complexity of HEVC decoding process is a imposing challenges on real-time playback of ultra-high resolution video. To accelerate the HEVC decoding process for ultra high resolution video, this paper introduces a data-level parallel video decoding method using slice and/or tile supported by HEVC. Moreover, deblocking filter process is further parallelized. The proposed method distributes independent decoding operations of each tile and/or each slice to multiple threads as well as deblocking filter operations. The experimental results show that the proposed method facilitates executions up to 2.0 times faster than the HEVC reference software for 4K videos.

On the Performance of Sample-Adaptive Product Quantizer for Noisy Channels (표본적응 프러덕트 양자기의 전송로 잡음에서의 성능 분석에 관한 연구)

  • Kim Dong Sik
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.42 no.3 s.303
    • /
    • pp.81-90
    • /
    • 2005
  • When we transmit signals, which are quantized by the vector quantizer (VQ), through noisy channels, the overall performance of the coding system is very dependent on the employed quantization scheme and the channel error effect. In order to design an optimal coding system, the source and channel coding scheme should be jointly optimized as in the channel-optimized VQ. As a suboptimal approach, we may consider the robust VQ (RVQ). In RVQ, we consider developing an index assignment function for mapping the output of quantizers to channel symbols so that the effect of the channel errors is minimized. Recently, a VQ, which can reduce the encoding complexity and is called the sample-adaptive product quantizer (SAPQ), has been proposed. SAPQ has very similar quantizer structure as to the product quantizer (PQ). However, the quantization performance can be better than PQ. Further, the encoding complexity and the memory requirement for the codebooks are lower than the regular full-search VQ case. In this paper, SAPQ is employed in order to design an RVQ to channel errors by reducing the vector dimension. Discussions on the codebook structure of SAPQ and experiments are introduced in an aspect of robustness to noisy channels.

Efficient Pipeline Architecture of CABAC in H.264/AVC (H.264/AVC의 효율적인 파이프라인 구조를 적용한 CABAC 하드웨어 설계)

  • Choi, Jin-Ha;Oh, Myung-Seok;Kim, Jae-Seok
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.45 no.7
    • /
    • pp.61-68
    • /
    • 2008
  • In this paper, we propose an efficient hardware architecture and algorithm to increase an encoding process rate and implement a hardware for CABAC (Context Adaptive Binary Arithmetic Coding) which is used with one of the entropy coding ways for the latest video compression technique, H.264/AVC (Advanced Video Coding). CABAC typically provides a better high compression performance maximum 15% compared with CAVLC. However, the complexity of operation of CABAC is significantly higher than the CAVLC. Because of complicated data dependency during the encoding process, the complexity of operation is higher. Therefore, various architectures were proposed to reduce an amount of operation. However, they have still latency on account of complicated data dependency. The proposed architecture has two techniques to implement efficient pipeline architecture. The one is quick calculation of 7, 8th bits used to calculate a probability is the first step in Binary arithmetic coding. The other is one step reduced pipeline arcbitecture when the type of the encoded symbols is MPS. By adopting these two techniques, the required processing time was reduced about 27-29% compared with previous architectures. It is designed in a hardware description language and total logic gate count is 19K using 0.18um standard cell library.

An Effective Error-Concealment Approach for Video Data Transmission over Internet (인터넷상의 비디오 데이타 전송에 효과적인 오류 은닉 기법)

  • 김진옥
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.6
    • /
    • pp.736-745
    • /
    • 2002
  • In network delivery of compressed video, packets may be lost if the channel is unreliable like Internet. Such losses tend to of cur in burst like continuous bit-stream error. In this paper, we propose an effective error-concealment approach to which an error resilient video encoding approach is applied against burst errors and which reduces a complexity of error concealment at the decoder using data hiding. To improve the performance of error concealment, a temporal and spatial error resilient video encoding approach at encoder is developed to be robust against burst errors. For spatial area of error concealment, block shuffling scheme is introduced to isolate erroneous blocks caused by packet losses. For temporal area of error concealment, we embed parity bits in content data for motion vectors between intra frames or continuous inter frames and recovery loss packet with it at decoder after transmission While error concealment is performed on error blocks of video data at decoder, it is computationally costly to interpolate error video block using neighboring information. So, in this paper, a set of feature are extracted at the encoder and embedded imperceptibly into the original media. If some part of the media data is damaged during transmission, the embedded features can be extracted and used for recovery of lost data with bi-direction interpolation. The use of data hiding leads to reduced complexity at the decoder. Experimental results suggest that our approach can achieve a reasonable quality for packet loss up to 30% over a wide range of video materials.

Lightweight video coding using spatial correlation and symbol-level error-correction channel code (공간적 유사성과 심볼단위 오류정정 채널 코드를 이용한 경량화 비디오 부호화 방법)

  • Ko, Bong-Hyuck;Shim, Hiuk-Jae;Jeon, Byeung-Woo
    • Journal of Broadcast Engineering
    • /
    • v.13 no.2
    • /
    • pp.188-199
    • /
    • 2008
  • In conventional video coding, encoder complexity is much higher than that of decoder. However, investigations for lightweight encoder to eliminate motion prediction/compensation claiming most complexity in encoder have recently become an important issue. The Wyner-Ziv coding is one of the representative schemes for the problem and, in this scheme, since encoder generates only parity bits of a current frame without performing any type of processes extracting correlation information between frames, it has an extremely simple structure compared to conventional coding techniques. However, in Wyner-Ziv coding, channel decoding errors occur when noisy side information is used in channel decoding process. These channel decoding errors appear more frequently, especially, when there is not enough correlation between frames to generate accurate side information and, as a result, those errors look like Salt & Pepper type noise in the reconstructed frame. Since this noise severely deteriorates subjective video quality even though such noise rarely occurs, previously we proposed a computationally extremely light encoding method based on selective median filter that corrects such noise using spatial correlation of a frame. However, in the previous method, there is a problem that loss of texture from filtering may exceed gain from error correction by the filter for video sequences having complex torture. Therefore, in this paper, we propose an improved lightweight encoding method that minimizes loss of texture detail from filtering by allowing information of texture and that of noise in side information to be utilized by the selective median filter. Our experiments have verified average PSNR gain of up to 0.84dB compared to the previous method.

Fast Partition Decision Using Rotation Forest for Intra-Frame Coding in HEVC Screen Content Coding Extension (회전 포레스트 분류기법을 이용한 HEVC 스크린 콘텐츠 화면 내 부호화 조기분할 결정 방법)

  • Heo, Jeonghwan;Jeong, Jechang
    • Journal of Broadcast Engineering
    • /
    • v.23 no.1
    • /
    • pp.115-125
    • /
    • 2018
  • This paper presents a fast partition decision framework for High Efficiency Video Coding (HEVC) Screen Content Coding (SCC) based on machine learning. Currently, the HEVC performs quad-tree block partitioning process to achieve optimal coding efficiency. Since this process requires a high computational complexity of the encoding device, the fast encoding process has been studied as determining the block structure early. However, in the case of the screen content video coding, it is difficult to apply the conventional early partition decision method because it shows different partition characteristics from natural content. The proposed method solves the problem by classifying the screen content blocks after partition decision, and it shows an increase of 3.11% BD-BR and 42% time reduction compared to the SCC common test condition.