• Title/Summary/Keyword: group decoding

Search Result 84, Processing Time 0.026 seconds

Quasi-Orthogonal STBC with Iterative Decoding in Bit Interleaved Coded Modulation

  • Sung, Chang-Kyung;Kim, Ji-Hoon;Lee, In-Kyu
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.33 no.4A
    • /
    • pp.426-433
    • /
    • 2008
  • In this paper, we present a method to improve the performance of the four transmit antenna quasi-orthogonal space-time block code (STBC) in the coded system. For the four transmit antenna case, the quasi-orthogonal STBC consists of two symbol groups which are orthogonal to each other, but intra group symbols are not. In uncoded system with the matched filter detection, constellation rotation can improve the performance. However, in coded systems, its gain is absorbed by the coding gain especially for lower rate code. We propose an iterative decoding method to improve the performance of quasi-orthogonal codes in coded systems. With conventional quasi-orthogonal STBC detection, the joint ML detection can be improved by iterative processing between the demapper and the decoder. Simulation results shows that the performance improvement is about 2dB at 1% frame error rate.

Multi-Sever based Distributed Coding based on HEVC/H.265 for Studio Quality Video Editing

  • Kim, Jongho;Lim, Sung-Chang;Jeong, Se-Yoon;Kim, Hui-Yong
    • Journal of Multimedia Information System
    • /
    • v.5 no.3
    • /
    • pp.201-208
    • /
    • 2018
  • High Efficiency Video Coding range extensions (HEVC RExt) is a kind of extension model of HEVC. HEVC RExt was specially designed for dealing the high quality images. HEVC RExt is very essential for studio editing which handle the very high quality and various type of images. There are some problems to dealing these massive data in studio editing. One of the most important procedure is re-encoding and decoding procedure during the editing. Various codecs are widely used for studio data editing. But most of the codecs have common problems to dealing the massive data in studio editing. First, the re-encoding and decoding processes are frequently occurred during the studio data editing and it brings enormous time-consuming and video quality loss. This paper, we suggest new video coding structure for the efficient studio video editing. The coding structure which is called "ultra-low delay (ULD)". It has the very simple and low-delayed referencing structure. To simplify the referencing structure, we can minimize the number of the frames which need decoding and re-encoding process. It also prevents the quality degradation caused by the frequent re-encoding. Various fast coding algorithms are also proposed for efficient editing such as tool-level optimization, multi-serve based distributed coding and SIMD (Single instruction, multiple data) based parallel processing. It can reduce the enormous computational complexity during the editing procedure. The proposed method shows 9500 times faster coding speed with negligible loss of quality. The proposed method also shows better coding gain compare to "intra only" structure. We can confirm that the proposed method can solve the existing problems of the studio video editing efficiently.

Complexity Reduction of Block-Layered QOSTC with Less Transmission Time (복잡도 감소와 전송시간이 덜 소요되는 블록 층의 준 직교 시공간코드 설계)

  • Abu Hanif, Mohammad;Lee, Moon-Ho;Hai, Han
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.49 no.7
    • /
    • pp.48-55
    • /
    • 2012
  • Because of increasing complexity in maximum-likelihood (ML) decoding of four of higher antenna scenario, Partial Interference Cancellation (PIC) group decoding could be the perfect solution to reduce the decoding complexity occurs in ML decoding. In this paper, we separate the symbols the users in the layered basis and find the equivalent channel matrix. Based on the equivalent channel matrix we provide the grouping scheme. In our paper, we construct a block wise transmission technique which will achieve the desired code rate and reduce the complexity and provide less transmission time. Finally we show the different grouping performance.

Generalized Sub-optimum Decoding for Space-Time Trellis Codes in Quasistatic Flat Fading Channel (준정적 플랫 페이딩 채널에서 시공간 트렐리스 부호의 일반화된 부최적 복호법)

  • Kim Young Ju;Shin Sang Sup;Kang Hyun-Soo
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.43 no.1 s.343
    • /
    • pp.89-94
    • /
    • 2006
  • We present a generalized version of principal ratio combining (PRC)[1], which is a near-optimum decoding scheme for space-time trellis codes in quasi-static flat fading environments. In [1], the performance penalty increases as the number of receive antennas increases. In the proposed scheme, receive antennas are divided into K groups, and the PRC decoding method is applied to each group. This shows a flexible tradeoff between performance and decoding complexity by choosing the appropriate K. Moreover, we also propose the performance index(PI) to easily predict the decoding performance among the possible different(receive antenna) configurations.

Performance Analysis of SOVA by Robust Equalization, Techniques in Nongaussian Noise Channel (비가우시안 잡음 채널에서 Robust 등화기법을 이용한 터보 부호의 SOVA 성능분석)

  • Soh, Surng-Ryurl;Lee, Chang-Bum;Kim, Yung-Kwon;Chung, Boo-Young
    • Journal of IKEEE
    • /
    • v.4 no.2 s.7
    • /
    • pp.257-265
    • /
    • 2000
  • Turbo Code decoder is an iterate decoding technology, which extracts extrinsic information from the bit to be decoded by calculating both forward and backward metrics in each decoding step, and uses the information to the next decoding step. Viterbi decoder, which is for a convolutional code, runs continuous mode, while Turbo Code decoder runs by block unit. There are algorithms used in a decoder : which are MAP(maximum a posteriori) algorithm requiring very complicated calculation and SOVA(soft output Viterbi algorithm) using Viterbi algorithm suggested by Hagenauer, and it is known that the decoding performance of MAP is better. The result of this make experimentation shows that the performance of SOVA, which has half complex algorithm compare to MAP, is almost same as the performance of MAP when the SOVA decoding performance is supplemented with Robust equalization techniques.

  • PDF

Efficient Local Decoding Using Bit Stream Map for High Resolution Video (비트 스트림 지도를 이용한 고해상도 영상의 효율적인 지역복호화)

  • Park Sungwon;Won Jongwoo;Lee Sunyoung;Kim Wookjoong;Kim Kyuheon;Jang Euee S
    • Journal of Broadcast Engineering
    • /
    • v.9 no.4 s.25
    • /
    • pp.391-401
    • /
    • 2004
  • In this paper, we introduce a novel coding method to efficiently enable spatial random access for high resolution video. In terms of resolution and display size, standard display devices (such as cathode-ray tubes. monitors. PDAs, and LCDs) do not sufficiently support high resolution video such as digital cinema and panoramic video. Currently, users have no choice but to view video at lower resolution as a result of down-sampling, or only a partial region of the video due to display size limitations. Our proposed method. which we call the B-map, represents the set of starting locations of the coded segments in a picture frame. This information, or B-map, is first sent to the decoder prior to the coded data stream of the frame and is then used for fast local decoding. To test our method, we compare our B-map with JPEG tiling and the JPEG Resynchronization marker. Experimental results show that the proposed coding method requires less overhead than existing methods during the same decoding time. The results show promise for future panoramic or digital cinema applications.

Complexity Analysis of Internet Video Coding (IVC) Decoding

  • Park, Sang-hyo;Dong, Tianyu;Jang, Euee S.
    • Journal of Multimedia Information System
    • /
    • v.4 no.4
    • /
    • pp.179-188
    • /
    • 2017
  • The Internet Video Coding (IVC) standard is due to be published by Moving Picture Experts Group (MPEG) for various Internet applications such as internet broadcast streaming. IVC aims at three things fundamentally: 1) forming IVC patents under a free of charge license, 2) reaching comparable compression performance to AVC/H.264 constrained Baseline Profile (cBP), and 3) maintaining computational complexity for feasible implementation of real-time encoding and decoding. MPEG experts have worked diligently on the intellectual property rights issues for IVC, and they reported that IVC already achieved the second goal (compression performance) and even showed comparable performance to even AVC/H.264 High Profile (HP). For the complexity issue, however, there has not been thorough analysis on IVC decoder. In this paper, we analyze the IVC decoder in view of the time complexity by evaluating running time. Through the experimental results, IVC is 3.6 times and 3.1 times more complex than AVC/H.264 cBP under constrained set (CS) 1 and CS2, respectively. Compared to AVC/H.264 HP, IVC is 2.8 times and 2.9 times slower in decoding time under CS1 and CS2, respectively. The most critical tool to be improved for lightweight IVC decoder is motion compensation process containing a resolution-adaptive interpolation filtering process.

Low Computational Complexity LDPC Decoding Algorithms for DVB-S2 Systems (DVB-S2 시스템을 위한 저복잡도 LDPC 복호 알고리즘)

  • Jung Ji-Won
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.16 no.10 s.101
    • /
    • pp.965-972
    • /
    • 2005
  • In this paper, we first review LDPC codes in general and a belief propagation algorithm that works in logarithm domain. LDPC codes, which is chosen for second generation digital video broadcasting standard, are required a large number of computation due to large size of coded block and iteration. Therefore, we presented two kinds of low computational algorithm for LDPC codes. First, sequential decoding with partial group is proposed. It has same H/W complexity, and fewer number of iteration's are required at same performance in comparison with conventional decoder algerian. Secondly, early detection method for reducing the computational complexity is proposed. Using a confidence criterion, some bit nodes and check node edges are detected early on during decoding. Through the simulation, we knew that the iteration number are reduced by half using subset algorithm and computational complexity of early detected method is about $50\%$ offs in case of check node update, $99\%$ offs in case of check node update compared to conventional scheme.

Low Computational Complexity LDPC Decoding Algorithms for 802.11n Standard (802.11n 규격에서의 저복잡도 LDPC 복호 알고리즘)

  • Kim, Min-Hyuk;Park, Tae-Doo;Jung, Ji-Won;Lee, Seong-Ro;Jung, Min-A
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.35 no.2C
    • /
    • pp.148-154
    • /
    • 2010
  • In this paper, we first review LDPC codes in general and a belief propagation algorithm that works in logarithm domain. LDPC codes, which is chosen 802.11n for wireless local access network(WLAN) standard are required a large number of computation due to large size of coded block and iteration. Therefore, we presented three kinds of low computational algorithm for LDPC codes. First, sequential decoding with partial group is proposed. It has same H/W complexity, and fewer number of iteration's are required at same performance in comparison with conventional decoder algorithm. Secondly, we have apply early stop algorithm. This method is reduced number of unnecessary iteration. Third, early detection method for reducing the computational complexity is proposed. Using a confidence criterion, some bit nodes and check node edges are detected early on during decoding. Through the simulation, we knew that the iteration number are reduced by half using subset algorithm and early stop algorithm is reduced more than one iteration and computational complexity of early detected method is about 30% offs in case of check node update, 94% offs in case of check node update compared to conventional scheme.