• Title/Summary/Keyword: Entropy encoding

Search Result 47, Processing Time 0.026 seconds

A Study on the Memory Saturation Prevention of the Entropy Encoder for He HDTV (HDTV용 엔트로피 부호화기의 메모리 포화 방지에 관한 연구)

  • 이선근;임순자;김환용
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.5A
    • /
    • pp.545-553
    • /
    • 2004
  • Expansion of network environment and multimedia demand universality of application service as HDTV, etc. During these processes, it is essential to process multimedia in real time in the wireless communication system based on mobile phone network and in the wire communication system due to fiber cable and xDSL. So, in this Paper the optimal memory allocation algorithm combines the merit of huffman encoding which is superior in simultaneous decoding ability and lempel-ziv that is distinguished in execution of compress is proposed to improve the channel transmission rate and processing speed in the compressing procedure and is verified in the entropy encoder of HDTV. Because the entropy encoder system using proposed optimal memory allocation algorithm has memory saturation prevention we confirms that the compressing ratio for moving pictures is superior than Huffman encoding and LZW.

Non-fixed Quantization Considering Entropy Encoding in HEVC (HEVC 엔트로피 부호화를 고려한 비균등 양자화 방법)

  • Gweon, Ryeong-Hee;Han, Woo-Jin;Lee, Yung-Lyul
    • Journal of Broadcast Engineering
    • /
    • v.16 no.6
    • /
    • pp.1036-1046
    • /
    • 2011
  • MPEG and VCEG have constituted a collaboration team called JCT-VC(Joint Collaborative Team on Video Coding) and have been developing HEVC(High Efficiency Video Coding) standard. All transform coefficients in a TU(Transform Unit) have been equally quantized according to the quantization and inverse quantization method which is used in HEVC standard. Such an equal quantization is not efficient because the transformed coefficients in the TU are not eqully distributed. Furthermore, the quantized coefficients which is positioned in later scanning order cannot be efficient due to the entropy scanning method. We suggest an algorithm that transform coefficients are quantized at different values according to the position in TU considering a scanning order of entropy encoding to improve the coding efficiency. The principle of this algorithm is that quantization and inverse quantization are carried out according to the scanning order which is in accordance with the statistical characteristic of distribution of quantized transform coefficients. The proposed algorithm shows on the average of 0.34% Y BD-rate compression rate improvement.

Nonlinear Quality Indices Based on a Novel Lempel-Ziv Complexity for Assessing Quality of Multi-Lead ECGs Collected in Real Time

  • Zhang, Yatao;Ma, Zhenguo;Dong, Wentao
    • Journal of Information Processing Systems
    • /
    • v.16 no.2
    • /
    • pp.508-521
    • /
    • 2020
  • We compared a novel encoding Lempel-Ziv complexity (ELZC) with three common complexity algorithms i.e., approximate entropy (ApEn), sample entropy (SampEn), and classic Lempel-Ziv complexity (CLZC) so as to determine a satisfied complexity and its corresponding quality indices for assessing quality of multi-lead electrocardiogram (ECG). First, we calculated the aforementioned algorithms on six artificial time series in order to compare their performance in terms of discerning randomness and the inherent irregularity within time series. Then, for analyzing sensitivity of the algorithms to content level of different noises within the ECG, we investigated their change trend in five artificial synthetic noisy ECGs containing different noises at several signal noise ratios. Finally, three quality indices based on the ELZC of the multi-lead ECG were proposed to assess the quality of 862 real 12-lead ECGs from the MIT databases. The results showed the ELZC could discern randomness and the inherent irregularity within six artificial time series, and also reflect content level of different noises within five artificial synthetic ECGs. The results indicated the AUCs of three quality indices of the ELZC had statistical significance (>0.500). The ELZC and its corresponding three indices were more suitable for multi-lead ECG quality assessment than the other three algorithms.

SMOOTHLY PERFECT 8-CONNECTED CONTOUR AND ITS CODING TECHNIQUE (평활한 완전 8방향 윤곽선과 이의 부호화 기법)

  • 조성호;김인철;이상욱
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1996.06a
    • /
    • pp.195-198
    • /
    • 1996
  • In this paper, we introduce the notion of the smoothly perfect 8-connected (SP8C) contour and propose a coding technique for the SP8C contours. Based on the contour simplification using the majority filter proposed by Gu[6], SP8C contours are extracted on the contour lattice from the segmented image. By noting that, unlike the perfect 8-connected contours, the SP8C contours are restricted to travel in only 3 different directions along the contours, we also propose two techniques for encoding the SP8C contours. The one is the modified version of the neighbouring direction segment coding by Kandeko[2], while the other is to employ the notion of the entropy coding. From the comparison in terms of the entropy, it is shown that the proposed SP8C contours require les bits in encoding the diagonal contours than the 4-connected contours employed by Gu. And computer simulations reveal that the contours can be efficiently encoded by the proposed technique.

  • PDF

Encoding of a run-length soruce using recursive indexing (줄길이 신호원의 순환지수 부호화)

  • 서재준;나상신
    • Journal of the Korean Institute of Telematics and Electronics A
    • /
    • v.33A no.7
    • /
    • pp.23-33
    • /
    • 1996
  • This paper deals with the design of a recursively-indexed binary code for facsimile soruces and its performance. Sources used here are run-lengths of white pixels form higher-resolution facsimile. The modified huffman code used for G.3 facsimile is chosen for the performance comparison. Experiments confirm the fact that recursive indexing preserves the entropy of a memoryless geometric source: the entropy of recursively-indexed physical surce iwth roughly geometric distributin remains within 2% of the empirical source entropy. The designed recursively-indexed binary codes consist of a code applied to text-type documents and to graphics - type documents is compared iwth that of the modified huffman code. Numerical resutls show that the modified huffman code performs well for text-type documents and not equally well for graphics-tyep documents. On the other hand, recursively-indexed binary codes have shown a better performance for graphics-type documents whose distribution are similar to a geometric distribution. Specifically, the code rates of recursively-indexed binary codes with 60 codewords are from 8% to 20% of the empirical source entropy smaller than that of th emodified huffman code with 91 codewords.

  • PDF

An efficient Pipelined Arithmetic Encoder for H.264/AVC (H.264/AVC를 위한 효율적인 Pipelined Arithmetic Encoder)

  • Yun, Jae-Bok;Park, Tae-Geun
    • Proceedings of the IEEK Conference
    • /
    • 2005.11a
    • /
    • pp.687-690
    • /
    • 2005
  • H.264/AVC에서 압축 효율을 향상시키기 위해 사용된 entropy coding중에 CABAC(Context-based Adaptive Binary Arithmetic Coding)은 하드웨어 복잡도가 높고 bit-serial 과정에서 data dependancy가 존재하기 때문에 빠른 연산이 어렵다. 본 논문에서는 adaptive arithmetic encoder와 정규화 과정을 효율적으로 구성하여 각 입력 심벌이 정규화 과정의 반복횟수에 관계없이 고정된 cycle에 encoding이 되도록 하였다. 제안한 구조는 pipeline으로 구성하기 용이하며, 이 경우 매 cycle에 한 입력 심벌의 encoding이 가능하다.

  • PDF

Design of the ICMEP Algorithm for the Highly Efficient Entropy Encoding (고효율 엔트로피 부호화를 위한 ICMEP 알고리즘 설계)

  • 이선근;임순자;김환용
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.41 no.4
    • /
    • pp.75-82
    • /
    • 2004
  • The channel transmission ratio is speeded up by the combination of the Huffman algorithm, the model scheme of the lossy transform having minimum average code lengths for the image information and good instantaneous decoding capability, with the Lempel-Ziv algorithm showing the fast processing performance during the compression process. In order to increase the processing speed during the compression process, ICMEP algorithm is proposed and the entropy encoder of HDTV is designed and inspected. The ICMEP entropy encoder have been designed by choosing the top-down method and consisted of the source codes and the test benches by the behavior expression with VHDL. As a simulation results, implemented ICMEP entropy encoder confirmed that whole system efficiency by memory saturation prevention and compressibility increase improves.

Multi-Symbol Binary Arithmetic Coding Algorithm for Improving Throughput in Hardware Implementation

  • Kim, Jin-Sung;Kim, Eung Sup;Lee, Kyujoong
    • Journal of Multimedia Information System
    • /
    • v.5 no.4
    • /
    • pp.273-276
    • /
    • 2018
  • In video compression standards, the entropy coding is essential to the high performance compression because redundancy of data symbols is removed. Binary arithmetic coding is one of high performance entropy coding methods. However, the dependency between consecutive binary symbols prevents improving the throughput. For the throughput enhancement, a new probability model is proposed for encoding multi-symbols at one time. In the proposed method, multi-symbol encoder is implemented with only adders and shifters, and the multiplication table for interval subdivision of binary arithmetic coding is removed. Compared to the compression ratio of CABAC of H.264/AVC, the performance degradation on average is only 1.4% which is negligible.

Anomaly-based Alzheimer's disease detection using entropy-based probability Positron Emission Tomography images

  • Husnu Baris Baydargil;Jangsik Park;Ibrahim Furkan Ince
    • ETRI Journal
    • /
    • v.46 no.3
    • /
    • pp.513-525
    • /
    • 2024
  • Deep neural networks trained on labeled medical data face major challenges owing to the economic costs of data acquisition through expensive medical imaging devices, expert labor for data annotation, and large datasets to achieve optimal model performance. The heterogeneity of diseases, such as Alzheimer's disease, further complicates deep learning because the test cases may substantially differ from the training data, possibly increasing the rate of false positives. We propose a reconstruction-based self-supervised anomaly detection model to overcome these challenges. It has a dual-subnetwork encoder that enhances feature encoding augmented by skip connections to the decoder for improving the gradient flow. The novel encoder captures local and global features to improve image reconstruction. In addition, we introduce an entropy-based image conversion method. Extensive evaluations show that the proposed model outperforms benchmark models in anomaly detection and classification using an encoder. The supervised and unsupervised models show improved performances when trained with data preprocessed using the proposed image conversion method.

Reduction of Air-pumping Noise based on a Genetic Algorithm (유전자 알고리즘을 이용한 타이어 공력소음의 저감)

  • Kim, Eui-Youl;Hwang, Sung-Wook;Kim, Byung-Hyun;Lee, Sang-Kwon
    • Transactions of the Korean Society for Noise and Vibration Engineering
    • /
    • v.22 no.1
    • /
    • pp.61-73
    • /
    • 2012
  • The paper presents the novel approach to solve some problems occurred in application of the genetic algorithm to the determination of the optimal tire pattern sequence in order to reduce the tire air-pumping noise which is generated by the repeated compression and expansion of the air cavity between tire pattern and road surface. The genetic algorithm has been used to find the optimal tire pattern sequence having a low level of tire air-pumping noise using the image based air-pumping model. In the genetic algorithm used in the previous researches, there are some problems in the encoding structure and the selection of objective function. The paper proposed single encoding element with five integers, divergent objective function based on evolutionary process and the optimal evolutionary rate based on Shannon entropy to solve the problems. The results of the proposed genetic algorithm with evolutionary process are compared with those of the randomized algorithm without evolutionary process on the two-dimensional normal distribution. It is confirmed that the genetic algorithm is more effective to reduce the peak value of the predicted tire air-pumping noise and the consistency and cohesion of the obtained simulation results are also improved in terms of probability.