• Title/Summary/Keyword: Compression processing time

Search Result 324, Processing Time 0.031 seconds

QuadTree-Based Lossless Image Compression and Encryption for Real-Time Processing (실시간 처리를 위한 쿼드트리 기반 무손실 영상압축 및 암호화)

  • Yoon, Jeong-Oh;Sung, Woo-Seok;Hwang, Chan-Sik
    • The KIPS Transactions:PartC
    • /
    • v.8C no.5
    • /
    • pp.525-534
    • /
    • 2001
  • Generally, compression and encryption procedures are performed independently in lossless image compression and encryption. When compression is followed by encryption, the compressed-stream should have the property of randomness because its entropy is decreased during the compression. However, when full data is compressed using image compression methods and then encrypted by encryption algorithms, real-time processing is unrealistic due to the time delay involved. In this paper, we propose to combine compression and encryption to reduce the overall processing time. It is method decomposing gray-scale image by means of quadtree compression algorithms and encrypting the structural part. Moreover, the lossless compression ratio can be increased using a transform that provides an decorrelated image and homogeneous region, and the encryption security can be improved using a reconstruction of the unencrypted quadtree data at each level. We confirmed the increased compression ratio, improved encryption security, and real-time processing by using computer simulations.

  • PDF

Study on the Similarity Functions for Image Compression (영상 압축을 위한 유사성 함수 연구)

  • Joo, Woo-Seok;Kang, Jong-Oh
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.8
    • /
    • pp.2133-2142
    • /
    • 1997
  • Compared with previous compression methods, fractal image compression drastically increases compression rate by using block-based encoding. Although decompression can be done in real time even with softwares, the most serious problem in utilizing the fractal method is the time required for the encoding. In this paper, we propose and verify i) an algorithm that reduces the encoding time by reducing the number of similarity searching on the basis of dimensional informations, and ii) an algorithm that enhances the quality of the restored image on the basis of brightness and contrast information. Finally, a method that enables fast compression with little quality degradation is proposed.

  • PDF

Intra-picture Block-matching Method for Codebook-based Texture Compression

  • Cui, Li;Jang, Euee S.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.10
    • /
    • pp.5063-5073
    • /
    • 2016
  • In this paper, an efficient texture compression method is proposed for fast rendering, which exploits the spatial correlation among blocks through intra-picture block matching. Texture mapping is widely used to enhance the visual quality of results in real-time rendering applications. For fast texture mapping, it is necessary to identify an effective trade-off between compression efficiency and computational complexity. The conventional compression methods utilized for image processing (e.g., JPEG) provide high compression efficiency while resulting in high complexity. Thus, low complexity methods, such as ETC1, are often used in real-time rendering applications. Although these methods can achieve low complexity, the compression efficiency is still lower than that of JPEG. To solve this problem, we propose a texture compression method by reducing the spatial redundancy between blocks in order to achieve the better compression performance than ETC1 while maintaining complexity that is lower than that of JPEG. Experimental results show that the proposed method achieves better compression efficiency than ETC1, and the decoding time is significantly reduced compared to JPEG while similar to ETC1.

Wavelet Compression Method with Minimum Delay for Mobile Tele-cardiology Applications (이동형 Tele-cardiology 시스템 적용을 위한 최저 지연을 가진 웨이브릿 압축 기법)

  • Kim Byoung-Soo;Yoo Sun-Kook;Lee Moon-Hyoung
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.53 no.11
    • /
    • pp.786-792
    • /
    • 2004
  • A wavelet based ECG data compression has become an attractive and efficient method in many mobile tele-cardiology applications. But large data size required for high compression performance leads a serious delay. In this paper, new wavelet compression method with minimum delay is proposed. It is based on deciding the type and compression ratio(CR) of block organically according to the standard deviation of input ECG data with minimum block size. Compression performances of the proposed algorithm for different MIT ECG Records were analyzed comparing other ECG compression algorithm. In addition to the processing delay measurement, compression efficiency and reconstruction sensitivity to error were also evaluated via random noise simulation models. The results show that the proposed algorithm has both lower PRD than other algorithm on same CR and minimum time in the data acquisition, processing and transmission.

Numerical Analysis of Residual Stresses and Birefringence in Injection/Compression Molded Center-gated Disks (II) - Effects of Processing Conditions - (사출/압축 성형 Center-gated 터스크에서의 잔류 응력과 복굴절의 수치 해석 (II) - 공정조건의 영향 -)

  • Lee, Young-Bok;Kwon, Tai-Hun;Yoon, Kyung-Hwan
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.26 no.11
    • /
    • pp.2355-2363
    • /
    • 2002
  • The accompanying paper, Part 1, has presented the physical modeling and basic numerical analysis results of both the flow-induced and thermally-induced residual stress and birefringence in injection molded center gated disks. The present paper, Part II, has attempted to investigate the effects of various processing conditions of injection/compression molding process on the residual stress and birefringence. The birefringence is significantly affected by injection melt temperature, packing pressure and packing time. Birefringence in the shell layer increases as melt temperature gets lower. The inner peak of birefringence increases with packing time and packing pressure. On the other hand, packing pressure, packing time and mold wall temperature affect the thermally-induced residual stress rather significantly in the shell layer, but insignificantly in the core region. Injection/compression molding has been found to reduce the birefringence in comparison with the conventional injection molding process. In particular, mold closing velocity and initial opening thickness in the compression stage of injection/compression molding process have significant effect on the flow-induced birefringence, but not on tile thermal residual stress and the thermally induced birefringence.

Fuzzy Logic-based Bit Compression Method for Distributed Face Recognition (분산 얼굴인식을 위한 퍼지로직 기반 비트 압축법)

  • Kim, Tae-Young;Noh, Chang-Hyeon;Lee, Jong-Sik
    • Journal of the Korea Society for Simulation
    • /
    • v.18 no.2
    • /
    • pp.9-17
    • /
    • 2009
  • A face database has contained a large amount of facial information data since face recognition was widely used. With the increase of facial information data, the face recognition based on distributed processing method has been noticed as a major topic. In existing studies, there were lack of discussion about the transferring method for large data. So, we proposed a fuzzy logic-based bit compression rate selection method for distributed face recognition. The proposed method selects an effective bit compression rate by fuzzy inference based on face recognition rate, processing time for recognition, and transferred bit length. And, we compared the facial recognition rate and the recognition time of the proposed method to those of facial information data with no compression and fixed bit compression rates. Experimental results demonstrates that the proposed method can reduce processing time for face recognition with a reasonable recognition rate.

Multi-Description Image Compression Coding Algorithm Based on Depth Learning

  • Yong Zhang;Guoteng Hui;Lei Zhang
    • Journal of Information Processing Systems
    • /
    • v.19 no.2
    • /
    • pp.232-239
    • /
    • 2023
  • Aiming at the poor compression quality of traditional image compression coding (ICC) algorithm, a multi-description ICC algorithm based on depth learning is put forward in this study. In this study, first an image compression algorithm was designed based on multi-description coding theory. Image compression samples were collected, and the measurement matrix was calculated. Then, it processed the multi-description ICC sample set by using the convolutional self-coding neural system in depth learning. Compressing the wavelet coefficients after coding and synthesizing the multi-description image band sparse matrix obtained the multi-description ICC sequence. Averaging the multi-description image coding data in accordance with the effective single point's position could finally realize the compression coding of multi-description images. According to experimental results, the designed algorithm consumes less time for image compression, and exhibits better image compression quality and better image reconstruction effect.

CNN based Image Restoration Method for the Reduction of Compression Artifacts (압축 왜곡 감소를 위한 CNN 기반 이미지 화질개선 알고리즘)

  • Lee, Yooho;Jun, Dongsan
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.5
    • /
    • pp.676-684
    • /
    • 2022
  • As realistic media are widespread in various image processing areas, image or video compression is one of the key technologies to enable real-time applications with limited network bandwidth. Generally, image or video compression cause the unnecessary compression artifacts, such as blocking artifacts and ringing effects. In this study, we propose a Deep Residual Channel-attention Network, so called DRCAN, which consists of an input layer, a feature extractor and an output layer. Experimental results showed that the proposed DRCAN can reduced the total memory size and the inference time by as low as 47% and 59%, respectively. In addition, DRCAN can achieve a better peak signal-to-noise ratio and structural similarity index measure for compressed images compared to the previous methods.

An Efficient Bit-Level Lossless Grayscale Image Compression Based on Adaptive Source Mapping

  • Al-Dmour, Ayman;Abuhelaleh, Mohammed;Musa, Ahmed;Al-Shalabi, Hasan
    • Journal of Information Processing Systems
    • /
    • v.12 no.2
    • /
    • pp.322-331
    • /
    • 2016
  • Image compression is an essential technique for saving time and storage space for the gigantic amount of data generated by images. This paper introduces an adaptive source-mapping scheme that greatly improves bit-level lossless grayscale image compression. In the proposed mapping scheme, the frequency of occurrence of each symbol in the original image is computed. According to their corresponding frequencies, these symbols are sorted in descending order. Based on this order, each symbol is replaced by an 8-bit weighted fixed-length code. This replacement will generate an equivalent binary source with an increased length of successive identical symbols (0s or 1s). Different experiments using Lempel-Ziv lossless image compression algorithms have been conducted on the generated binary source. Results show that the newly proposed mapping scheme achieves some dramatic improvements in regards to compression ratios.

Hardware Architecture and its Design of Real-Time Video Compression Processor for Motion JPEG2000 (Motion JPEG2000을 위한 실시간 비디오 압축 프로세서의 하드웨어 구조 및 설계)

  • 서영호;김동욱
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.53 no.1
    • /
    • pp.1-9
    • /
    • 2004
  • In this paper, we proposed a hardware(H/W) structure which can compress and recontruct the input image in real time operation and implemented it into a FPGA platform using VHDL(VHSIC Hardware Description Language). All the image processing element to process both compression and reconstruction in a FPGA were considered each of them was mapped into a H/W with the efficient structure for FPGA. We used the DWT(discrete wavelet transform) which transforms the data from spatial domain to the frequency domain, because use considered the motion JPEG2000 as the application. The implemented H/W is separated to both the data path part and the control part. The data path part consisted of the image processing blocks and the data processing blocks. The image processing blocks consisted of the DWT Kernel for the filtering by DWT, Quantizer/Huffman Encoder, Inverse Adder/Buffer for adding the low frequency coefficient to the high frequency one in the inverse DWT operation, and Huffman Decoder. Also there existed the interface blocks for communicating with the external application environments and the timing blocks for buffering between the internal blocks. The global operations of the designed H/W are the image compression and the reconstruction, and it is operated by the unit or a field synchronized with the A/D converter. The implemented H/W used the 54%(12943) LAB(Logic Array Block) and 9%(28352) ESB(Embedded System Block) in the APEX20KC EP20K600CB652-7 FPGA chip of ALTERA, and stably operated in the 70MHz clock frequency. So we verified the real time operation. that is. processing 60 fields/sec(30 frames/sec).