• Title/Summary/Keyword: Terms Image compression

Search Result 87, Processing Time 0.025 seconds

An Error-Resilient Image Compression Base on the Zerotree Wavelet Algorithm (오류에 강인한 제로트리 웨이블릿 영상 압축)

  • 장우영;송환종;손광훈
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.25 no.7A
    • /
    • pp.1028-1036
    • /
    • 2000
  • In this paper, an error-resilient image compression technique using wavelet transform is proposed. The zerotree technique that uses properties of statistics, energy and directions of wavelet coefficients in the space-frequency domain shows effective compression results. Since it is highly sensitive to the propagation of channel errors, evena single bit error degrades the whole image quality severely. In the proposed algorithm, the image is encoded by the SPIHT(Set Partitioning in Hierarchical Trees) algorithm using the zerotree coding technique. Encoded bitstreams are partitioned into some blocks using the subband correlations and then fixed-length blocks are made by using the effective bit reorganization algorithm. finally, an effective bit allocation technique is used to limit error propagation in each block. Therefore, in low BER the proposed algorithm shows similar compression performance to the zerotree compression technique and in high BER it shows better performance in terms of PSNR than the conventional methods.

  • PDF

Digital Watermarking of JPEG Image Based on Human Visual System (인간 시각 시스템에 기반 한 JPEG 영상의 디지털 워터마킹)

  • Bae, Sung-Ho
    • The KIPS Transactions:PartB
    • /
    • v.11B no.2
    • /
    • pp.125-132
    • /
    • 2004
  • In this paper, a watermark inserting method according to the sensitivity of human visual system and minimizing distortion of original DCT coefficients in DCT transform domain is proposed. The proposed method inserts a more robust watermark in the insensitive block of human vision by reordering the blocks according to the human visual system which is appropriate to the JPEG image compression. It also enhances the invisibility and robustness in high compression rate in terms of the watermark inserting method within the block which minimizes distortions of original DCT coefficients. The computer simulation results show that the proposed method maintains high image quality and good robustness in high compression rate compared with conventional watermarking method.

APBT-JPEG Image Coding Based on GPU

  • Wang, Chengyou;Shan, Rongyang;Zhou, Xiao
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.4
    • /
    • pp.1457-1470
    • /
    • 2015
  • In wireless multimedia sensor networks (WMSN), the latency of transmission is an increasingly problem. With the improvement of resolution, the time cost in image and video compression is more and more, which seriously affects the real-time of WMSN. In JPEG system, the core of the system is DCT, but DCT-JPEG is not the best choice. Block-based DCT transform coding has serious blocking artifacts when the image is highly compressed at low bit rates. APBT is used in this paper to solve that problem, but APBT does not have a fast algorithm. In this paper, we analyze the structure in JPEG and propose a parallel framework to speed up the algorithm of JPEG on GPU. And we use all phase biorthogonal transform (APBT) to replace the discrete cosine transform (DCT) for the better performance of reconstructed image. Therefore, parallel APBT-JPEG is proposed to solve the real-time of WMSN and the blocking artifacts in DCT-JPEG in this paper. We use the CUDA toolkit based on GPU which is released by NVIDIA to design the parallel algorithm of APBT-JPEG. Experimental results show that the maximum speedup ratio of parallel algorithm of APBT-JPEG can reach more than 100 times with a very low version GPU, compared with conventional serial APBT-JPEG. And the reconstructed image using the proposed algorithm has better performance than the DCT-JPEG in terms of objective quality and subjective effect. The proposed parallel algorithm based on GPU of APBT also can be used in image compression, video compression, the edge detection and some other fields of image processing.

All Phase Discrete Sine Biorthogonal Transform and Its Application in JPEG-like Image Coding Using GPU

  • Shan, Rongyang;Zhou, Xiao;Wang, Chengyou;Jiang, Baochen
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.9
    • /
    • pp.4467-4486
    • /
    • 2016
  • Discrete cosine transform (DCT) based JPEG standard significantly improves the coding efficiency of image compression, but it is unacceptable event in serious blocking artifacts at low bit rate and low efficiency of high-definition image. In the light of all phase digital filtering theory, this paper proposes a novel transform based on discrete sine transform (DST), which is called all phase discrete sine biorthogonal transform (APDSBT). Applying APDSBT to JPEG scheme, the blocking artifacts are reduced significantly. The reconstructed image of APDSBT-JPEG is better than that of DCT-JPEG in terms of objective quality and subjective effect. For improving the efficiency of JPEG coding, the structure of JPEG is analyzed. We analyze key factors in design and evaluation of JPEG compression on the massive parallel graphics processing units (GPUs) using the compute unified device architecture (CUDA) programming model. Experimental results show that the maximum speedup ratio of parallel algorithm of APDSBT-JPEG can reach more than 100 times with a very low version GPU. Some new parallel strategies are illustrated in this paper for improving the performance of parallel algorithm. With the optimal strategy, the efficiency can be improved over 10%.

저전송률 영상압축에 있어서의 후처리 기법

  • 이주흥;정제창;최병욱
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1996.06a
    • /
    • pp.233-236
    • /
    • 1996
  • A new method of blocking effects reduction is proposed in this paper for use in low bitrate image coding. We use 28 DCT kernel functions of which boundary values are linearly independent, and Gram-Schmidt process is applied to the boundary values in order to obtain 28 boundary-orthonormal basis images. Then we use these basis images to obtain the correction terms for blocking artifacts reduction. A threshold of block discontinuity is introduced for improvement of visual quality by reducing image blurring. We also investigate the number of basis images needed for efficient blocking artifacts reduction when the compression ratio changes.

  • PDF

Image Scrambling for One-Chip JPEG Applications (One-Chip JPEG 적용을 위한 영상 스크램블링)

  • 권정익;원치선;김재공
    • Proceedings of the Korea Institutes of Information Security and Cryptology Conference
    • /
    • 1994.11a
    • /
    • pp.193-202
    • /
    • 1994
  • In this paper, we investigate possible scrambling methods for the JPEG(Joint Photographic Export Group) still image compression standard. In particular, we compare the conventional line rotation and line permutation methods to the DCT block scrambling in terms of the number of bits to be increased and the easiness of buffer control. Computer simulation results show that the DCT block scrambling method is suitable for both data security and buffer control in one-chip JPEG applications.

  • PDF

A Buffer-constrained Adaptive Quantization Algorithm for Image Compression (버퍼제약에 의한 영상압축 적응양자화 알고리듬)

  • 박대철;정두영
    • Journal of Korea Multimedia Society
    • /
    • v.5 no.3
    • /
    • pp.249-254
    • /
    • 2002
  • We consider a buffer-constrained adaptive quantization algorithm for image compression. Buffer control algorithm was considered with source coding scheme by some researchers and recently a formal description of the algorithm in terms of rate-distortion has been developed. We propose a buffer control algorithm that incorporates the buffer occupancy into the Lagrange multiplier form in a rate-distortion cost measure. Although the proposed algorithm provides the suboptimal performance as opposed to the optimal Vieterbi algorithm, it can be implemented with very low computaional complexity. In addition stability of this buffer control algorithm has been mentioned briefly using Liapnov stability theory.

  • PDF

IMAGE QUALITY OPTIMIZATION BASED ON WAVELET FILTER DESIGN AND WAVELET DECOMPOSITION IN JPEG2000

  • Quan, Do;Ho, Yo-Sung
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2009.01a
    • /
    • pp.7-12
    • /
    • 2009
  • In JPEG2000, the Cohen-Daubechies-Feauveau (CDF) 9/7-tap wavelet filter adopted in lossy compression is implemented by the lifting scheme or by the convolution scheme while the LeGall 5/3-tap wavelet filter adopted in lossless compression is implemented just by the lifting scheme. However, these filters are not optimal in terms of Peak Signal-to-Noise Ratio (PSNR) values, and irrational coefficients of wavelet filters are complicated. In this paper, we proposed a method to optimize image quality based on wavelet filter design and on wavelet decomposition. First, we propose a design of wavelet filters by selecting the most appropriate rational coefficients of wavelet filters. These filters are shown to have better performance than previous wavelet ones. Then, we choose the most appropriate wavelet decomposition to get the optimal PSNR values of images.

  • PDF

Multidimensional uniform cubic lattice vector quantization for wavelet transform coding (웨이브렛변환 영상 부호화를 위한 다차원 큐빅 격자 구조 벡터 양자화)

  • 황재식;이용진;박현욱
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.22 no.7
    • /
    • pp.1515-1522
    • /
    • 1997
  • Several image coding algorithms have been developed for the telecommunication and multimedia systems with high image quality and high compression ratio. In order to achieve low entropy and distortion, the system should pay great cost of computation time and memory. In this paper, the uniform cubic lattice is chosen for Lattice Vector Quantization (LVQ) because of its generic simplicity. As a transform coding, the Discrete Wavelet Transform (DWT) is applied to the images because of its multiresolution property. The proposed algorithm is basically composed of the biorthogonal DWT and the uniform cubic LVQ. The multiresolution property of the DWT is actively used to optimize the entropy and the distortion on the basis of the distortion-rate function. The vector codebooks are also designed to be optimal at each subimage which is analyzed by the biorthogonal DWT. For compression efficiency, the vector codebook has different dimension depending on the variance of subimage. The simulation results show that the performance of the proposed coding mdthod is superior to the others in terms of the computation complexity and the PSNR in the range of entropy below 0.25 bpp.

  • PDF

Very deep super-resolution for efficient cone-beam computed tomographic image restoration

  • Hwang, Jae Joon;Jung, Yun-Hoa;Cho, Bong-Hae;Heo, Min-Suk
    • Imaging Science in Dentistry
    • /
    • v.50 no.4
    • /
    • pp.331-337
    • /
    • 2020
  • Purpose: As cone-beam computed tomography (CBCT) has become the most widely used 3-dimensional (3D) imaging modality in the dental field, storage space and costs for large-capacity data have become an important issue. Therefore, if 3D data can be stored at a clinically acceptable compression rate, the burden in terms of storage space and cost can be reduced and data can be managed more efficiently. In this study, a deep learning network for super-resolution was tested to restore compressed virtual CBCT images. Materials and Methods: Virtual CBCT image data were created with a publicly available online dataset (CQ500) of multidetector computed tomography images using CBCT reconstruction software (TIGRE). A very deep super-resolution (VDSR) network was trained to restore high-resolution virtual CBCT images from the low-resolution virtual CBCT images. Results: The images reconstructed by VDSR showed better image quality than bicubic interpolation in restored images at various scale ratios. The highest scale ratio with clinically acceptable reconstruction accuracy using VDSR was 2.1. Conclusion: VDSR showed promising restoration accuracy in this study. In the future, it will be necessary to experiment with new deep learning algorithms and large-scale data for clinical application of this technology.