• Title/Summary/Keyword: Video Compression

Search Result 775, Processing Time 0.022 seconds

Basis Function Truncation Effect of the Gabor Cosine and Sine Transform (Gabor 코사인과 사인 변환의 기저함수 절단 효과)

  • Lee, Juck-Sik
    • The KIPS Transactions:PartB
    • /
    • v.11B no.3
    • /
    • pp.303-308
    • /
    • 2004
  • The Gabor cosine and sine transform can be applied to image and video compression algorithm by representing image frequency components locally The computational complexity of forward and inverse matrix transforms used in the compression and decompression requires O($N^3$)operations. In this paper, the length of basis functions is truncated to produce a sparse basis matrix, and the computational burden of transforms reduces to deal with image compression and reconstruction in a real-time processing. As the length of basis functions is decreased, the truncation effects to the energy of basis functions are examined and the change in various Qualify measures is evaluated. Experiment results show that 11 times fewer multiplication/addition operations are achieved with less than 1% performance change.

BTC-based Image Compression using Pattern (패턴을 이용한 블록 절단 부호화 기반의 영상 압축)

  • Kim, Cheonshik;Oh, Jae-Whan
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.15 no.3
    • /
    • pp.77-83
    • /
    • 2015
  • Block Truncation Coding, or BTC, is a type of lossy image compression technique for grayscale images. It divides the original images into blocks and then reduces the number of grey levels in each block to compute the mean and standard deviation. BTC has also been adapted to video compression. Another variation of BTC is Absolute Moment Block Truncation Coding. AMBTC is computationally simpler than BTC. In this paper, we proposed new image compression method based on BTC, which is applied patterns to improve compression rate and image quality. This method make two codebooks to extract 36 and 64 patterns from the highest frequency patterns in BTC. When you are compressing an image, you compare many block patterns to that of codebook and use to compress indexes of identical patterns. We experiment our proposed scheme with 36 patterns and the experimental results showed the compression rate of 1.37 bpp. In this paper, our proposed scheme showed higher compression rate rather than that of BTC. In experiment, we used standard images for the performance evaluation.

Pyramid Feature Compression with Inter-Level Feature Restoration-Prediction Network (계층 간 특징 복원-예측 네트워크를 통한 피라미드 특징 압축)

  • Kim, Minsub;Sim, Donggyu
    • Journal of Broadcast Engineering
    • /
    • v.27 no.3
    • /
    • pp.283-294
    • /
    • 2022
  • The feature map used in the network for deep learning generally has larger data than the image and a higher compression rate than the image compression rate is required to transmit the feature map. This paper proposes a method for transmitting a pyramid feature map with high compression rate, which is used in a network with an FPN structure that has robustness to object size in deep learning-based image processing. In order to efficiently compress the pyramid feature map, this paper proposes a structure that predicts a pyramid feature map of a level that is not transmitted with pyramid feature map of some levels that transmitted through the proposed prediction network to efficiently compress the pyramid feature map and restores compression damage through the proposed reconstruction network. Suggested mAP, the performance of object detection for the COCO data set 2017 Train images of the proposed method, showed a performance improvement of 31.25% in BD-rate compared to the result of compressing the feature map through VTM12.0 in the rate-precision graph, and compared to the method of performing compression through PCA and DeepCABAC, the BD-rate improved by 57.79%.

3-D Lossy Volumetric Medical Image Compression with Overlapping method and SPIHT Algorithm and Lifting Steps (Overlapping method와 SPIHT Algorithm과 Lifting Steps을 이용한 3차원 손실 의료 영상 압축 방법)

  • 김영섭
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.4 no.3
    • /
    • pp.263-269
    • /
    • 2003
  • This paper focuses on lossy medical image compression methods for medical images that operate on three-dimensional(3D) irreversible integer wavelet transform. We offer an application of the Set Partitioning in Hierarchical Trees(SPIHT) algorithm〔l-3〕to medical images, using a 3-D wavelet decomposition and a 3-D spatial dependence tree. The wavelet decomposition is accomplished with integer wavelet filters implemented with the lifting method, where careful scaling and truncations keep the integer precision small and the transform unitary. As the compression rate increases, the boundaries between adjacent coding units become increasingly visible. Unlike video, the volume image is examined under static condition, and must not exhibit such boundary artifacts. In order to eliminate them, we utilize overlapping at axial boundaries between adjacent coding units. We have tested our encoder on medical images using different integer filters. Results show that our algorithm with certain filters performs as well. The improvement is visibly manifested as fewer ringing artifacts and noticeably better reconstruction of low contrast.

  • PDF

APBT-JPEG Image Coding Based on GPU

  • Wang, Chengyou;Shan, Rongyang;Zhou, Xiao
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.4
    • /
    • pp.1457-1470
    • /
    • 2015
  • In wireless multimedia sensor networks (WMSN), the latency of transmission is an increasingly problem. With the improvement of resolution, the time cost in image and video compression is more and more, which seriously affects the real-time of WMSN. In JPEG system, the core of the system is DCT, but DCT-JPEG is not the best choice. Block-based DCT transform coding has serious blocking artifacts when the image is highly compressed at low bit rates. APBT is used in this paper to solve that problem, but APBT does not have a fast algorithm. In this paper, we analyze the structure in JPEG and propose a parallel framework to speed up the algorithm of JPEG on GPU. And we use all phase biorthogonal transform (APBT) to replace the discrete cosine transform (DCT) for the better performance of reconstructed image. Therefore, parallel APBT-JPEG is proposed to solve the real-time of WMSN and the blocking artifacts in DCT-JPEG in this paper. We use the CUDA toolkit based on GPU which is released by NVIDIA to design the parallel algorithm of APBT-JPEG. Experimental results show that the maximum speedup ratio of parallel algorithm of APBT-JPEG can reach more than 100 times with a very low version GPU, compared with conventional serial APBT-JPEG. And the reconstructed image using the proposed algorithm has better performance than the DCT-JPEG in terms of objective quality and subjective effect. The proposed parallel algorithm based on GPU of APBT also can be used in image compression, video compression, the edge detection and some other fields of image processing.

A Fast Search Algorithm for Sub-Pixel Motion Estimation (부화소 움직임 추정을 위한 고속 탐색 기법)

  • Park, Dong-Kyun;Jo, Seong-Hyeon;Cho, Hyo-Moon;Lee, Jong-Hwa
    • Proceedings of the KIEE Conference
    • /
    • 2007.04a
    • /
    • pp.26-28
    • /
    • 2007
  • The motion estimation is the most important technique in the image compression of the video standards. In the case of next generation standards in the video codec as H.264, a high compression-efficiency can be also obtained by using a motion compensation. To obtain the accurate motion search, a motion estimation should be achieved up to 1/2 pixel and 1/4 pixel uiuts. To do this, the computational complexity is increased although the image compression rate is increased. Therefore, in this paper, we propose the advanced sub-pixel block matching algorithm to reduce the computational complexity by using a statistical characteristics of SAD(Sum of Absolute Difference). Generally, the probability of the minimum SAD values is high when searching point is in the distance 1 from the reference point. Thus, we reduced the searching area and then we can overcome the computational complexity problem. The main concept of proposed algorithm, which based on TSS(Three Step Search) method, first we find three minimum SAD points which is in integer distance unit, and then, in second step, the optimal point is in 1/2 pixel unit either between the most minimum SAD value point and the second minimum SAD point or between the most minimum SAD value point and the third minimum SAD point In third step, after finding the smallest SAD value between two SAD values on 1/2 pixel unit, the final optimized point is between the most minimum SAD value and the result value of the third step, in 1/2 pixel unit i.e., 1/4 pixel unit in totally. The conventional TSS method needs an eight.. search points in the sub-pixel steps in 1/2 pixel unit and also an eight search points in 1/4 pixel, to detect the optimal point. However, in proposed algorithm, only total five search points are needed. In the result. 23 % improvement of processing speed is obtained.

  • PDF

VLSI architecture design of CAVLC entropy encoder/decoder for H.264/AVC (H.264/AVC를 위한 CAVLC 엔트로피 부/복호화기의 VLSI 설계)

  • Lee Dae-joon;Jeong Yong-jin
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.30 no.5C
    • /
    • pp.371-381
    • /
    • 2005
  • In this paper, we propose an advanced hardware architecture for the CAVLC entropy encoder/decoder engine for real time video compression. The CAVLC (Context-based Adaptive Variable Length Coding) is a lossless compression method in H.264/AVC and it has high compression efficiency but has computational complexity. The reference memory size is optimized using partitioned storing method and memory reuse method which are based on partiality of memory referencing. We choose the hardware architecture which has the most suitable one in several encoder/decoder architectures for the mobile devices and improve its performance using parallel processing. The proposed architecture has been verified by ARM-interfaced emulation board using Altera Excalibur and also synthesized on Samsung 0.18 um CMOS technology. The synthesis result shows that the encoder can process about 300 CIF frames/s at 150MHz and the decoder can process about 250 CIF frames/s at 140Mhz. The hardware architectures are being used as core modules when implementing a complete H.264/AVC video encoder/decoder chip for real-time multimedia application.

Analysis of Intra Prediction for Digital Watermarking based on HEVC (HEVC기반의 디지털 워터마킹을 위한 인트라 예측의 분석)

  • Seo, Young-Ho;Kim, Bora;Kim, Dong-Wook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.19 no.5
    • /
    • pp.1189-1198
    • /
    • 2015
  • Recently, with rapid development of digital broadcasting technology, high-definition video service increased interest and demand. supplied mobile and image device support that improve 4~16 time existing Full HD. Such as high-definition contents supply, proposed compression for high-efficiency video codec (HEVC). Therefore, watermarking technology is necessary applying HEVC for protecting ownership and intellectual property. In this paper, analysis of prediction mode in intra frame and study feasibility of watermarking in re-encoding based HEVC. Proposed detect un-changed blocks in intra frame, using the result of analysis prediction mode.

A Multi-Channel Trick Mode Play Algorithm and Hardware Implementation of H.264/AVC for Surveillance Applications (H.264/AVC 감시 어플리케이션용 멀티 채널 트릭 모드 재생 알고리즘 및 하드웨어 구현)

  • Jo, Hyeonsu;Hong, Youpyo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.41 no.12
    • /
    • pp.1834-1843
    • /
    • 2016
  • DVRs are the most common recording and displaying devices used for surveillance. Video compression plays a key role in DVRs for saving storage; the video compression standard, H.264/AVC, has recently become the dominant choice for DVRs. DVRs require various display modes, such as fast-forward, backward play, and pause; these are called trick modes. The implementation of precise trick mode play requires a very high decoding capability or a very intelligent scheme in order to handle the high computation complexity. The complexity is increased in many surveillance applications where more than one camera is used to monitor multiple spots or to monitor the same area using various angles. An implementation of a trick mode play and a frame buffer management scheme for the hardware-based H.264/AVC codec for multi-channel is presented in this paper. The experimental results show that exact trick mode play is possible using a standard H.264/AVC video codec with keyframe encoding feature at the expense of bitstream size increase.

MPEG-DASH based 3D Point Cloud Content Configuration Method (MPEG-DASH 기반 3차원 포인트 클라우드 콘텐츠 구성 방안)

  • Kim, Doohwan;Im, Jiheon;Kim, Kyuheon
    • Journal of Broadcast Engineering
    • /
    • v.24 no.4
    • /
    • pp.660-669
    • /
    • 2019
  • Recently, with the development of three-dimensional scanning devices and multi-dimensional array cameras, research is continuously conducted on techniques for handling three-dimensional data in application fields such as AR (Augmented Reality) / VR (Virtual Reality) and autonomous traveling. In particular, in the AR / VR field, content that expresses 3D video as point data has appeared, but this requires a larger amount of data than conventional 2D images. Therefore, in order to serve 3D point cloud content to users, various technological developments such as highly efficient encoding / decoding and storage, transfer, etc. are required. In this paper, V-PCC bit stream created using V-PCC encoder proposed in MPEG-I (MPEG-Immersive) V-PCC (Video based Point Cloud Compression) group, It is defined by the MPEG-DASH (Dynamic Adaptive Streaming over HTTP) standard, and provides to be composed of segments. Also, in order to provide the user with the information of the 3D coordinate system, the depth information parameter of the signaling message is additionally defined. Then, we design a verification platform to verify the technology proposed in this paper, and confirm it in terms of the algorithm of the proposed technology.