• Title/Summary/Keyword: Motion transform

Search Result 461, Processing Time 0.023 seconds

A Moving Picture Coding Method Based on Region Segmentation Using Genetic Algorithm (유전적 알고리즘을 이용한 동화상의 영역분할 부호화 방법)

  • Jung, Nam-Chae
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.10 no.1
    • /
    • pp.32-39
    • /
    • 2009
  • In this paper, the method of region segmentation using genetic algorithm is proposed for an improvement of efficiency in moving picture coding. A genetic algorithm is the method that searches a large probing space using only a function value for a optimal combination consecutively. By progressing both motion presumption and region segmentation at once, we can assign the motion vector in a image to a small block or a pixel respectively, and transform the capacity of coding and a signal to noise rate into a problem of optimization. That is to say, there is close correlation between region segmentation and motion presumption in motion-compensated prediction coding. This is to optimize the capacity of coding and a S/N ratio. This is to arrange the motion vector in each block of picture according to the state of optimization. Therefore, we examined both the data type of genetic algorithm and the method of data processing to obtain the results of optimal region segmentation in this paper. And we confirmed the validity of a proposed method using the test pictures by means of computer simulation.

  • PDF

A full-Hardwired Low-Power MPEG4@SP Video Encoder for Mobile Applications (모바일 향 저전력 동영상 압축을 위한 고집적 MPEG4@SP 동영상 압축기)

  • Shin, Sun Young;Park, Hyun Sang
    • Journal of Broadcast Engineering
    • /
    • v.10 no.3
    • /
    • pp.392-400
    • /
    • 2005
  • Highly integrated MPEG-4@SP video compression engine, VideoCore, is proposed for mobile application. The primary components of video compression require the high memory bandwidth since they access the external memory frequently. They include motion estimation, motion compensation, quantization, discrete cosine transform, variable length coding, and so on. The motion estimation processor adopted in VideoCore utilizes the small-size local memories such that the video compression system accesses external memory as less frequently as possible. The entire video compression system is divided into two distinct sub-systems: the integer-unit motion estimation part and the others, and both operate concurrently in a pipelined architecture. Thus the VideoCore enables the real-time high-quality video compression with a relatively low operation frequency.

Embedded Compression Codec Algorithm for Motion Compensated Wavelet Video Coding System (움직임 보상된 웨이블릿 기반의 비디오 코딩 시스템에 적용 가능한 임베디드 압축 코덱 알고리즘)

  • Kim, Song-Ju
    • The Journal of the Korea Contents Association
    • /
    • v.12 no.3
    • /
    • pp.77-83
    • /
    • 2012
  • In this paper, a low-complexity embedded compression (EC) Codec algorithm for the wavelet video coder is applied to reduce excessive external memory requirements. The EC algorithm is used to achieve a fixed compression ratio of 50 % under the near-lossless-compression constraint. The EC technique can reduce the 50 % memory requirement for intermediate low-frequency coefficients during multiple discrete wavelet transform stages compared with direct implementation of the wavelet video encoder of this paper. Furthermore, the EC scheme based on a forward adaptive quantization and fixed length coding can save bandwidth and size of buffer between DWT and SPIHT to 50 %. Simulation results show that our EC algorithm present only PSNR degradation of 0.179 and 0.162 dB in average when the target bit-rate of the video coder are 1 and 0.5 bpp, respectively.

Analysis on the special quantitative variation of dot model by the position transform

  • Kim, Jeong-lae;Kim, Kyung-seop
    • International Journal of Advanced Culture Technology
    • /
    • v.5 no.3
    • /
    • pp.67-72
    • /
    • 2017
  • Transform variation technique is constituted the vibration status of the flash-gap recognition level (FGRL) on the distribution recognition function. The recognition level condition by the distribution recognition function system is associated with the scattering vibration system. As to search a position of the dot model, we are consisted of the distribution value with character point by the output signal. The concept of recognition level is composed the reference of flash-gap level for variation signal by the distribution vibration function. For displaying a variation of the FGRL of the maximum-average in terms of the vibration function, and distribution position vibration that was the a distribution value of the far variation of the $Dis-rf-FA-{\alpha}_{MAX-AVG}$ with $5.74{\pm}1.12$ units, that was the a distribution value of the convenient variation of the $Dis-rf-CO-{\alpha}_{MAX-AVG}$ with $1.64{\pm}0.16$ units, that was the a distribution value of the flank variation of the $Dis-rf-FL-{\alpha}_{MAX-AVG}$ with $0.74{\pm}0.24$ units, that was the a distribution value of the vicinage variation of the $Dis-rf-VI-{\alpha}_{MAX-AVG}$ with $0.12{\pm}0.01$ units. The scattering vibration will be to evaluate at the ability of the vibration function with character point by the distribution recognition level on the FGRL that is showed the flash-gap function by the recognition level system. Scattering recognition system will be possible to control of a function by the special signal and to use a distribution data of scattering vibration level.

Design of Architecture of Programmable Stack-based Video Processor with VHDL (VHDL을 이용한 프로그램 가능한 스택 기반 영상 프로세서 구조 설계)

  • 박주현;김영민
    • Journal of the Korean Institute of Telematics and Electronics C
    • /
    • v.36C no.4
    • /
    • pp.31-43
    • /
    • 1999
  • The main goal of this paper is to design a high performance SVP(Stack based Video Processor) for network applications. The SVP is a comprehensive scheme; 'better' in the sense that it is an optimal selection of previously proposed enhancements of a stack machine and a video processor. This can process effectively object-based video data using a S-RISC(Stack-based Reduced Instruction Set Computer) with a semi -general-purpose architecture having a stack buffer for OOP(Object-Oriented Programming) with many small procedures at running programs. And it includes a vector processor that can improve the MPEG coding speed. The vector processor in the SVP can execute advanced mode motion compensation, motion prediction by half pixel and SA-DCT(Shape Adaptive-Discrete Cosine Transform) of MPEG-4. Absolutors and halfers in the vector processor make this architecture extensive to a encoder. We also designed a VLSI stack-oriented video processor using the proposed architecture of stack-oriented video decoding. It was designed with O.5$\mu\textrm{m}$ 3LM standard-cell technology, and has 110K logic gates and 12 Kbits SRAM internal buffer. The operating frequency is 50MHz. This executes algorithms of video decoding for QCIF 15fps(frame per second), maximum rate of VLBV(Very Low Bitrate Video) in MPEG-4.

  • PDF

Hybrid Super-Resolution Algorithm Robust to Cut-Change (컷 전환에 적응적인 혼합형 초고해상도 기법)

  • Kwon, Soon-Chan;Lim, Jong-Myeong;Yoo, Jisang
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.17 no.7
    • /
    • pp.1672-1686
    • /
    • 2013
  • In this paper, we propose a hybrid super-resolution algorithm robust to cut-change. Existing single-frame based super-resolution algorithms are usually fast, but quantity of information for interpolation is limited. Although the existing multi-frame based super-resolution algorithms generally robust to this problem, the performance of algorithm strongly depends on motions of input video. Furthemore at boundary of cut, applying of the algorithm is limited. In the proposed method, we detect a define boundary of cut using cut-detection algorithm. Then we adaptively apply a single-frame based super-resolution method to detected cut. Additionally, we propose algorithms of normalizing motion vector and analyzing pattern of edge to solve various problems of existing super-resolution algorithms. The experimental results show that the proposed algorithm has better performance than other conventional interpolation methods.

Super-Resolution Algorithm by Motion Estimation with Sub-Pixel Accuracy using 6-Tap FIR Filter (6-Tap FIR 필터를 이용한 부화소 단위 움직임 추정을 통한 초해상도 기법)

  • Kwon, Soon-Chan;Yoo, Ji-Sang
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37 no.6A
    • /
    • pp.464-472
    • /
    • 2012
  • In this paper, we propose a new super-resolution algorithm that uses successive frames by applying the block matching motion estimation algorithm. Usually, single frame super-resolution algorithms are based on probability or discrete wavelet transform (DWT) approach to extract high-frequency components of the input image, but only limited information is available for these algorithms. To solve this problem, various multiple-frame based super-resolution algorithms are proposed. The accuracy of registration between frames is a very important factor for the good performance of an algorithm. We therefore propose an algorithm using 6-Tap FIR filter to increase the accuracy of the image registration with sub-pixel unit. Proposed algorithm shows better performance than other conventional interpolation based algorithms such as nearest neighborhood, bi-linear and bi-cubic methods and results in about the same image quality as DWT based super-resolution algorithm.

Fire-Flame Detection Using Fuzzy Logic (퍼지 로직을 이용한 화재 불꽃 감지)

  • Hwang, Hyun-Jae;Ko, Byoung-Chul
    • The KIPS Transactions:PartB
    • /
    • v.16B no.6
    • /
    • pp.463-470
    • /
    • 2009
  • In this paper, we propose the advanced fire-flame detection algorithm using camera image for better performance than previous sensors-based systems which is limited on small area. Also, previous works using camera image were depend on a lot of heuristic thresholds or required an additional computation time. To solve these problems, we use statistical values and divide image into blocks to reduce the processing time. First, from the captured image, candidate flame regions are detected by a background model and fire colored models of the fire-flame. After the probability models are formed using the change of luminance, wavelet transform and the change of motion on time axis, they are used for membership function of fuzzy logic. Finally, the result function is made by the defuzzification, and the probability value of fire-flame is estimated. The proposed system has shown better performance when it compared to Toreyin's method which perform well among existing algorithms.

Analysis of Statistical Properties of Propagation Errors in DCT Coefficient-Dropping Transcoder (DCT 계수 제거 트랜스코딩에서의 전파 왜곡의 통계적 특성 분석)

  • Kim, Jin-Soo;Kim, Jae-Gon
    • Journal of Broadcast Engineering
    • /
    • v.12 no.6
    • /
    • pp.555-564
    • /
    • 2007
  • Discrete cosine transform (DCT) coefficient dropping is well recognized as an efficient rate adaptation transcoding in motion-compensated (MC)-DCT based MPEG-compressed videos. However, in this scheme, the errors incurred by the DCT coefficient-dropping are propagated and often result in significant visual quality degradation. This paper presents two propositions describing well the statistical properties of propagated errors. That is, we propose that the DCT error of the current frame is not correlated to the propagated errors of the previous frames. We also propose that the overall distortions in a given frame can be approximated as the sum of the DCT error of the current frame and the propagated errors from the previous frames. Then, it is shown that several computer simulations with different video sequences verify the effectiveness of the proposed statistical analyses.

Lightweight Attention-Guided Network with Frequency Domain Reconstruction for High Dynamic Range Image Fusion

  • Park, Jae Hyun;Lee, Keuntek;Cho, Nam Ik
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2022.06a
    • /
    • pp.205-208
    • /
    • 2022
  • Multi-exposure high dynamic range (HDR) image reconstruction, the task of reconstructing an HDR image from multiple low dynamic range (LDR) images in a dynamic scene, often produces ghosting artifacts caused by camera motion and moving objects and also cannot deal with washed-out regions due to over or under-exposures. While there has been many deep-learning-based methods with motion estimation to alleviate these problems, they still have limitations for severely moving scenes. They also require large parameter counts, especially in the case of state-of-the-art methods that employ attention modules. To address these issues, we propose a frequency domain approach based on the idea that the transform domain coefficients inherently involve the global information from whole image pixels to cope with large motions. Specifically we adopt Residual Fast Fourier Transform (RFFT) blocks, which allows for global interactions of pixels. Moreover, we also employ Depthwise Overparametrized convolution (DO-conv) blocks, a convolution in which each input channel is convolved with its own 2D kernel, for faster convergence and performance gains. We call this LFFNet (Lightweight Frequency Fusion Network), and experiments on the benchmarks show reduced ghosting artifacts and improved performance up to 0.6dB tonemapped PSNR compared to recent state-of-the-art methods. Our architecture also requires fewer parameters and converges faster in training.

  • PDF