• Title/Summary/Keyword: Complexity of Computation

Search Result 606, Processing Time 0.03 seconds

A Design of Modified Euclidean Algorithm using Finite State Machine (FSM을 이용한 수정된 유클리드 알고리즘 설계)

  • Kang, Sung-Jin
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.11 no.6
    • /
    • pp.2202-2206
    • /
    • 2010
  • In this paper, an architecture for modified Euclidean(ME) algorithm is proposed, which is using finite-state machine(FSM) instead of degree computation. Since the proposed architecture does not have degree computation circuits, it is possible to reduce the hardware complexity of RS(Reed-Solomon) decoder, so that a very high-speed RS decoder can be implemented. RS(255,239) decoder with the proposed architecture is implemented using Verilog-HDL and requires about 13% fewer gate counts than conventional one.

A COMPUTATION METHOD IN PERFORMANCE EVALUATION IN CELLULAR COMMUNICATION NETWORK UNDER THE GENERAL DISTRIBUTION MODEL

  • Kim, Kyung-Sup
    • Journal of the Korean Society for Industrial and Applied Mathematics
    • /
    • v.12 no.2
    • /
    • pp.119-131
    • /
    • 2008
  • The paper considers the computation method in the performance evaluation of cellular network in the phase-type distribution assumptions that the channel holding times induced from mobility are modeled by well-fitted distributions to reflect an actual situation. When ww consider a phase-type distribution model instead of exponential distribution, the complexity of the computation increase exponential even though the accuracy is improved. We consider an efficient numerical algorithm to compute the performance evaluations in cellular networks such as a handoff call dropping probability, new call blocking probability, and handoff arrival rate. Numerical experiment shows that numerical analysis results are well approximated to the results of simulation.

  • PDF

SOC Verification Based on WGL

  • Du, Zhen-Jun;Li, Min
    • Journal of Korea Multimedia Society
    • /
    • v.9 no.12
    • /
    • pp.1607-1616
    • /
    • 2006
  • The growing market of multimedia and digital signal processing requires significant data-path portions of SoCs. However, the common models for verification are not suitable for SoCs. A novel model--WGL (Weighted Generalized List) is proposed, which is based on the general-list decomposition of polynomials, with three different weights and manipulation rules introduced to effect node sharing and the canonicity. Timing parameters and operations on them are also considered. Examples show the word-level WGL is the only model to linearly represent the common word-level functions and the bit-level WGL is especially suitable for arithmetic intensive circuits. The model is proved to be a uniform and efficient model for both bit-level and word-level functions. Then Based on the WGL model, a backward-construction logic-verification approach is presented, which reduces time and space complexity for multipliers to polynomial complexity(time complexity is less than $O(n^{3.6})$ and space complexity is less than $O(n^{1.5})$) without hierarchical partitioning. Finally, a construction methodology of word-level polynomials is also presented in order to implement complex high-level verification, which combines order computation and coefficient solving, and adopts an efficient backward approach. The construction complexity is much less than the existing ones, e.g. the construction time for multipliers grows at the power of less than 1.6 in the size of the input word without increasing the maximal space required. The WGL model and the verification methods based on WGL show their theoretical and applicable significance in SoC design.

  • PDF

Fast Motion Estimation Algorithm Using Limited Sub-blocks (제한된 서브블록을 이용한 고속 움직임 추정 알고리즘)

  • Kim Seong-Hee;Oh Jeong-Su
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.31 no.3C
    • /
    • pp.258-263
    • /
    • 2006
  • Each pixel in a matching block does not equally contribute to block matching and the matching error is greatly affected by image complexity. On the basis of the facts, this paper proposes a fast motion estimation algorithm using some sub-blocks selected by the image complexity. The proposed algorithm divides a matching block into 16 sub-blocks, computes the image complexity in every sub-block, executes partial block matching using some sub-blocks with large complexity, and detects a motion vector. The simulation results show that the proposed algorithm brings about negligible image degradation, but can reduce a large amount of computation in comparison with conventional algorithms.

Architecture of 2-D DCT processor adopting accuracy comensator (정확도 보상기를 적용한 2차원 이산 코사인 변환 프로세서의 구조)

  • 김견수;장순화;김재호;손경식
    • Journal of the Korean Institute of Telematics and Electronics A
    • /
    • v.33A no.10
    • /
    • pp.168-176
    • /
    • 1996
  • This paper presetns a 2-D DCT architecture adopting accurac y compensator for reducing the hardware complexity and increasing processing speed in VL\ulcornerSI implementation. In the application fields such as moving pictures experts group (MPEG) and joint photographic experts group (JPEG), 2-D DCT processor must be implemented precisely enough to meet the accuracy specifications of the ITU-T H.261. Almost all of 2-D DCT processors have been implemented using many multiplications and accumulations of matrices and vectors. The number of multiplications and accumulations seriously influence on comlexity and speed of 20D DCT processor. In 2-D DCT with fixed-point calculations, the computation bit width must be sufficiently large for the above accuracy specifications. It makes the reduction of hardware complexity hard. This paper proposes the accuracy compensator which compensates the accuracy of the finite word length calculation. 2-D DCT processor with the proposed accuracy compensator shows fairly reduced hardware complexity and improved processing speed.

  • PDF

URBAN COMPLEXITY ESTIMATION INDICES BASED ON 3D DISCRETE WAVELET TRANSFORM OF REMOTELY SENSED IMAGERY;THE PRELIMINARY INTERPRETATION WITH LAND COVER MAP

  • Yoo, Hee-Young;Lee, Ki-Won;Kwon, Byung-Doo
    • Proceedings of the KSRS Conference
    • /
    • 2007.10a
    • /
    • pp.405-409
    • /
    • 2007
  • Each class in remotely sensed imagery has different spectral and spatial characteristics. Natural features have relatively smaller spatial changes than spectral changes. Meanwhile, urban area in which buildings, roads, and cars are included is inclined to face more changes of spatial variation than spectral one. This study aims to propose the new urban complexity index (UCI) based on the 3D DWT computation of remotely sensed imageries considering these characteristics. And then we analyze relation between index and land cover map. The 3DWUCI values are related to class and the indices of urban area are greater than natural area. The proposed UCI could be used to express effectively the standard of urban complexity over a wide area.

  • PDF

A Fast and Low-complexity Motion Estimation for HEVC

  • Kim, Sungoh;Park, Chansik;Chun, Hyungju;Kim, Jaemoon
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2013.06a
    • /
    • pp.173-175
    • /
    • 2013
  • In this paper, we propose a fast and low-complexity Motion Estimation (ME) algorithm for High Efficiency Video Coding (HEVC). Motion estimation occupies 77~81% of the amount of computation in HEVC. After all, the main key of codec implementation is to find a fast and low-complexity motion estimation algorithm and architecture. The proposed algorithm uses only 1% of the amount of operations compared to full search algorithm while maintaining compression performance with slight loss of 0.6% (BDBR).

  • PDF

A Variable Step-Size NLMS Algorithm with Low Complexity

  • Chung, Ik-Joo
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.3E
    • /
    • pp.93-98
    • /
    • 2009
  • In this paper, we propose a new VSS-NLMS algorithm through a simple modification of the conventional NLMS algorithm, which leads to a low complexity algorithm with enhanced performance. The step size of the proposed algorithm becomes smaller as the error signal is getting orthogonal to the input vector. We also show that the proposed algorithm is an approximated normalized version of the KZ-algorithm and requires less computation than the KZ-algorithm. We carried out a performance comparison of the proposed algorithm with the conventional NLMS and other VSS algorithms using an adaptive channel equalization model. It is shown that the proposed algorithm presents good convergence characteristics under both stationary and non-stationary environments despites its low complexity.

DESIGN-ORIENTED AERODYNAMIC ANALYSES OF HELICOPTER ROTOR IN HOVER (정지비행 헬리콥터 로터의 설계를 위한 공력해석)

  • Jung H.J.;Kim T.S.;Son C.H.;Joh C.Y.
    • Journal of computational fluids engineering
    • /
    • v.11 no.3 s.34
    • /
    • pp.1-7
    • /
    • 2006
  • Euler and Navier-Stokes flow analyses for helicopter rotor in hover were performed as low and high fidelity analysis models respectively for the future multidisciplinary design optimization(MDO). These design-oriented analyses possess several attributes such as variable complexity, sensitivity-computation capability and modularity which analysis models involved in MDO are recommended to provide with. To realize PC-based analyses for both fidelity models, reduction of flow domain was made by appling farfield boundary condition based on 3-dimensional point sink with simple momentum theory and also periodic boundary condition in the azimuthal direction. Correlations of thrust, torque and their sensitivities between low and high complexity models were tried to evaluate the applicability of these analysis models in MDO process. It was found that the low-fidelity Euler analysis model predicted inaccurate sensitivity derivatives at relatively high angle of attack.

Improved Method for the Macroblock-Level Deblocking Scheme

  • Le, Thanh Ha;Jung, Seung-Won;Baek, Seung-Jin;Ko, Sung-Jea
    • ETRI Journal
    • /
    • v.33 no.2
    • /
    • pp.194-200
    • /
    • 2011
  • This paper presents a deblocking method for video compression in which the blocking artifacts are effectively extracted and eliminated based on both spatial and frequency domain operations. Firstly, we use a probabilistic approach to analyze the performance of the conventional macroblock-level deblocking scheme. Then, based on the results of the analysis, an algorithm to reduce the computational complexity is introduced. Experimental results show that the proposed algorithm outperforms the conventional video coding methods in terms of computation complexity while coding efficiency is maintained.