• Title/Summary/Keyword: standard algorithm

Search Result 2,739, Processing Time 0.03 seconds

A Study on High Speed LDPC Decoder Algorithm Based on DVB-S2 Standard (멀티미디어 기반 해상통신을 위한 DVB-S2 기반 고속 LDPC 복호를 위한 알고리즘에 관한 연구)

  • Jung, Ji Won;Kwon, Hae Chan;Kim, Yeong Ju;Park, Sang Hyuk;Lee, Seong Ro
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.38C no.3
    • /
    • pp.311-317
    • /
    • 2013
  • In this paper, we proposed high speed LDPC decoding algorithm based on DVB-S2 standard for applying marine communications in order to multimedia transmission. For implementing the high speed LDPC decoder, HSS algorithm which reduce the iteration numbers without performance degradation is applied. In HSS algorithm, check node update units are update at the same time of bit node update. HSS can be accelerated to the decoding speed because it does not need to separate calculation of the bit nodes, However, check node calculation blocks need many clocks because of just one memory is used. Therefore, this paper proposed partial memory structure in order to reduced the delay and high speed decoder is possible. The results of the simulation, when the max number of iteration set to 30 times, decoding throughput of HSS algorithm is 326 Mbit/s and decoding speed of proposed algorithm is 2.29 Gbit/s. So, decoding speed of proposed algorithm more than 7 times could be obtained compared to the HSS algorithm.

A Simple Connection Pruning Algorithm and its Application to Simulated Random Signal Classification (연결자 제거를 위한 간단한 알고리즘과 모의 랜덤 신호 분류에의 응용)

  • Won, Yong-Gwan;Min, Byeong-Ui
    • The Transactions of the Korea Information Processing Society
    • /
    • v.3 no.2
    • /
    • pp.381-389
    • /
    • 1996
  • A simple modification of the standard back-propagation algorithm to eliminate redundant connections(weights and biases) is described. It was motivated by speculations from the distribution of the magnitudes of the weights and the biases, analysis of the classification boundary, and the nonlinearity of the sigmoid function. After initial training, this algorithm eliminates all connections of which magnitude is below a threshold by setting them to zero. The algorithm then conducts retraining in which all weights and biases are adjusted to allow important ones to recover. In studies with Boolean functions, the algorithm reconstructed the theoretical minimum architecture and eliminated the connections which are not necessary to solve the functions. For simulated random signal classification problems, the algorithm produced the result which is consistent with the idea that easier problems require simpler networks and yield lower misclassification rates. Furthermore, in comparison, our algorithm produced better generalization than the standard algorithm by reducing over fitting and pattern memorization problems.

  • PDF

Low Computational Complexity LDPC Decoding Algorithms for 802.11n Standard (802.11n 규격에서의 저복잡도 LDPC 복호 알고리즘)

  • Kim, Min-Hyuk;Park, Tae-Doo;Jung, Ji-Won;Lee, Seong-Ro;Jung, Min-A
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.35 no.2C
    • /
    • pp.148-154
    • /
    • 2010
  • In this paper, we first review LDPC codes in general and a belief propagation algorithm that works in logarithm domain. LDPC codes, which is chosen 802.11n for wireless local access network(WLAN) standard are required a large number of computation due to large size of coded block and iteration. Therefore, we presented three kinds of low computational algorithm for LDPC codes. First, sequential decoding with partial group is proposed. It has same H/W complexity, and fewer number of iteration's are required at same performance in comparison with conventional decoder algorithm. Secondly, we have apply early stop algorithm. This method is reduced number of unnecessary iteration. Third, early detection method for reducing the computational complexity is proposed. Using a confidence criterion, some bit nodes and check node edges are detected early on during decoding. Through the simulation, we knew that the iteration number are reduced by half using subset algorithm and early stop algorithm is reduced more than one iteration and computational complexity of early detected method is about 30% offs in case of check node update, 94% offs in case of check node update compared to conventional scheme.

Harmony search based, improved Particle Swarm Optimizer for minimum cost design of semi-rigid steel frames

  • Hadidi, Ali;Rafiee, Amin
    • Structural Engineering and Mechanics
    • /
    • v.50 no.3
    • /
    • pp.323-347
    • /
    • 2014
  • This paper proposes a Particle Swarm Optimization (PSO) algorithm, which is improved by making use of the Harmony Search (HS) approach and called HS-PSO algorithm. A computer code is developed for optimal sizing design of non-linear steel frames with various semi-rigid and rigid beam-to-column connections based on the HS-PSO algorithm. The developed code selects suitable sections for beams and columns, from a standard set of steel sections such as American Institute of Steel Construction (AISC) wide-flange W-shapes, such that the minimum total cost, which comprises total member plus connection costs, is obtained. Stress and displacement constraints of AISC-LRFD code together with the size constraints are imposed on the frame in the optimal design procedure. The nonlinear moment-rotation behavior of connections is modeled using the Frye-Morris polynomial model. Moreover, the P-${\Delta}$ effects of beam-column members are taken into account in the non-linear structural analysis. Three benchmark design examples with several types of connections are presented and the results are compared with those of standard PSO and of other researches as well. The comparison shows that the proposed HS-PSO algorithm performs better both than the PSO and the Big Bang-Big Crunch (BB-BC) methods.

Development of an Optimized Feature Extraction Algorithm for Throat Signal Analysis

  • Jung, Young-Giu;Han, Mun-Sung;Lee, Sang-Jo
    • ETRI Journal
    • /
    • v.29 no.3
    • /
    • pp.292-299
    • /
    • 2007
  • In this paper, we present a speech recognition system using a throat microphone. The use of this kind of microphone minimizes the impact of environmental noise. Due to the absence of high frequencies and the partial loss of formant frequencies, previous systems using throat microphones have shown a lower recognition rate than systems which use standard microphones. To develop a high performance automatic speech recognition (ASR) system using only a throat microphone, we propose two methods. First, based on Korean phonological feature theory and a detailed throat signal analysis, we show that it is possible to develop an ASR system using only a throat microphone, and propose conditions of the feature extraction algorithm. Second, we optimize the zero-crossing with peak amplitude (ZCPA) algorithm to guarantee the high performance of the ASR system using only a throat microphone. For ZCPA optimization, we propose an intensification of the formant frequencies and a selection of cochlear filters. Experimental results show that this system yields a performance improvement of about 4% and a reduction in time complexity of 25% when compared to the performance of a standard ZCPA algorithm on throat microphone signals.

  • PDF

The Improved Deblocking Algorithm for Low-bit Rate H.264/AVC (Low-bit Rate H.264/AVC 비디오에 적합한 개선된 디블럭킹 알고리즘)

  • Kwon, Dong-Jin;Kwak, Nae-Joung;Ryu, Sung-Pil
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2006.11a
    • /
    • pp.499-502
    • /
    • 2006
  • H.264/MPEG4 Advanced Video coding joint standard needs deblocking filter of decoder. We propose a better deblocking algorithm ensuring picture quality even if it is low bit-rate and bandwidth in MPEG-4 video. The complexity diminishes in proposed deblocking algorithm because it uses only simple shift, addition and comparison. We handle dividing into complexity area, medium area and simple area after counting boundary intensity of mask block to identify presence of block effects. As a result of experiment, we make certain of that block effects reduces in proposed deblocking algorithm.

  • PDF

Early Coding Unit-Splitting Termination Algorithm for High Efficiency Video Coding (HEVC)

  • Goswami, Kalyan;Kim, Byung-Gyu;Jun, Dongsan;Jung, Soon-Heung;Choi, Jin Soo
    • ETRI Journal
    • /
    • v.36 no.3
    • /
    • pp.407-417
    • /
    • 2014
  • A new-generation video coding standard, named High Efficiency Video Coding (HEVC), has recently been developed by JCT-VC. This new standard provides a significant improvement in picture quality, especially for high-resolution videos. However, one the most important challenges in HEVC is time complexity. A quadtree-based structure is created for the encoding and decoding processes and the rate-distortion (RD) cost is calculated for all possible dimensions of coding units in the quadtree. This provides a high encoding quality, but also causes computational complexity. We focus on a reduction scheme of the computational complexity and propose a new approach that can terminate the quadtree-based structure early, based on the RD costs of the parent and current levels. Our proposed algorithm is compared with HEVC Test Model version 10.0 software and a previously proposed algorithm. Experimental results show that our algorithm provides a significant time reduction for encoding, with only a small loss in video quality.

A Joint Resource Allocation Scheme for Relay Enhanced Multi-cell Orthogonal Frequency Division Multiple Networks

  • Fu, Yaru;Zhu, Qi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.7 no.2
    • /
    • pp.288-307
    • /
    • 2013
  • This paper formulates resource allocation for decode-and-forward (DF) relay assisted multi-cell orthogonal frequency division multiple (OFDM) networks as an optimization problem taking into account of inter-cell interference and users fairness. To maximize the transmit rate of system we propose a joint interference coordination, subcarrier and power allocation algorithm. To reduce the complexity, this semi-distributed algorithm divides the primal optimization into three sub-optimization problems, which transforms the mixed binary nonlinear programming problem (BNLP) into standard convex optimization problems. The first layer optimization problem is used to get the optimal subcarrier distribution index. The second is to solve the problem that how to allocate power optimally in a certain subcarrier distribution order. Based on the concept of equivalent channel gain (ECG) we transform the max-min function into standard closed expression. Subsequently, with the aid of dual decomposition, water-filling theorem and iterative power allocation algorithm the optimal solution of the original problem can be got with acceptable complexity. The third sub-problem considers dynamic co-channel interference caused by adjacent cells and redistributes resources to achieve the goal of maximizing system throughput. Finally, simulation results are provided to corroborate the proposed algorithm.

COMPARISON OF RED TIDE DETECTION BY A NEW RED TIDE INDEX METHOD AND STANDARD BIO-OPTICAL ALGORITHM APPLIED TO SEA WIFS IMAGERY IN OPTICALLY COMPLEX CASE-II WATERS

  • Shanmugam Palanisamy;Ahn Yu-Hwan
    • Proceedings of the KSRS Conference
    • /
    • 2005.10a
    • /
    • pp.445-449
    • /
    • 2005
  • Various methods to detect the phytoplankton/red tide blooms in the oceanic waters have been developed and tested on satellite ocean color imagery since the last two and half decades, but accurate detection of blooms with these methods remains challenging in optically complex turbid waters, mainly because of the eventual interference of absorbing and scattering properties of dissolved organic and particulate inorganic matters with these methods. The present study introduces a new method called Red tide Index (Rl), providing indices which behave as a good measure of detecting red tide algal blooms in high scattering and absorbing waters of the Korean South Sea and Yellow Sea. The effectiveness of this method in identifying and locating red tides is compared with the standard Ocean Chlorophyll 4 (OC4) bio-optical algorithm applied to SeaWiFS ocean imagery, acquired during two bloom episodes on 27 March 2002 and 28 September 2003. The result revealed that OC4 bio-optical algorithm falsely identifies red tide blooms in areas abundance in colored dissolved organic and particulate inorganic matter constituents associated with coastal areas, estuaries and river mouths, whereas red tide index provides improved capability of detecting, predicting and monitoring of these blooms in both clear and turbid waters.

  • PDF

A Pivot And Probe Algorithm(PARA) for Network Optimization

  • Moonsig Kang;Kim, Young-Moon
    • Korean Management Science Review
    • /
    • v.15 no.1
    • /
    • pp.1-12
    • /
    • 1998
  • This paper discusses a new algorithm, the PAPANET (Pivot And Probe Algorithm for NETwork optimization), for solving linear, capacitated linear network flow problem (NPs), PAPANET is a variation and specialization of the Pivot And Probe Algorithm (PAPA) developed by Sethi and Thompson, published in 1983-1984. PAPANET first solves an initial relaxed NP (RNP) with all the nodes from the original problem and a limited set of arcs (possibly all the artificial and slack arcs). From the arcs not considered in the current relaxation, we PROBE to identify candidate arcs that violate the current solution's dual constraints maximally. Candidate arcs are added to the RNP, and this new RNP is solved to optimality. This candidate pricing procedure and pivoting continue until all the candidate arcs price unfavorably and all of the dual constraints corresponding to the other, so-called noncandidate arcs, are satisfied. The implementation of PAPANET requires significantly fewer arcs and less solution CPU time than is required by the standard network simplex method implementation upon which it is based. Computational tests on randomly generated NPs indicate that our PAPANET implementation requires up to 40-50% fewer pivots and 30-40% less solution CPU time than is required by the comparable standard network simplex implementation from which it is derived.

  • PDF