• Title/Summary/Keyword: polynomial approach

Search Result 274, Processing Time 0.022 seconds

SOC Verification Based on WGL

  • Du, Zhen-Jun;Li, Min
    • Journal of Korea Multimedia Society
    • /
    • v.9 no.12
    • /
    • pp.1607-1616
    • /
    • 2006
  • The growing market of multimedia and digital signal processing requires significant data-path portions of SoCs. However, the common models for verification are not suitable for SoCs. A novel model--WGL (Weighted Generalized List) is proposed, which is based on the general-list decomposition of polynomials, with three different weights and manipulation rules introduced to effect node sharing and the canonicity. Timing parameters and operations on them are also considered. Examples show the word-level WGL is the only model to linearly represent the common word-level functions and the bit-level WGL is especially suitable for arithmetic intensive circuits. The model is proved to be a uniform and efficient model for both bit-level and word-level functions. Then Based on the WGL model, a backward-construction logic-verification approach is presented, which reduces time and space complexity for multipliers to polynomial complexity(time complexity is less than $O(n^{3.6})$ and space complexity is less than $O(n^{1.5})$) without hierarchical partitioning. Finally, a construction methodology of word-level polynomials is also presented in order to implement complex high-level verification, which combines order computation and coefficient solving, and adopts an efficient backward approach. The construction complexity is much less than the existing ones, e.g. the construction time for multipliers grows at the power of less than 1.6 in the size of the input word without increasing the maximal space required. The WGL model and the verification methods based on WGL show their theoretical and applicable significance in SoC design.

  • PDF

Automatic NURBS Surface Generation from Unorganized Point Cloud Data (임의의 점 군 데이터로부터 NURBS 곡면의 자동생성)

  • Yoo, Dong-Jin
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.23 no.9 s.186
    • /
    • pp.200-207
    • /
    • 2006
  • In this paper a new approach which combines implicit surface scheme and NURBS surface interpolation method is proposed in order to generate a complete surface model from unorganized point cloud data. In the method a base surface was generated by creating smooth implicit surface from the input point cloud data through which the actual surface would pass. The implicit surface was defined by a combination of shape functions including quadratic polynomial function, cubic polynomial functions and radial basis function using adaptive domain decomposition method. In this paper voxel data which can be extracted easily from the base implicit surface were used in order to generate rectangular net with good quality using the normal projection and smoothing scheme. After generating the interior points and tangential vectors in each rectangular region considering the required accuracy, the NURBS surface were constructed by interpolating the rectangular array of points using boundary tangential vectors which assure C$^1$ continuity between rectangular patches. The validity and effectiveness of this new approach was demonstrated by performing numerical experiments for the various types of point cloud data.

MULTI-BLOCK BOUNDARY VALUE METHODS FOR ORDINARY DIFFERENTIAL AND DIFFERENTIAL ALGEBRAIC EQUATIONS

  • OGUNFEYITIMI, S.E.;IKHILE, M.N.O.
    • Journal of the Korean Society for Industrial and Applied Mathematics
    • /
    • v.24 no.3
    • /
    • pp.243-291
    • /
    • 2020
  • In this paper, multi-block generalized backward differentiation methods for numerical solutions of ordinary differential and differential algebraic equations are introduced. This class of linear multi-block methods is implemented as multi-block boundary value methods (MB2 VMs). The root distribution of the stability polynomial of the new class of methods are determined using the Wiener-Hopf factorization of a matrix polynomial for the purpose of their correct implementation. Numerical tests, showing the potential of such methods for output of multi-block of solutions of the ordinary differential equations in the new approach are also reported herein. The methods which output multi-block of solutions of the ordinary differential equations on application, are unlike the conventional linear multistep methods which output a solution at a point or the conventional boundary value methods and multi-block methods which output only a block of solutions per step. The MB2 VMs introduced herein is a novel approach at developing very large scale integration methods (VLSIM) in the numerical solution of differential equations.

Exploiting Neural Network for Temporal Multi-variate Air Quality and Pollutant Prediction

  • Khan, Muneeb A.;Kim, Hyun-chul;Park, Heemin
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.2
    • /
    • pp.440-449
    • /
    • 2022
  • In recent years, the air pollution and Air Quality Index (AQI) has been a pivotal point for researchers due to its effect on human health. Various research has been done in predicting the AQI but most of these studies, either lack dense temporal data or cover one or two air pollutant elements. In this paper, a hybrid Convolutional Neural approach integrated with recurrent neural network architecture (CNN-LSTM), is presented to find air pollution inference using a multivariate air pollutant elements dataset. The aim of this research is to design a robust and real-time air pollutant forecasting system by exploiting a neural network. The proposed approach is implemented on a 24-month dataset from Seoul, Republic of Korea. The predicted results are cross-validated with the real dataset and compared with the state-of-the-art techniques to evaluate its robustness and performance. The proposed model outperforms SVM, SVM-Polynomial, ANN, and RF models with 60.17%, 68.99%, 14.6%, and 6.29%, respectively. The model performs SVM and SVM-Polynomial in predicting O3 by 78.04% and 83.79%, respectively. Overall performance of the model is measured in terms of Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE) and the Root Mean Square Error (RMSE).

Target Observability Analysis of Time-to-go Polynomial Guidance Law (Time-to-go 다항식 유도 법칙의 표적 가관측성 분석)

  • Lee, Chang-Hun;Kim, Tae-Hun;Tahk, Min-Jea
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.38 no.7
    • /
    • pp.664-672
    • /
    • 2010
  • This paper provides the target observability analysis of time-to-go polynomial guidance law with bearing-only measurement. In this study, a direct approach is used to analyze the target observability. Since the observability condition of a constant-velocity target is given by the function of LOS angle only, the target observability characteristic is determined by substituting the closed form solution of LOS angle to the observability condition directly. The analysis results show that the target observability is depended on the choice of guidance gain, initial intercept condition and guidance command shape. After that this mathematical analysis result is evaluated and demonstrated by number of simulation.

The Grid Type Quadratic Assignment Problem Algorithm (그리드형 2차 할당문제 알고리즘)

  • Lee, Sang-Un
    • Journal of the Korea Society of Computer and Information
    • /
    • v.19 no.4
    • /
    • pp.91-99
    • /
    • 2014
  • TThis paper suggests an heuristic polynomial time algorithm to solve the optimal solution for QAP (quadratic assignment problem). While Hungarian algorithm is most commonly used for a linear assignment, there is no polynomial time algorithm for the QAP. The proposed algorithm derives a grid type layout among unit distances of a distance matrix. And, we apply max-flow/min-distance approach to assign this grid type layout in such a descending order way that the largest flow is matched to the smallest unit distance from flow matrix. Evidences from implementation results of the proposed algorithm on various numerical grid type QAP examples show that a solution to the QAP could be obtained by a polynomial algorithm.

A Novel Soft Computing Technique for the Shortcoming of the Polynomial Neural Network

  • Kim, Dongwon;Huh, Sung-Hoe;Seo, Sam-Jun;Park, Gwi-Tae
    • International Journal of Control, Automation, and Systems
    • /
    • v.2 no.2
    • /
    • pp.189-200
    • /
    • 2004
  • In this paper, we introduce a new soft computing technique that dwells on the ideas of combining fuzzy rules in a fuzzy system with polynomial neural networks (PNN). The PNN is a flexible neural architecture whose structure is developed through the modeling process. Unfortunately, the PNN has a fatal drawback in that it cannot be constructed for nonlinear systems with only a small amount of input variables. To overcome this limitation in the conventional PNN, we employed one of three principal soft computing components such as a fuzzy system. As such, a space of input variables is partitioned into several subspaces by the fuzzy system and these subspaces are utilized as new input variables to the PNN architecture. The proposed soft computing technique is achieved by merging the fuzzy system and the PNN into one unified framework. As a result, we can find a workable synergistic environment and the main characteristics of the two modeling techniques are harmonized. Thus, the proposed method alleviates the problems of PNN while providing superb performance. Identification results of the three-input nonlinear static function and nonlinear system with two inputs will be demonstrated to demonstrate the performance of the proposed approach.

Polynomial Approximation Approach to ECG Analysis and Tele-monitoring (다항식 근사를 이용한 심전도 분석 및 원격 모니터링)

  • Yu, Kee-Ho;Jeong, Gu-Young;Jung, Sung-Nam;No, Tae-Soo
    • Proceedings of the KSME Conference
    • /
    • 2001.06b
    • /
    • pp.42-47
    • /
    • 2001
  • Analyzing the ECG signal, we can find heart disease, for example, arrhythmia and myocardial infarction, etc. Particularly, detecting arrhythmia is more important, because serious arrhythmia can take away the life from patients within ten minutes. In this paper, we would like to introduce the signal processing for ECG analysis and the device made for wireless communication of ECG data. In the signal processing, the wavelet transform decomposes the ECG signal into high and low frequency components using wavelet function. Recomposing the high frequency bands including QRS complex, we can detect QRS complex and eliminate the noise from the original ECG signal. To recognize the ECG signal pattern, we adopted the polynomial approximation partially and statistical method. The ECG signal is divided into small parts based on QRS complex, and then, each part is approximated to the polynomials. Comparing the approximated ECG pattern with the database, we can detect and classify the heart disease. The ECG detection device consists of amplifier, filters, A/D converter and RF module. After amplification and filtering, the ECG signal is fed through the A/D converter to be digitalized. The digital ECG data is transmitted to the personal computer through the RF transceiver module and serial port.

  • PDF

A New Design Approach for Optimization of GA-based SOPNN (GA 기반 자기구성 다항식 뉴럴 네트워크의 최적화를 위한 새로운 설계 방법)

  • Park, Ho-Sung;Park, Byoung-Jun;Park, Keon-Jun;Oh, Sung-Kwun
    • Proceedings of the KIEE Conference
    • /
    • 2003.07d
    • /
    • pp.2627-2629
    • /
    • 2003
  • In this paper, we propose a new architecture of Genetic Algorithms(GAs)-based Self-Organizing Polynomial Neural Networks(SOPNN). The conventional SOPNN is based on the extended Group Method of Data Handling(GMDH) method and utilized the polynomial order (viz. linear, quadratic, and modified quadratic) as well as the number of node inputs fixed (selected in advance by designer) at Polynomial Neurons (or nodes) located in each layer through a growth process of the network. Moreover it does not guarantee that the SOPNN generated through learning has the optimal network architecture. But the proposed GA-based SOPNN enable the architecture to be a structurally more optimized networks, and to be much more flexible and preferable neural network than the conventional SOPNN. An aggregate performance index with a weighting factor is proposed in order to achieve a sound balance between approximation and generalization (predictive) abilities of the model. To evaluate the performance of the GA-based SOPNN, the model is experimented with using nonlinear system data.

  • PDF

On Practical Efficiency of Locally Parametric Nonparametric Density Estimation Based on Local Likelihood Function

  • Kang, Kee-Hoon;Han, Jung-Hoon
    • Communications for Statistical Applications and Methods
    • /
    • v.10 no.2
    • /
    • pp.607-617
    • /
    • 2003
  • This paper offers a practical comparison of efficiency between local likelihood approach and conventional kernel approach in density estimation. The local likelihood estimation procedure maximizes a kernel smoothed log-likelihood function with respect to a polynomial approximation of the log likelihood function. We use two types of data driven bandwidths for each method and compare the mean integrated squares for several densities. Numerical results reveal that local log-linear approach with simple plug-in bandwidth shows better performance comparing to the standard kernel approach in heavy tailed distribution. For normal mixture density cases, standard kernel estimator with the bandwidth in Sheather and Jones(1991) dominates the others in moderately large sample size.