• Title/Summary/Keyword: Generalization Error

Search Result 110, Processing Time 0.024 seconds

Multilayer Neural Network Using Delta Rule: Recognitron III (텔타규칙을 이용한 다단계 신경회로망 컴퓨터:Recognitron III)

  • 김춘석;박충규;이기한;황희영
    • The Transactions of the Korean Institute of Electrical Engineers
    • /
    • v.40 no.2
    • /
    • pp.224-233
    • /
    • 1991
  • The multilayer expanson of single layer NN (Neural Network) was needed to solve the linear seperability problem as shown by the classic example using the XOR function. The EBP (Error Back Propagation ) learning rule is often used in multilayer Neural Networks, but it is not without its faults: 1)D.Rimmelhart expanded the Delta Rule but there is a problem in obtaining Ca from the linear combination of the Weight matrix N between the hidden layer and the output layer and H, wich is the result of another linear combination between the input pattern and the Weight matrix M between the input layer and the hidden layer. 2) Even if using the difference between Ca and Da to adjust the values of the Weight matrix N between the hidden layer and the output layer may be valid is correct, but using the same value to adjust the Weight matrixd M between the input layer and the hidden layer is wrong. Recognitron III was proposed to solve these faults. According to simulation results, since Recognitron III does not learn the three layer NN itself, but divides it into several single layer NNs and learns these with learning patterns, the learning time is 32.5 to 72.2 time faster than EBP NN one. The number of patterns learned in a EBP NN with n input and output cells and n+1 hidden cells are 2**n, but n in Recognitron III of the same size. [5] In the case of pattern generalization, however, EBP NN is less than Recognitron III.

  • PDF

Improved Pattern Recoginition Coding System of a Handwriting Character with 3D (3D Magnetic Ball을 이용한 필기체 인식 향상 Coding System)

  • Sim, Kyu Seung;Lee, Jae Hong;Lee, Byoung Yup
    • The Journal of the Korea Contents Association
    • /
    • v.13 no.9
    • /
    • pp.10-19
    • /
    • 2013
  • This Paper proposed the development of new magnetic sensor and recognition system to expendite pattern recognition of a handwriting character. Received character graphics should be performed the session and balancing and no extraction of end points, bend points and juntions separately. The Artifical intelligence algorithm is adapted to structure snalysis and recognition process by individual basic letter dictionary except for the handwriing character graphic dictionaryimproving error of recognition algorithm and enomous dictionary for generalization. In this Paper, recognition rate of the received character are compared with pre registered character at letter dictionary for performance test of magnetic ball sensor. As a result of unicode conversion and eomparison, the artificial intelligence study have recognition rate more than 95% at initial recognition rate of 70%.

A Study on the Trend Change Point of NBUE-property

  • Kim, Dae-Kyung
    • Communications for Statistical Applications and Methods
    • /
    • v.3 no.2
    • /
    • pp.275-282
    • /
    • 1996
  • A life distribution F with survival function $\overline{F}$=1-F, finite mean $\mu$ and mean residual life m(t) is said to be NBUE(NWUE) if m(t)$\leq$($\geq$) .$\mu$ for t$\geq$0. This NBUE property can equivalently be characterized by the fact that $\varphi$(u)$\geq$($\leq$)u for 0$\leq$u$\leq$1, where $\varphi$(u) is the scaled total-time-on test transform of F. A generalization of the NBUE properties is that there is a value of p such that $\varphi$(u)\geq.u$ for 0$\leq$u$\leq$p and $\varphi$(u)\leq$$\leq$u$\leq$1, or vice versa. This means that we have a trend change in the NBUE property. In this paper we point out an error of Klefsjo's paper (1988). He erroneously takes advantage of trend change point of failure rate to calculate the empirical test size and power in lognormal distribution. We solves the trend change point of mean residual lifetime and recalculate the empirical test size and power of Klefsjo (1988) in mocensoring case.

  • PDF

A Panel Analysis on the Cross Border E-commerce Trade: Evidence from ASEAN Countries

  • HE, Yugang;WANG, Jingnan
    • The Journal of Asian Finance, Economics and Business
    • /
    • v.6 no.2
    • /
    • pp.95-104
    • /
    • 2019
  • Along with the economic globalization and network generalization, this provides a good opportunity to the development of cross-border e-commerce trade. Based on this background, this paper sets ASEAN countries as an example to exploit the determinants of cross-border e-commerce trade including the export and the import, respectively. The panel data from the year of 1998 to 2016 will be employed to estimate the relationship between cross-border e-commerce trade and relevant variables under the dynamic ordinary least squares and the error correction model. The findings of this paper show that there is a long-run relationship between cross-border e-commerce trade and relevant variables. Generally speaking, the GDP(+) and real exchange rate(-export & +import) have an effect on cross-border e-commerce trade. However, the population (+) and the terms of trade (-) only have an effect on cross-border e-commerce import. The empirical evidences show that the GDP and the real exchange rate always affect the development of cross-border e-commerce trade. Therefore, all ASEAN countries should try their best to develop the economic growth and focus on the exchange rate regime so as to meet the need of cross-border e-commerce trade development.

Bayesian and maximum likelihood estimations from exponentiated log-logistic distribution based on progressive type-II censoring under balanced loss functions

  • Chung, Younshik;Oh, Yeongju
    • Communications for Statistical Applications and Methods
    • /
    • v.28 no.5
    • /
    • pp.425-445
    • /
    • 2021
  • A generalization of the log-logistic (LL) distribution called exponentiated log-logistic (ELL) distribution on lines of exponentiated Weibull distribution is considered. In this paper, based on progressive type-II censored samples, we have derived the maximum likelihood estimators and Bayes estimators for three parameters, the survival function and hazard function of the ELL distribution. Then, under the balanced squared error loss (BSEL) and the balanced linex loss (BLEL) functions, their corresponding Bayes estimators are obtained using Lindley's approximation (see Jung and Chung, 2018; Lindley, 1980), Tierney-Kadane approximation (see Tierney and Kadane, 1986) and Markov Chain Monte Carlo methods (see Hastings, 1970; Gelfand and Smith, 1990). Here, to check the convergence of MCMC chains, the Gelman and Rubin diagnostic (see Gelman and Rubin, 1992; Brooks and Gelman, 1997) was used. On the basis of their risks, the performances of their Bayes estimators are compared with maximum likelihood estimators in the simulation studies. In this paper, research supports the conclusion that ELL distribution is an efficient distribution to modeling data in the analysis of survival data. On top of that, Bayes estimators under various loss functions are useful for many estimation problems.

Multiple-threshold asymmetric volatility models for financial time series (비대칭 금융 시계열을 위한 다중 임계점 변동성 모형)

  • Lee, Hyo Ryoung;Hwang, Sun Young
    • The Korean Journal of Applied Statistics
    • /
    • v.35 no.3
    • /
    • pp.347-356
    • /
    • 2022
  • This article is concerned with asymmetric volatility models for financial time series. A generalization of standard single-threshold volatility model is discussed via multiple-threshold in which we specialize to twothreshold case for ease of presentation. An empirical illustration is made by analyzing S&P500 data from NYSE (New York Stock Exchange). For comparison measures between competing models, parametric bootstrap method is used to generate forecast distributions from which summary statistics of CP (Coverage Probability) and PE (Prediction Error) are obtained. It is demonstrated that our suggestion is useful in the field of asymmetric volatility analysis.

Simultaneous Motion Recognition Framework using Data Augmentation based on Muscle Activation Model (근육 활성화 모델 기반의 데이터 증강을 활용한 동시 동작 인식 프레임워크)

  • Sejin Kim;Wan Kyun Chung
    • The Journal of Korea Robotics Society
    • /
    • v.19 no.2
    • /
    • pp.203-212
    • /
    • 2024
  • Simultaneous motion is essential in the activities of daily living (ADL). For motion intention recognition, surface electromyogram (sEMG) and corresponding motion label is necessary. However, this process is time-consuming and it may increase the burden of the user. Therefore, we propose a simultaneous motion recognition framework using data augmentation based on muscle activation model. The model consists of multiple point sources to be optimized while the number of point sources and their initial parameters are automatically determined. From the experimental results, it is shown that the framework has generated the data which are similar to the real one. This aspect is quantified with the following two metrics: structural similarity index measure (SSIM) and mean squared error (MSE). Furthermore, with k-nearest neighbor (k-NN) or support vector machine (SVM), the classification accuracy is also enhanced with the proposed framework. From these results, it can be concluded that the generalization property of the training data is enhanced and the classification accuracy is increased accordingly. We expect that this framework reduces the burden of the user from the excessive and time-consuming data acquisition.

Prediction of the remaining time and time interval of pebbles in pebble bed HTGRs aided by CNN via DEM datasets

  • Mengqi Wu;Xu Liu;Nan Gui;Xingtuan Yang;Jiyuan Tu;Shengyao Jiang;Qian Zhao
    • Nuclear Engineering and Technology
    • /
    • v.55 no.1
    • /
    • pp.339-352
    • /
    • 2023
  • Prediction of the time-related traits of pebble flow inside pebble-bed HTGRs is of great significance for reactor operation and design. In this work, an image-driven approach with the aid of a convolutional neural network (CNN) is proposed to predict the remaining time of initially loaded pebbles and the time interval of paired flow images of the pebble bed. Two types of strategies are put forward: one is adding FC layers to the classic classification CNN models and using regression training, and the other is CNN-based deep expectation (DEX) by regarding the time prediction as a deep classification task followed by softmax expected value refinements. The current dataset is obtained from the discrete element method (DEM) simulations. Results show that the CNN-aided models generally make satisfactory predictions on the remaining time with the determination coefficient larger than 0.99. Among these models, the VGG19+DEX performs the best and its CumScore (proportion of test set with prediction error within 0.5s) can reach 0.939. Besides, the remaining time of additional test sets and new cases can also be well predicted, indicating good generalization ability of the model. In the task of predicting the time interval of image pairs, the VGG19+DEX model has also generated satisfactory results. Particularly, the trained model, with promising generalization ability, has demonstrated great potential in accurately and instantaneously predicting the traits of interest, without the need for additional computational intensive DEM simulations. Nevertheless, the issues of data diversity and model optimization need to be improved to achieve the full potential of the CNN-aided prediction tool.

An Intelligent Intrusion Detection Model Based on Support Vector Machines and the Classification Threshold Optimization for Considering the Asymmetric Error Cost (비대칭 오류비용을 고려한 분류기준값 최적화와 SVM에 기반한 지능형 침입탐지모형)

  • Lee, Hyeon-Uk;Ahn, Hyun-Chul
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.4
    • /
    • pp.157-173
    • /
    • 2011
  • As the Internet use explodes recently, the malicious attacks and hacking for a system connected to network occur frequently. This means the fatal damage can be caused by these intrusions in the government agency, public office, and company operating various systems. For such reasons, there are growing interests and demand about the intrusion detection systems (IDS)-the security systems for detecting, identifying and responding to unauthorized or abnormal activities appropriately. The intrusion detection models that have been applied in conventional IDS are generally designed by modeling the experts' implicit knowledge on the network intrusions or the hackers' abnormal behaviors. These kinds of intrusion detection models perform well under the normal situations. However, they show poor performance when they meet a new or unknown pattern of the network attacks. For this reason, several recent studies try to adopt various artificial intelligence techniques, which can proactively respond to the unknown threats. Especially, artificial neural networks (ANNs) have popularly been applied in the prior studies because of its superior prediction accuracy. However, ANNs have some intrinsic limitations such as the risk of overfitting, the requirement of the large sample size, and the lack of understanding the prediction process (i.e. black box theory). As a result, the most recent studies on IDS have started to adopt support vector machine (SVM), the classification technique that is more stable and powerful compared to ANNs. SVM is known as a relatively high predictive power and generalization capability. Under this background, this study proposes a novel intelligent intrusion detection model that uses SVM as the classification model in order to improve the predictive ability of IDS. Also, our model is designed to consider the asymmetric error cost by optimizing the classification threshold. Generally, there are two common forms of errors in intrusion detection. The first error type is the False-Positive Error (FPE). In the case of FPE, the wrong judgment on it may result in the unnecessary fixation. The second error type is the False-Negative Error (FNE) that mainly misjudges the malware of the program as normal. Compared to FPE, FNE is more fatal. Thus, when considering total cost of misclassification in IDS, it is more reasonable to assign heavier weights on FNE rather than FPE. Therefore, we designed our proposed intrusion detection model to optimize the classification threshold in order to minimize the total misclassification cost. In this case, conventional SVM cannot be applied because it is designed to generate discrete output (i.e. a class). To resolve this problem, we used the revised SVM technique proposed by Platt(2000), which is able to generate the probability estimate. To validate the practical applicability of our model, we applied it to the real-world dataset for network intrusion detection. The experimental dataset was collected from the IDS sensor of an official institution in Korea from January to June 2010. We collected 15,000 log data in total, and selected 1,000 samples from them by using random sampling method. In addition, the SVM model was compared with the logistic regression (LOGIT), decision trees (DT), and ANN to confirm the superiority of the proposed model. LOGIT and DT was experimented using PASW Statistics v18.0, and ANN was experimented using Neuroshell 4.0. For SVM, LIBSVM v2.90-a freeware for training SVM classifier-was used. Empirical results showed that our proposed model based on SVM outperformed all the other comparative models in detecting network intrusions from the accuracy perspective. They also showed that our model reduced the total misclassification cost compared to the ANN-based intrusion detection model. As a result, it is expected that the intrusion detection model proposed in this paper would not only enhance the performance of IDS, but also lead to better management of FNE.

Residual Strength of Corroded Reinforced Concrete Beams Using an Adaptive Model Based on ANN

  • Imam, Ashhad;Anifowose, Fatai;Azad, Abul Kalam
    • International Journal of Concrete Structures and Materials
    • /
    • v.9 no.2
    • /
    • pp.159-172
    • /
    • 2015
  • Estimation of the residual strength of corroded reinforced concrete beams has been studied from experimental and theoretical perspectives. The former is arduous as it involves casting beams of various sizes, which are then subjected to various degrees of corrosion damage. The latter are static; hence cannot be generalized as new coefficients need to be re-generated for new cases. This calls for dynamic models that are adaptive to new cases and offer efficient generalization capability. Computational intelligence techniques have been applied in Construction Engineering modeling problems. However, these techniques have not been adequately applied to the problem addressed in this paper. This study extends the empirical model proposed by Azad et al. (Mag Concr Res 62(6):405-414, 2010), which considered all the adverse effects of corrosion on steel. We proposed four artificial neural networks (ANN) models to predict the residual flexural strength of corroded RC beams using the same data from Azad et al. (2010). We employed two modes of prediction: through the correction factor ($C_f$) and through the residual strength ($M_{res}$). For each mode, we studied the effect of fixed and random data stratification on the performance of the models. The results of the ANN models were found to be in good agreement with experimental values. When compared with the results of Azad et al. (2010), the ANN model with randomized data stratification gave a $C_f$-based prediction with up to 49 % improvement in correlation coefficient and 92 % error reduction. This confirms the reliability of ANN over the empirical models.