• Title/Summary/Keyword: arithmetic errors

Search Result 50, Processing Time 0.021 seconds

An Approach for the NHPP Software Reliability Model Using Erlang Distribution (어랑 분포를 이용한 NHPP 소프트웨어 신뢰성장 모형에 관한 연구)

  • Kim Hee-Cheul;Choi Yue-Soon;Park Jong-Goo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.10 no.1
    • /
    • pp.7-14
    • /
    • 2006
  • The finite failure NHPP models proposed in the literature exhibit either constant, monotonic increasing or monotonic decreasing failure occurrence rates per fault. In this paper, we propose the Erlang reliability model, which can capture the increasing nature of the failure occurrence rate per fault. Equations to estimate the parameters of the Erlang finite failure NHPP model based on failure data collected in the form of inter-failure times are developed. For the sake of proposing shape parameter of the Erlang distribution, we used to the goodness-of-fit test of distribution. Data set, where the underlying failure process could not be adequately described by the existing models, which motivated the development of the Erlang model. Analysis of the failure data set which led us to the Erlang model, using arithmetic and Laplace trend tests, goodness-of-fit test, bias tests is presented.

NHPP Software Reliability Model based on Generalized Gamma Distribution (일반화 감마 분포를 이용한 NHPP 소프트웨어 신뢰도 모형에 관한 연구)

  • Kim, Hee-Cheul
    • Journal of the Korea Society of Computer and Information
    • /
    • v.10 no.6 s.38
    • /
    • pp.27-36
    • /
    • 2005
  • Finite failure NHPP models presented in the literature exhibit either constant, monotonic increasing or monotonic decreasing failure occurrence rates Per fault. This Paper Proposes reliability model using the generalized gamma distribution, which can capture the monotonic increasing(or monotonic decreasing) nature of the failure occurrence rate per fault. Equations to estimate the parameters of the generalized gamma finite failure NHPP model based on failure data collected in the form of interfailure times are developed. For the sake of proposing shape parameter of the generalized gamma distribution, used to the special pattern. Data set, where the underlying failure process could not be adequately described by the knowing models, which motivated the development of the gamma or Weibull model. Analysis of failure data set for the generalized gamma modell, using arithmetic and Laplace trend tests . goodness-of-fit test, bias tests is presented.

  • PDF

A Study of Counting Ability of Mathematical Low Achievers in Low Grades of Elementary School (초등학교 저학년 수학부진아의 수세기 능력 연구)

  • Kim, Soo-Mi
    • School Mathematics
    • /
    • v.12 no.2
    • /
    • pp.137-150
    • /
    • 2010
  • This study has planned to understand the counting ability of mathematical low achievers(MLAs) in low grades of elementary school. For this, four MLAs of second graders were chosen as a experimental group and four MLAs of fourth graders and two mid-achievers of second grades were chosen as a comparison group. Every students were individually interviewed and their data were analyzed with three respects: accuracy, speed, efficiency. The results are the following three. First, the experimental group is still not used to counting. They counted slowly, made many mistakes and errors, and used inefficient strategies during counting. Second, It is hard to expect natural improvement of their counting performance as growing up. Finally, the explicit correlation between counting ability and arithmetic ability is not drown in this study, but the possibility still remains. These results tell us that we need to stress counting activity to MLAs when they are low graders.

  • PDF

The Study for NHPP Software Reliability Growth Model based on Exponentiated Exponential Distribution (지수화 지수 분포에 의존한 NHPP 소프트웨어 신뢰성장 모형에 관한 연구)

  • Kim, Hee-Cheul
    • Journal of the Korea Society of Computer and Information
    • /
    • v.11 no.5 s.43
    • /
    • pp.9-18
    • /
    • 2006
  • Finite failure NHPP models presented in the literature exhibit either constant, monotonic increasing or monotonic decreasing failure occurrence rates per fault. In this paper, Goel-Okumoto and Yamada-Ohba-Osaki model was reviewed, proposes the exponentiated exponential distribution reliability model, which maked out efficiency substituted for gamma and Weibull model(2 parameter shape illustrated by Gupta and Kundu(2001) Algorithm to estimate the parameters used to maximum likelihood estimator and bisection method, model selection based on SSE, AIC statistics and Kolmogorov distance, for the sake of efficient model, was employed. Analysis of failure using NTDS data set for the sake of proposing shape parameter of the exponentiated exponential distribution was employed. This analysis of failure data compared with the exponentiated exponential distribution model and the existing model (using arithmetic and Laplace trend tests, bias tests) is presented.

  • PDF

The Study for NHPP Software Reliability Model based on Chi-Square Distribution (카이제곱 NHPP에 의한 소프트웨어 신뢰성 모형에 관한 연구)

  • Kim, Hee-Cheul
    • Journal of the Korea Society of Computer and Information
    • /
    • v.11 no.1 s.39
    • /
    • pp.45-53
    • /
    • 2006
  • Finite failure NHPP models presented in the literature exhibit either constant, monotonic increasing or monotonic decreasing failure occurrence rates per fault. In this paper, Goel-Okumoto and Yamada-Ohba-Osaki model was reviewed, proposes the $x^2$ reliability model, which can capture the increasing nature of the failure occurrence rate per fault. Algorithm to estimate the parameters used to maximum likelihood estimator and bisection method, model selection based on SSE, AIC statistics and Kolmogorov distance, for the sake of efficient model, was employed. Analysis of failure using real data set, SYS2(Allen P.Nikora and Michael R.Lyu), for the sake of proposing shape parameter of the $x^2$ distribution using the degree of freedom, was employed. This analysis of failure data compared with the $x^2$ model and the existing model using arithmetic and Laplace trend tests, Kolmogorov test is presented.

  • PDF

Evaluation of Raingauge Network using Area Average Rainfall Estimation and the Estimation Error (면적평균강우량 산정을 통한 강우관측망 평가 및 추정오차)

  • Lee, Ji Ho;Jun, Hwan Don
    • Journal of Wetlands Research
    • /
    • v.16 no.1
    • /
    • pp.103-112
    • /
    • 2014
  • Area average rainfall estimation is important to determine the exact amount of the available water resources and the essential input data for rainfall-runoff analysis. Like that, the necessary criterion for accurate area average rainfall estimate is the uniform spatial distribution of raingauge network. In this study, we suggest the spatial distribution evaluation methodology of raingauge network to estimate better area average rainfall and after the suggested method is applied to Han River and Geum River basin. The spatial distribution of rainfall network can be quantified by the nearest neighbor index. In order to evaluate the effects of the spatial distribution of rainfall network by each basin, area average rainfall was estimated by arithmetic mean method, the Thiessen's weighting method and estimation theory for 2013's rainfall event, and evaluated the involved errors by each cases. As a result, it can be found that the estimation error at the best basin of spatial distribution was lower than the worst basin of spatial distribution.

Improved Real-Time Mean-Shift Face Tracking by Readjusting Detected Face Region Histogram (검출된 얼굴 영역 히스토그램 재조정을 통한 개선된 실시간 평균이동 얼굴 추적 방식)

  • Kim, Gui-sik;Lee, Jae-sung
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2013.10a
    • /
    • pp.195-198
    • /
    • 2013
  • Recognition and Tracking of interesting object is the significant field in Computer Vision. Mean-Shift algorithm have chronic problems that some errors are occurred when histogram of tracking area is similar to another area. in this paper, we propose to solve the problem. Each algorithm blocks skin color filtering, face detect and Mean-Shift started consecutive order assists better operation of the next algorithm. Avoid to operations of the overhead of tracking area similar to a histogram distribution areas overlap only consider the number of white pixels by running the Viola-Jones algorithm, simple arithmetic increases the convergence of the Mean-Shift. The experimental results, it comes to 78% or more of white pixels in the Mean-Shift search area, only if the recognition of the face area when it is configured to perform a Viola-Jones algorithm is tracking the object, was 100 percent successful.

  • PDF

A New Program to Design Residual Treatment Trains at Water Treatment Plants (정수장 배출수처리시설 설계 프로그램의 개발)

  • Bae, Byung-Uk;Her, Kuk;Joo, Dae-Sung;Jeong, Yeon-Gu;Kim, Young-Il;Ha, Chang-Won
    • Journal of Korean Society of Environmental Engineers
    • /
    • v.29 no.3
    • /
    • pp.277-282
    • /
    • 2007
  • For more accurate and practical design of the residual treatment train at water treatment plants(WTPs), a computational program based on the commercial spreadsheet, Microsoft Excel, was developed. The computational program for the design of a residual treatment train(DRTT) works in three steps which estimate the residual production to be treated, analyze the mass balance, and determine the size of each unit process. Of particular interest in the DRTT program, is provision for a filter backwash recycle system consisting of surge tank and sedimentation basin for more efficient recycling of backwash water. When the DRTT program was applied to the Chungju WTP, the program was very beneficial in avoiding errors which might have occurred during arithmetic calculations and in reducing the time needed to get the output. It is anticipated that the DRTT program could be used for design of new WTPs as well as the rehabilitation of existing ones.

Speaker-Adaptive Speech Synthesis based on Fuzzy Vector Quantizer Mapping and Neural Networks (퍼지 벡터 양자화기 사상화와 신경망에 의한 화자적응 음성합성)

  • Lee, Jin-Yi;Lee, Gwang-Hyeong
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.1
    • /
    • pp.149-160
    • /
    • 1997
  • This paper is concerned with the problem of speaker-adaptive speech synthes is method using a mapped codebook designed by fuzzy mapping on FLVQ (Fuzzy Learning Vector Quantization). The FLVQ is used to design both input and reference speaker's codebook. This algorithm is incorporated fuzzy membership function into the LVQ(learning vector quantization) networks. Unlike the LVQ algorithm, this algorithm minimizes the network output errors which are the differences of clas s membership target and actual membership values, and results to minimize the distances between training patterns and competing neurons. Speaker Adaptation in speech synthesis is performed as follow;input speaker's codebook is mapped a reference speaker's codebook in fuzzy concepts. The Fuzzy VQ mapping replaces a codevector preserving its fuzzy membership function. The codevector correspondence histogram is obtained by accumulating the vector correspondence along the DTW optimal path. We use the Fuzzy VQ mapping to design a mapped codebook. The mapped codebook is defined as a linear combination of reference speaker's vectors using each fuzzy histogram as a weighting function with membership values. In adaptive-speech synthesis stage, input speech is fuzzy vector-quantized by the mapped codcbook, and then FCM arithmetic is used to synthesize speech adapted to input speaker. The speaker adaption experiments are carried out using speech of males in their thirties as input speaker's speech, and a female in her twenties as reference speaker's speech. Speeches used in experiments are sentences /anyoung hasim nika/ and /good morning/. As a results of experiments, we obtained a synthesized speech adapted to input speaker.

  • PDF

Ensemble Learning with Support Vector Machines for Bond Rating (회사채 신용등급 예측을 위한 SVM 앙상블학습)

  • Kim, Myoung-Jong
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.2
    • /
    • pp.29-45
    • /
    • 2012
  • Bond rating is regarded as an important event for measuring financial risk of companies and for determining the investment returns of investors. As a result, it has been a popular research topic for researchers to predict companies' credit ratings by applying statistical and machine learning techniques. The statistical techniques, including multiple regression, multiple discriminant analysis (MDA), logistic models (LOGIT), and probit analysis, have been traditionally used in bond rating. However, one major drawback is that it should be based on strict assumptions. Such strict assumptions include linearity, normality, independence among predictor variables and pre-existing functional forms relating the criterion variablesand the predictor variables. Those strict assumptions of traditional statistics have limited their application to the real world. Machine learning techniques also used in bond rating prediction models include decision trees (DT), neural networks (NN), and Support Vector Machine (SVM). Especially, SVM is recognized as a new and promising classification and regression analysis method. SVM learns a separating hyperplane that can maximize the margin between two categories. SVM is simple enough to be analyzed mathematical, and leads to high performance in practical applications. SVM implements the structuralrisk minimization principle and searches to minimize an upper bound of the generalization error. In addition, the solution of SVM may be a global optimum and thus, overfitting is unlikely to occur with SVM. In addition, SVM does not require too many data sample for training since it builds prediction models by only using some representative sample near the boundaries called support vectors. A number of experimental researches have indicated that SVM has been successfully applied in a variety of pattern recognition fields. However, there are three major drawbacks that can be potential causes for degrading SVM's performance. First, SVM is originally proposed for solving binary-class classification problems. Methods for combining SVMs for multi-class classification such as One-Against-One, One-Against-All have been proposed, but they do not improve the performance in multi-class classification problem as much as SVM for binary-class classification. Second, approximation algorithms (e.g. decomposition methods, sequential minimal optimization algorithm) could be used for effective multi-class computation to reduce computation time, but it could deteriorate classification performance. Third, the difficulty in multi-class prediction problems is in data imbalance problem that can occur when the number of instances in one class greatly outnumbers the number of instances in the other class. Such data sets often cause a default classifier to be built due to skewed boundary and thus the reduction in the classification accuracy of such a classifier. SVM ensemble learning is one of machine learning methods to cope with the above drawbacks. Ensemble learning is a method for improving the performance of classification and prediction algorithms. AdaBoost is one of the widely used ensemble learning techniques. It constructs a composite classifier by sequentially training classifiers while increasing weight on the misclassified observations through iterations. The observations that are incorrectly predicted by previous classifiers are chosen more often than examples that are correctly predicted. Thus Boosting attempts to produce new classifiers that are better able to predict examples for which the current ensemble's performance is poor. In this way, it can reinforce the training of the misclassified observations of the minority class. This paper proposes a multiclass Geometric Mean-based Boosting (MGM-Boost) to resolve multiclass prediction problem. Since MGM-Boost introduces the notion of geometric mean into AdaBoost, it can perform learning process considering the geometric mean-based accuracy and errors of multiclass. This study applies MGM-Boost to the real-world bond rating case for Korean companies to examine the feasibility of MGM-Boost. 10-fold cross validations for threetimes with different random seeds are performed in order to ensure that the comparison among three different classifiers does not happen by chance. For each of 10-fold cross validation, the entire data set is first partitioned into tenequal-sized sets, and then each set is in turn used as the test set while the classifier trains on the other nine sets. That is, cross-validated folds have been tested independently of each algorithm. Through these steps, we have obtained the results for classifiers on each of the 30 experiments. In the comparison of arithmetic mean-based prediction accuracy between individual classifiers, MGM-Boost (52.95%) shows higher prediction accuracy than both AdaBoost (51.69%) and SVM (49.47%). MGM-Boost (28.12%) also shows the higher prediction accuracy than AdaBoost (24.65%) and SVM (15.42%)in terms of geometric mean-based prediction accuracy. T-test is used to examine whether the performance of each classifiers for 30 folds is significantly different. The results indicate that performance of MGM-Boost is significantly different from AdaBoost and SVM classifiers at 1% level. These results mean that MGM-Boost can provide robust and stable solutions to multi-classproblems such as bond rating.