• Title/Summary/Keyword: Bayesian Testing

Search Result 148, Processing Time 0.024 seconds

Estimation of genetic parameters and trends for production traits of dairy cattle in Thailand using a multiple-trait multiple-lactation test day model

  • Buaban, Sayan;Puangdee, Somsook;Duangjinda, Monchai;Boonkum, Wuttigrai
    • Asian-Australasian Journal of Animal Sciences
    • /
    • v.33 no.9
    • /
    • pp.1387-1399
    • /
    • 2020
  • Objective: The objective of this study was to estimate the genetic parameters and trends for milk, fat, and protein yields in the first three lactations of Thai dairy cattle using a 3-trait,-3-lactation random regression test-day model. Methods: Data included 168,996, 63,388, and 27,145 test-day records from the first, second, and third lactations, respectively. Records were from 19,068 cows calving from 1993 to 2013 in 124 herds. (Co) variance components were estimated by Bayesian methods. Gibbs sampling was used to obtain posterior distributions. The model included herd-year-month of testing, breed group-season of calving-month in tested milk group, linear and quadratic age at calving as fixed effects, and random regression coefficients for additive genetic and permanent environmental effects, which were defined as modified constant, linear, quadratic, cubic and quartic Legendre coefficients. Results: Average daily heritabilities ranged from 0.36 to 0.48 for milk, 0.33 to 0.44 for fat and 0.37 to 0.48 for protein yields; they were higher in the third lactation for all traits. Heritabilities of test-day milk and protein yields for selected days in milk were higher in the middle than at the beginning or end of lactation, whereas those for test-day fat yields were high at the beginning and end of lactation. Genetics correlations (305-d yield) among production yields within lactations (0.44 to 0.69) were higher than those across lactations (0.36 to 0.68). The largest genetic correlation was observed between the first and second lactation. The genetic trends of 305-d milk, fat and protein yields were 230 to 250, 25 to 29, and 30 to 35 kg per year, respectively. Conclusion: A random regression model seems to be a flexible and reliable procedure for the genetic evaluation of production yields. It can be used to perform breeding value estimation for national genetic evaluation in the Thai dairy cattle population.

Probabilistic Calibration of Computer Model and Application to Reliability Analysis of Elasto-Plastic Insertion Problem (컴퓨터모델의 확률적 보정 및 탄소성 압착문제의 신뢰도분석 응용)

  • Yoo, Min Young;Choi, Joo Ho
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.37 no.9
    • /
    • pp.1133-1140
    • /
    • 2013
  • A computer model is a useful tool that provides solution via physical modeling instead of expensive testing. In reality, however, it often does not agree with the experimental data owing to simplifying assumption and unknown or uncertain input parameters. In this study, a Bayesian approach is proposed to calibrate the computer model in a probabilistic manner using the measured data. The elasto-plastic analysis of a pyrotechnically actuated device (PAD) is employed to demonstrate this approach, which is a component that delivers high power in remote environments by the combustion of a self-contained energy source. A simple mathematical model that quickly evaluates the performance is developed. Unknown input parameters are calibrated conditional on the experimental data using the Markov Chain Monte Carlo algorithm, which is a modern computational statistics method. Finally, the results are applied to determine the reliability of the PAD.

Junk-Mail Filtering by Mail Address Validation and Title-Content Weighting (메일 주소 유효성과 제목-내용 가중치 기법에 의한 스팸 메일 필터링)

  • Kang Seung-Shik
    • Journal of Korea Multimedia Society
    • /
    • v.9 no.2
    • /
    • pp.255-263
    • /
    • 2006
  • It is common that a junk mail has an inconsistency of mail addresses between those of the mail headers and the mail recipients. In addition, users easily know that an email is a junk or legitimate mail only by looking for the title of the email. In this paper, we tried to apply the filtering classifiers of mail address validation check and the combination method of title-content weighting to improve the performance of junk mail filtering system. In order to verify the effectiveness of the proposed method, we performed an experiment by applying them to Naive Bayesian classifier. The experiment includes the unit testing and the combination of the filtering techniques. As a result, we found that our method improved 11.6% of recall and 2.1% of precision that it contributed the enhancement of the junk mail filtering system.

  • PDF

An N-version Learning Approach to Enhance the Prediction Accuracy of Classification Systems in Genetics-based Learning Environments (유전학 기반 학습 환경하에서 분류 시스템의 성능 향상을 위한 엔-버전 학습법)

  • Kim, Yeong-Jun;Hong, Cheol-Ui
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.7
    • /
    • pp.1841-1848
    • /
    • 1999
  • DELVAUX is a genetics-based inductive learning system that learns a rule-set, which consists of Bayesian classification rules, from sets of examples for classification tasks. One problem that DELVAUX faces in the rule-set learning process is that, occasionally, the learning process ends with a local optimum without finding the best rule-set. Another problem is that, occasionally, the learning process ends with a rule-set that performs well for the training examples but not for the unknown testing examples. This paper describes efforts to alleviate these two problems centering on the N-version learning approach, in which multiple rule-sets are learning and a classification system is constructed with those learned rule-sets to improve the overall performance of a classification system. For the implementation of the N-version learning approach, we propose a decision-making scheme that can draw a decision using multiple rule-sets and a genetic algorithm approach to find a good combination of rule-sets from a set of learned rule-sets. We also present empirical results that evaluate the effect of the N-version learning approach in the DELVAUX learning environment.

  • PDF

An R package UnifiedDoseFinding for continuous and ordinal outcomes in Phase I dose-finding trials

  • Pan, Haitao;Mu, Rongji;Hsu, Chia-Wei;Zhou, Shouhao
    • Communications for Statistical Applications and Methods
    • /
    • v.29 no.4
    • /
    • pp.421-439
    • /
    • 2022
  • Phase I dose-finding trials are essential in drug development. By finding the maximum tolerated dose (MTD) of a new drug or treatment, a Phase I trial establishes the recommended doses for later-phase testing. The primary toxicity endpoint of interest is often a binary variable, which describes an event of a patient who experiences dose-limiting toxicity. However, there is a growing interest in dose-finding studies regarding non-binary outcomes, defined by either the weighted sum of rates of various toxicity grades or a continuous outcome. Although several novel methods have been proposed in the literature, accessible software is still lacking to implement these methods. This study introduces a newly developed R package, UnifiedDoseFinding, which implements three phase I dose-finding methods with non-binary outcomes (Quasi- and Robust Quasi-CRM designs by Yuan et al. (2007) and Pan et al. (2014), gBOIN design by Mu et al. (2019), and by a method by Ivanova and Kim (2009)). For each of the methods, UnifiedDoseFinding provides corresponding functions that begin with next that determines the dose for the next cohort of patients, select, which selects the MTD defined by the non-binary toxicity endpoint when the trial is completed, and get oc, which obtains the operating characteristics. Three real examples are provided to help practitioners use these methods. The R package UnifiedDoseFinding, which is accessible in R CRAN, provides a user-friendly tool to facilitate the implementation of innovative dose-finding studies with nonbinary outcomes.

Prediction of skewness and kurtosis of pressure coefficients on a low-rise building by deep learning

  • Youqin Huang;Guanheng Ou;Jiyang Fu;Huifan Wu
    • Wind and Structures
    • /
    • v.36 no.6
    • /
    • pp.393-404
    • /
    • 2023
  • Skewness and kurtosis are important higher-order statistics for simulating non-Gaussian wind pressure series on low-rise buildings, but their predictions are less studied in comparison with those of the low order statistics as mean and rms. The distribution gradients of skewness and kurtosis on roofs are evidently higher than those of mean and rms, which increases their prediction difficulty. The conventional artificial neural networks (ANNs) used for predicting mean and rms show unsatisfactory accuracy in predicting skewness and kurtosis owing to the limited capacity of shallow learning of ANNs. In this work, the deep neural networks (DNNs) model with the ability of deep learning is introduced to predict the skewness and kurtosis on a low-rise building. For obtaining the optimal generalization of the DNNs model, the hyper parameters are automatically determined by Bayesian Optimization (BO). Moreover, for providing a benchmark for future studies on predicting higher order statistics, the data sets for training and testing the DNNs model are extracted from the internationally open NIST-UWO database, and the prediction errors of all taps are comprehensively quantified by various error metrices. The results show that the prediction accuracy in this study is apparently better than that in the literature, since the correlation coefficient between the predicted and experimental results is 0.99 and 0.75 in this paper and the literature respectively. In the untrained cornering wind direction, the distributions of skewness and kurtosis are well captured by DNNs on the whole building including the roof corner with strong non-normality, and the correlation coefficients between the predicted and experimental results are 0.99 and 0.95 for skewness and kurtosis respectively.

Predictive Clustering-based Collaborative Filtering Technique for Performance-Stability of Recommendation System (추천 시스템의 성능 안정성을 위한 예측적 군집화 기반 협업 필터링 기법)

  • Lee, O-Joun;You, Eun-Soon
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.1
    • /
    • pp.119-142
    • /
    • 2015
  • With the explosive growth in the volume of information, Internet users are experiencing considerable difficulties in obtaining necessary information online. Against this backdrop, ever-greater importance is being placed on a recommender system that provides information catered to user preferences and tastes in an attempt to address issues associated with information overload. To this end, a number of techniques have been proposed, including content-based filtering (CBF), demographic filtering (DF) and collaborative filtering (CF). Among them, CBF and DF require external information and thus cannot be applied to a variety of domains. CF, on the other hand, is widely used since it is relatively free from the domain constraint. The CF technique is broadly classified into memory-based CF, model-based CF and hybrid CF. Model-based CF addresses the drawbacks of CF by considering the Bayesian model, clustering model or dependency network model. This filtering technique not only improves the sparsity and scalability issues but also boosts predictive performance. However, it involves expensive model-building and results in a tradeoff between performance and scalability. Such tradeoff is attributed to reduced coverage, which is a type of sparsity issues. In addition, expensive model-building may lead to performance instability since changes in the domain environment cannot be immediately incorporated into the model due to high costs involved. Cumulative changes in the domain environment that have failed to be reflected eventually undermine system performance. This study incorporates the Markov model of transition probabilities and the concept of fuzzy clustering with CBCF to propose predictive clustering-based CF (PCCF) that solves the issues of reduced coverage and of unstable performance. The method improves performance instability by tracking the changes in user preferences and bridging the gap between the static model and dynamic users. Furthermore, the issue of reduced coverage also improves by expanding the coverage based on transition probabilities and clustering probabilities. The proposed method consists of four processes. First, user preferences are normalized in preference clustering. Second, changes in user preferences are detected from review score entries during preference transition detection. Third, user propensities are normalized using patterns of changes (propensities) in user preferences in propensity clustering. Lastly, the preference prediction model is developed to predict user preferences for items during preference prediction. The proposed method has been validated by testing the robustness of performance instability and scalability-performance tradeoff. The initial test compared and analyzed the performance of individual recommender systems each enabled by IBCF, CBCF, ICFEC and PCCF under an environment where data sparsity had been minimized. The following test adjusted the optimal number of clusters in CBCF, ICFEC and PCCF for a comparative analysis of subsequent changes in the system performance. The test results revealed that the suggested method produced insignificant improvement in performance in comparison with the existing techniques. In addition, it failed to achieve significant improvement in the standard deviation that indicates the degree of data fluctuation. Notwithstanding, it resulted in marked improvement over the existing techniques in terms of range that indicates the level of performance fluctuation. The level of performance fluctuation before and after the model generation improved by 51.31% in the initial test. Then in the following test, there has been 36.05% improvement in the level of performance fluctuation driven by the changes in the number of clusters. This signifies that the proposed method, despite the slight performance improvement, clearly offers better performance stability compared to the existing techniques. Further research on this study will be directed toward enhancing the recommendation performance that failed to demonstrate significant improvement over the existing techniques. The future research will consider the introduction of a high-dimensional parameter-free clustering algorithm or deep learning-based model in order to improve performance in recommendations.

Robo-Advisor Algorithm with Intelligent View Model (지능형 전망모형을 결합한 로보어드바이저 알고리즘)

  • Kim, Sunwoong
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.39-55
    • /
    • 2019
  • Recently banks and large financial institutions have introduced lots of Robo-Advisor products. Robo-Advisor is a Robot to produce the optimal asset allocation portfolio for investors by using the financial engineering algorithms without any human intervention. Since the first introduction in Wall Street in 2008, the market size has grown to 60 billion dollars and is expected to expand to 2,000 billion dollars by 2020. Since Robo-Advisor algorithms suggest asset allocation output to investors, mathematical or statistical asset allocation strategies are applied. Mean variance optimization model developed by Markowitz is the typical asset allocation model. The model is a simple but quite intuitive portfolio strategy. For example, assets are allocated in order to minimize the risk on the portfolio while maximizing the expected return on the portfolio using optimization techniques. Despite its theoretical background, both academics and practitioners find that the standard mean variance optimization portfolio is very sensitive to the expected returns calculated by past price data. Corner solutions are often found to be allocated only to a few assets. The Black-Litterman Optimization model overcomes these problems by choosing a neutral Capital Asset Pricing Model equilibrium point. Implied equilibrium returns of each asset are derived from equilibrium market portfolio through reverse optimization. The Black-Litterman model uses a Bayesian approach to combine the subjective views on the price forecast of one or more assets with implied equilibrium returns, resulting a new estimates of risk and expected returns. These new estimates can produce optimal portfolio by the well-known Markowitz mean-variance optimization algorithm. If the investor does not have any views on his asset classes, the Black-Litterman optimization model produce the same portfolio as the market portfolio. What if the subjective views are incorrect? A survey on reports of stocks performance recommended by securities analysts show very poor results. Therefore the incorrect views combined with implied equilibrium returns may produce very poor portfolio output to the Black-Litterman model users. This paper suggests an objective investor views model based on Support Vector Machines(SVM), which have showed good performance results in stock price forecasting. SVM is a discriminative classifier defined by a separating hyper plane. The linear, radial basis and polynomial kernel functions are used to learn the hyper planes. Input variables for the SVM are returns, standard deviations, Stochastics %K and price parity degree for each asset class. SVM output returns expected stock price movements and their probabilities, which are used as input variables in the intelligent views model. The stock price movements are categorized by three phases; down, neutral and up. The expected stock returns make P matrix and their probability results are used in Q matrix. Implied equilibrium returns vector is combined with the intelligent views matrix, resulting the Black-Litterman optimal portfolio. For comparisons, Markowitz mean-variance optimization model and risk parity model are used. The value weighted market portfolio and equal weighted market portfolio are used as benchmark indexes. We collect the 8 KOSPI 200 sector indexes from January 2008 to December 2018 including 132 monthly index values. Training period is from 2008 to 2015 and testing period is from 2016 to 2018. Our suggested intelligent view model combined with implied equilibrium returns produced the optimal Black-Litterman portfolio. The out of sample period portfolio showed better performance compared with the well-known Markowitz mean-variance optimization portfolio, risk parity portfolio and market portfolio. The total return from 3 year-period Black-Litterman portfolio records 6.4%, which is the highest value. The maximum draw down is -20.8%, which is also the lowest value. Sharpe Ratio shows the highest value, 0.17. It measures the return to risk ratio. Overall, our suggested view model shows the possibility of replacing subjective analysts's views with objective view model for practitioners to apply the Robo-Advisor asset allocation algorithms in the real trading fields.