• Title/Summary/Keyword: bayesian predictive model

Search Result 77, Processing Time 0.027 seconds

Rapid seismic vulnerability assessment by new regression-based demand and collapse models for steel moment frames

  • Kia, M.;Banazadeh, M.;Bayat, M.
    • Earthquakes and Structures
    • /
    • v.14 no.3
    • /
    • pp.203-214
    • /
    • 2018
  • Predictive demand and collapse fragility functions are two essential components of the probabilistic seismic demand analysis that are commonly developed based on statistics with enormous, costly and time consuming data gathering. Although this approach might be justified for research purposes, it is not appealing for practical applications because of its computational cost. Thus, in this paper, Bayesian regression-based demand and collapse models are proposed to eliminate the need of time-consuming analyses. The demand model developed in the form of linear equation predicts overall maximum inter-story drift of the lowto mid-rise regular steel moment resisting frames (SMRFs), while the collapse model mathematically expressed by lognormal cumulative distribution function provides collapse occurrence probability for a given spectral acceleration at the fundamental period of the structure. Next, as an application, the proposed demand and collapse functions are implemented in a seismic fragility analysis to develop fragility and consequently seismic demand curves of three example buildings. The accuracy provided by utilization of the proposed models, with considering computation reduction, are compared with those directly obtained from Incremental Dynamic analysis, which is a computer-intensive procedure.

Bayesian Computation for Superposition of MUSA-OKUMOTO and ERLANG(2) processes (MUSA-OKUMOTO와 ERLANG(2)의 중첩과정에 대한 베이지안 계산 연구)

  • 최기헌;김희철
    • The Korean Journal of Applied Statistics
    • /
    • v.11 no.2
    • /
    • pp.377-387
    • /
    • 1998
  • A Markov Chain Monte Carlo method with data augmentation is developed to compute the features of the posterior distribution. For each observed failure epoch, we introduced latent variables that indicates with component of the Superposition model. This data augmentation approach facilitates specification of the transitional measure in the Markov Chain. Metropolis algorithms along with Gibbs steps are proposed to preform the Bayesian inference of such models. for model determination, we explored the Pre-quential conditional predictive Ordinate(PCPO) criterion that selects the best model with the largest posterior likelihood among models using all possible subsets of the component intensity functions. To relax the monotonic intensity function assumptions, we consider in this paper Superposition of Musa-Okumoto and Erlang(2) models. A numerical example with simulated dataset is given.

  • PDF

Bayesian bi-level variable selection for genome-wide survival study

  • Eunjee Lee;Joseph G. Ibrahim;Hongtu Zhu
    • Genomics & Informatics
    • /
    • v.21 no.3
    • /
    • pp.28.1-28.13
    • /
    • 2023
  • Mild cognitive impairment (MCI) is a clinical syndrome characterized by the onset and evolution of cognitive impairments, often considered a transitional stage to Alzheimer's disease (AD). The genetic traits of MCI patients who experience a rapid progression to AD can enhance early diagnosis capabilities and facilitate drug discovery for AD. While a genome-wide association study (GWAS) is a standard tool for identifying single nucleotide polymorphisms (SNPs) related to a disease, it fails to detect SNPs with small effect sizes due to stringent control for multiple testing. Additionally, the method does not consider the group structures of SNPs, such as genes or linkage disequilibrium blocks, which can provide valuable insights into the genetic architecture. To address the limitations, we propose a Bayesian bi-level variable selection method that detects SNPs associated with time of conversion from MCI to AD. Our approach integrates group inclusion indicators into an accelerated failure time model to identify important SNP groups. Additionally, we employ data augmentation techniques to impute censored time values using a predictive posterior. We adapt Dirichlet-Laplace shrinkage priors to incorporate the group structure for SNP-level variable selection. In the simulation study, our method outperformed other competing methods regarding variable selection. The analysis of Alzheimer's Disease Neuroimaging Initiative (ADNI) data revealed several genes directly or indirectly related to AD, whereas a classical GWAS did not identify any significant SNPs.

Multiple imputation and synthetic data (다중대체와 재현자료 작성)

  • Kim, Joungyoun;Park, Min-Jeong
    • The Korean Journal of Applied Statistics
    • /
    • v.32 no.1
    • /
    • pp.83-97
    • /
    • 2019
  • As society develops, the dissemination of microdata has increased to respond to diverse analytical needs of users. Analysis of microdata for policy making, academic purposes, etc. is highly desirable in terms of value creation. However, the provision of microdata, whose usefulness is guaranteed, has a risk of exposure of personal information. Several methods have been considered to ensure the protection of personal information while ensuring the usefulness of the data. One of these methods has been studied to generate and utilize synthetic data. This paper aims to understand the synthetic data by exploring methodologies and precautions related to synthetic data. To this end, we first explain muptiple imputation, Bayesian predictive model, and Bayesian bootstrap, which are basic foundations for synthetic data. And then, we link these concepts to the construction of fully/partially synthetic data. To understand the creation of synthetic data, we review a real longitudinal synthetic data example which is based on sequential regression multivariate imputation.

A Study on the War Simulation and Prediction Using Bayesian Inference (베이지안 추론을 이용한 전쟁 시뮬레이션과 예측 연구)

  • Lee, Seung-Lyong;Yoo, Byung Joo;Youn, Sangyoun;Bang, Sang-Ho;Jung, Jae-Woong
    • The Journal of the Korea Contents Association
    • /
    • v.21 no.11
    • /
    • pp.77-86
    • /
    • 2021
  • A method of constructing a war simulation based on Bayesian Inference was proposed as a method of constructing heterogeneous historical war data obtained with a time difference into a single model. A method of applying a linear regression model can be considered as a method of predicting future battles by analyzing historical war results. However it is not appropriate for two heterogeneous types of historical data that reflect changes in the battlefield environment due to different times to be suitable as a single linear regression model and violation of the model's assumptions. To resolve these problems a Bayesian inference method was proposed to obtain a post-distribution by assuming the data from the previous era as a non-informative prior distribution and to infer the final posterior distribution by using it as a prior distribution to analyze the data obtained from the next era. Another advantage of the Bayesian inference method is that the results sampled by the Markov Chain Monte Carlo method can be used to infer posterior distribution or posterior predictive distribution reflecting uncertainty. In this way, it has the advantage of not only being able to utilize a variety of information rather than analyzing it with a classical linear regression model, but also continuing to update the model by reflecting additional data obtained in the future.

A BAYESIAN APPROACH FOR A DECOMPOSITION MODEL OF SOFTWARE RELIABILITY GROWTH USING A RECORD VALUE STATISTICS

  • Choi, Ki-Heon;Kim, Hee-Cheul
    • Journal of applied mathematics & informatics
    • /
    • v.8 no.1
    • /
    • pp.243-252
    • /
    • 2001
  • The points of failure of a decomposition process are defined to be the union of the points of failure from two component point processes for software reliability systems. Because sampling from the likelihood function of the decomposition model is difficulty, Gibbs Sampler can be applied in a straightforward manner. A Markov Chain Monte Carlo method with data augmentation is developed to compute the features of the posterior distribution. For model determination, we explored the prequential conditional predictive ordinate criterion that selects the best model with the largest posterior likelihood among models using all possible subsets of the component intensity functions. A numerical example with a simulated data set is given.

Forecasting Government Bond Yields in Thailand: A Bayesian VAR Approach

  • BUABAN, Wantana;SETHAPRAMOTE, Yuthana
    • The Journal of Asian Finance, Economics and Business
    • /
    • v.9 no.3
    • /
    • pp.181-193
    • /
    • 2022
  • This paper seeks to investigate major macroeconomic factors and bond yield interactions in Thai bond markets, with the goal of forecasting future bond yields. This study examines the best predictive yields for future bond yields at different maturities of 1-, 3-, 5-, 7-, and 10-years using time series data of economic indicators covering the period from 1998 to 2020. The empirical findings support the hypothesis that macroeconomic factors influence bond yield fluctuations. In terms of forecasting future bond yields, static predictions reveal that in most cases, the BVAR model offers the best predictivity of bond rates at various maturities. Furthermore, the BVAR model has the best performance in dynamic rolling-window, forecasting bond yields with various maturities for 2-, 4-, and 8-quarters. The findings of this study imply that the BVAR model forecasts future yields more accurately and consistently than other competitive models. Our research could help policymakers and investors predict bond yield changes, which could be important in macroeconomic policy development.

Bayesian Survival Analysis of High-Dimensional Microarray Data for Mantle Cell Lymphoma Patients

  • Moslemi, Azam;Mahjub, Hossein;Saidijam, Massoud;Poorolajal, Jalal;Soltanian, Ali Reza
    • Asian Pacific Journal of Cancer Prevention
    • /
    • v.17 no.1
    • /
    • pp.95-100
    • /
    • 2016
  • Background: Survival time of lymphoma patients can be estimated with the help of microarray technology. In this study, with the use of iterative Bayesian Model Averaging (BMA) method, survival time of Mantle Cell Lymphoma patients (MCL) was estimated and in reference to the findings, patients were divided into two high-risk and low-risk groups. Materials and Methods: In this study, gene expression data of MCL patients were used in order to select a subset of genes for survival analysis with microarray data, using the iterative BMA method. To evaluate the performance of the method, patients were divided into high-risk and low-risk based on their scores. Performance prediction was investigated using the log-rank test. The bioconductor package "iterativeBMAsurv" was applied with R statistical software for classification and survival analysis. Results: In this study, 25 genes associated with survival for MCL patients were identified across 132 selected models. The maximum likelihood estimate coefficients of the selected genes and the posterior probabilities of the selected models were obtained from training data. Using this method, patients could be separated into high-risk and low-risk groups with high significance (p<0.001). Conclusions: The iterative BMA algorithm has high precision and ability for survival analysis. This method is capable of identifying a few predictive variables associated with survival, among many variables in a set of microarray data. Therefore, it can be used as a low-cost diagnostic tool in clinical research.

Detecting the Influential Observation Using Intrinsic Bayes Factors

  • Chung, Younshik
    • Journal of the Korean Statistical Society
    • /
    • v.29 no.1
    • /
    • pp.81-94
    • /
    • 2000
  • For the balanced variance component model, sometimes intraclass correlation coefficient is of interest. If there is little information about the parameter, then the reference prior(Berger and Bernardo, 1992) is widely used. Pettit nd Young(1990) considered a measrue of the effect of a single observation on a logarithmic Bayes factor. However, under such a reference prior, the Bayes factor depends on the ratio of unspecified constants. In order to discard this problem, influence diagnostic measures using the intrinsic Bayes factor(Berger and Pericchi, 1996) is presented. Finally, one simulated dataset is provided which illustrates the methodology with appropriate simulation based computational formulas. In order to overcome the difficult Bayesian computation, MCMC methods, such as Gibbs sampler(Gelfand and Smith, 1990) and Metropolis algorithm, are empolyed.

  • PDF

Methodology for Variable Optimization in Injection Molding Process (사출 성형 공정에서의 변수 최적화 방법론)

  • Jung, Young Jin;Kang, Tae Ho;Park, Jeong In;Cho, Joong Yeon;Hong, Ji Soo;Kang, Sung Woo
    • Journal of Korean Society for Quality Management
    • /
    • v.52 no.1
    • /
    • pp.43-56
    • /
    • 2024
  • Purpose: The injection molding process, crucial for plastic shaping, encounters difficulties in sustaining product quality when replacing injection machines. Variations in machine types and outputs between different production lines or factories increase the risk of quality deterioration. In response, the study aims to develop a system that optimally adjusts conditions during the replacement of injection machines linked to molds. Methods: Utilizing a dataset of 12 injection process variables and 52 corresponding sensor variables, a predictive model is crafted using Decision Tree, Random Forest, and XGBoost. Model evaluation is conducted using an 80% training data and a 20% test data split. The dependent variable, classified into five characteristics based on temperature and pressure, guides the prediction model. Bayesian optimization, integrated into the selected model, determines optimal values for process variables during the replacement of injection machines. The iterative convergence of sensor prediction values to the optimum range is visually confirmed, aligning them with the target range. Experimental results validate the proposed approach. Results: Post-experiment analysis indicates the superiority of the XGBoost model across all five characteristics, achieving a combined high performance of 0.81 and a Mean Absolute Error (MAE) of 0.77. The study introduces a method for optimizing initial conditions in the injection process during machine replacement, utilizing Bayesian optimization. This streamlined approach reduces both time and costs, thereby enhancing process efficiency. Conclusion: This research contributes practical insights to the optimization literature, offering valuable guidance for industries seeking streamlined and cost-effective methods for machine replacement in injection molding.