• Title/Summary/Keyword: 표본 포함확률

Search Result 57, Processing Time 0.026 seconds

Markov Chain Monte Carlo Simulation to Estimate Material Properties of a Layered Half-space (층상 반무한 지반의 물성치 추정을 위한 마르코프 연쇄 몬테카를로 모사 기법)

  • Jin Ho Lee;Hieu Van Nguyen;Se Hyeok Lee
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.36 no.3
    • /
    • pp.203-211
    • /
    • 2023
  • A Markov chain Monte Carlo (MCMC) simulation is proposed for probabilistic full waveform inversion (FWI) in a layered half-space. Dynamic responses on the half-space surface are estimated using the thin-layer method when a harmonic vertical force is applied. Subsequently, a posterior probability distribution function and the corresponding objective function are formulated to minimize the difference between estimations and observed data as well as that of model parameters from prior information. Based on the gradient of the objective function, a proposal distribution and an acceptance probability for MCMC samples are proposed. The proposed MCMC simulation is applied to several layered half-space examples. It is demonstrated that the proposed MCMC simulation for probabilistic FWI can estimate probabilistic material properties such as the shear-wave velocities of a layered half-space.

Derivation of Plotting Position Formulas Considering the Coefficients of Skewness for the GEV Distribution (왜곡도 계수를 고려한 GEV 분포의 도시위치공식 유도)

  • Kim, Soo-Young;Heo, Jun-Haeng;Choi, Min-Young
    • Journal of Korea Water Resources Association
    • /
    • v.44 no.2
    • /
    • pp.85-96
    • /
    • 2011
  • Probability plotting position is generally used for the graphical analysis of the annual maximum quantile and the estimation of exceedance probability to display the fitness between sample and an appropriate probability distribution. In addition, it is used to apply a specific goodness of fit test. Plotting position formula to define the probability plotting position has been studied in many researches. Especially, the GEV distribution which is an important probability distribution to analyze the frequency of hydrologic data was popular. In this study, the theoretical reduced variates are derived using the mean value of order statistics to derived an appropriate plotting position formula for the GEV distribution. In addition, various forms of plotting position formula considering various sample sizes and coefficients of skewness related with shape parameters are applied. The parameters of plotting position formulas are estimated using the genetic algorithm. The accuracy of derived plotting position formula is estimated by the errors between the theoretical reduced variates and those by various plotting position formulas including the derived ones in this study. As a result, the errors by derived plotting position formula is the smallest at the range of shape parameter with -0.25~0.10.

Exact simulataneous confidence interval for the case of four means using TK procedure (Tukey-Kramer방법을 이용한 4개 평균에 관한 정확한 동시 신뢰구간의 통계적 계산 방법)

  • 김병천;김화선;조신섭
    • The Korean Journal of Applied Statistics
    • /
    • v.2 no.1
    • /
    • pp.18-34
    • /
    • 1989
  • The problem of simultaneously estimating the pairwise differences of means of four independent normal populations with equal variances is considered. A statistical computing procedure involving a trivariate t density constructs the exact confidence intervals with simultaneous co verage probability equal to $1-\alpha$. For equal sample sizes, the new procedure is the same as the Tukey studentized range procedure. With unequal sample sizes, in the sense of efficiency for confidence interval lengths and experimentwise error rates, the procedure is superior to the various generalized Tukey procedures.

Theoretical Considerations for the Agresti-Coull Type Confidence Interval in Misclassified Binary Data (오분류된 이진자료에서 Agresti-Coull유형의 신뢰구간에 대한 이론적 고찰)

  • Lee, Seung-Chun
    • Communications for Statistical Applications and Methods
    • /
    • v.18 no.4
    • /
    • pp.445-455
    • /
    • 2011
  • Although misclassified binary data occur frequently in practice, the statistical methodology available for the data is rather limited. In particular, the interval estimation of population proportion has relied on the classical Wald method. Recently, Lee and Choi (2009) developed a new confidence interval by applying the Agresti-Coull's approach and showed the efficiency of their proposed confidence interval numerically, but a theoretical justification has not been explored yet. Therefore, a Bayesian model for the misclassified binary data is developed to consider the Agresti-Coull confidence interval from a theoretical point of view. It is shown that the Agresti-Coull confidence interval is essentially a Bayesian confidence interval.

A Comparison of the Interval Estimations for the Difference in Paired Areas under the ROC Curves (대응표본에서 AUC차이에 대한 신뢰구간 추정에 관한 고찰)

  • Kim, Hee-Young
    • Communications for Statistical Applications and Methods
    • /
    • v.17 no.2
    • /
    • pp.275-292
    • /
    • 2010
  • Receiver operating characteristic(ROC) curves can be used to assess the accuracy of tests measured on ordinal or continuous scales. The most commonly used measure for the overall diagnostic accuracy of diagnostic tests is the area under the ROC curve(AUC). When two ROC curves are constructed based on two tests performed on the same individuals, statistical analysis on differences between AUCs must take into account the correlated nature of the data. This article focuses on confidence interval estimation of the difference between paired AUCs. We compare nonparametric, maximum likelihood, bootstrap and generalized pivotal quantity methods, and conduct a monte carlo simulation to investigate the probability coverage and expected length of the four methods.

A study on collecting representative food samples for the 10th Korean standard foods composition table (국가표준식품성분 데이터베이스 대표시료 선정을 위한 표본설계)

  • Kim, Jinheum;Hwang, Hae-Won;Cho, Yu Jung;Park, Jinwoo
    • The Korean Journal of Applied Statistics
    • /
    • v.33 no.2
    • /
    • pp.215-228
    • /
    • 2020
  • Under Article 19, Paragraph 1 of the Food Industry Promotion Act, Rural Development Administration renews the Korean foods composition table every five years. Before the publication of the tenth revision of the Korean foods composition table in 2021, this paper suggests methods for collecting representative samples of 182 highly consumed foods in Korea. Food markets are categorized by their distribution channels, which are supermarkets and local markets. Eight samples are collected from each category by applying the National Food and Nutrient Analysis Program (NFNAP)'s stratified multi-stage sampling. The NFNAP was implemented in 1997 as a collaborative food composition research effort between the National Institute of Health (NIH) and the US Department of Agriculture (USDA) to secure reliable estimates for the nutrient content of food and beverages consumed by the US population. Selected supermarkets for selecting representative food samples are Emart Kayang, Homeplus Siheung, Lottemart Dongducheon, Emart Suwon, Lottemart Dunsan, Lottemart Yeosu, Emart Ulsan, and Hanaroclub Ulsan. Selected local markets also are Doksandongusijang in Geumcheon-gu and Pungnapsijang in Songpa-gu, Seoul, Ilsansijang in Ilsanseo-gu, Goyang, Unamsijang in Buk-gu, Gwangju, Beopdongsijang in Daedeok-gu, Daejeon, Bongnaesijang in Yeongdo-gu and Jwadongjaeraesijang in Haeundae-gu, Busan, and Jungangsijang in Jinhae-gu, Changwon.

UC Model with ARIMA Trend and Forecasting U.S. GDP (ARIMA 추세의 비관측요인 모형과 미국 GDP에 대한 예측력)

  • Lee, Young Soo
    • International Area Studies Review
    • /
    • v.21 no.4
    • /
    • pp.159-172
    • /
    • 2017
  • In a typical trend-cycle decomposition of GDP, the trend component is usually assumed to follow a random walk process. This paper considers an ARIMA trend and assesses the validity of the ARIMA trend model. I construct univariate and bivariate unobserved-components(UC) models, allowing the ARIMA trend. Estimation results using U.S. data are favorable to the ARIMA trend models. I, also, compare the forecasting performance of the UC models. Dynamic pseudo-out-of-sample forecasting exercises are implemented with recursive estimations. I find that the bivariate model outperforms the univariate model, the smoothed estimates of trend and cycle components deliver smaller forecasting errors compared to the filtered estimates, and, most importantly, allowing for the ARIMA trend can lead to statistically significant gains in forecast accuracy, providing support for the ARIMA trend model. It is worthy of notice that trend shocks play the main source of the output fluctuation if the ARIMA trend is allowed in the UC model.

Stochastic disaggregation of daily rainfall based on K-Nearest neighbor resampling method (K번째 최근접 표본 재추출 방법에 의한 일 강우량의 추계학적 분해에 대한 연구)

  • Park, HeeSeong;Chung, GunHui
    • Journal of Korea Water Resources Association
    • /
    • v.49 no.4
    • /
    • pp.283-291
    • /
    • 2016
  • As the infrastructures and populations are the condensed in the mega city, urban flood management becomes very important due to the severe loss of lives and properties. For the more accurate calculation of runoff from the urban catchment, hourly or even minute rainfall data have been utilized. However, the time steps of the measured or forecasted data under climate change scenarios are longer than hourly, which causes the difficulty on the application. In this study, daily rainfall data was disaggregated into hourly using the stochastic method. Based on the historical hourly precipitation data, Gram Schmidt orthonormalization process and K-Nearest Neighbor Resampling (KNNR) method were applied to disaggregate daily precipitation into hourly. This method was originally developed to disaggregate yearly runoff data into monthly. Precipitation data has smaller probability density than runoff data, therefore, rainfall patterns considering the previous and next days were proposed as 7 different types. Disaggregated rainfall was resampled from the only same rainfall patterns to improve applicability. The proposed method was applied rainfall data observed at Seoul weather station where has 52 years hourly rainfall data and the disaggregated hourly data were compared to the measured data. The proposed method might be applied to disaggregate the climate change scenarios.

An Improved Structural Reliability Analysis using Moving Least Squares Approximation (이동최소제곱근사법을 이용한 개선된 구조 신뢰성 해석)

  • Kang, Soo-Chang;Koh, Hyun-Moo
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.28 no.6A
    • /
    • pp.835-842
    • /
    • 2008
  • The response surface method (RSM) is widely adopted for the structural reliability analysis because of its numerical efficiency. However, the RSM is still time consuming for large-scale applications and sometimes shows large errors in the calculation of sensitivity of reliability index with respect to random variables. Therefore, this study proposes a new RSM in which moving least squares (MLS) approximation is applied. Least squares approximation generally used in the common RSM gives equal weight to the coefficients of the response surface function (RSF). On the other hand, The MLS approximation gives higher weight to the experimental points closer to the design point, which yields the RSF more similar to the limit state at the design point. In the procedure of the proposed method, a linear RSF is constructed initially and then a quadratic RSF is formed using the axial experimental points selected from the reduced region where the design point is likely to exist. The RSF is updated successively by adding one more experimental point to the previously sampled experimental points. In order to demonstrate the effectiveness of the proposed method, mathematical problems and ten-bar truss are considered as numerical examples. As a result, the proposed method shows better accuracy and computational efficiency than the common RSM.

Bootstrap estimation of the standard error of treatment effect with double propensity score adjustment (이중 성향점수 보정 방법을 이용한 처리효과 추정치의 표준오차 추정: 붓스트랩의 적용)

  • Lim, So Jung;Jung, Inkyung
    • The Korean Journal of Applied Statistics
    • /
    • v.30 no.3
    • /
    • pp.453-462
    • /
    • 2017
  • Double propensity score adjustment is an analytic solution to address bias due to incomplete matching. However, it is difficult to estimate the standard error of the estimated treatment effect when using double propensity score adjustment. In this study, we propose two bootstrap methods to estimate the standard error. The first is a simple bootstrap method that involves drawing bootstrap samples from the matched sample using the propensity score as well as estimating the standard error from the bootstrapped samples. The second is a complex bootstrap method that draws bootstrap samples first from the original sample and then applies the propensity score matching to each bootstrapped sample. We examined the performances of the two methods using simulations under various scenarios. The estimates of standard error using the complex bootstrap were closer to the empirical standard error than those using the simple bootstrap. The simple bootstrap methods tended to underestimate. In addition, the coverage rates of a 95% confidence interval using the complex bootstrap were closer to the advertised rate of 0.95. We applied the two methods to a real data example and found also that the estimate of the standard error using the simple bootstrap was smaller than that using the complex bootstrap.