• Title/Summary/Keyword: Statistical samples

Search Result 1,651, Processing Time 0.029 seconds

An analysis of Mathematical Knowledge for Teaching of statistical estimation (통계적 추정을 가르치기 위한 수학적 지식(MKT)의 분석)

  • Choi, Min Jeong;Lee, Jong Hak;Kim, Won Kyung
    • The Mathematical Education
    • /
    • v.55 no.3
    • /
    • pp.317-334
    • /
    • 2016
  • Knowledge and data interpretation on statistical estimation was important to have statistical literacy that current curriculum was said not to satisfy. The author investigated mathematics teachers' MKT on statistical estimation concerning interpretation of confidence interval by using questionnaire and interview. SMK of teachers' confidence was limited to the area of textbooks to be difficult to interpret data of real life context. Most of teachers wrongly understood SMK of interpretation of confidence interval to have influence upon PCK making correction of students' wrong concept. SMK of samples and sampling distribution that were basic concept of reliability and confidence interval cognized representation of samples rather exactly not to understand importance and value of not only variability but also size of the sample exactly, and not to cognize appropriateness and needs of each stage from sampling to confidence interval estimation to have great difficulty at proper teaching of statistical estimation. PCK that had teaching method had problem of a lot of misconception. MKT of sample and sampling distribution that interpreted confidence interval had almost no relation with teachers' experience to require opportunity for development of teacher professionalism. Therefore, teachers were asked to estimate statistic and to get confidence interval and to understand concept of the sample and think much of not only relationship of each concept but also validity of estimated values, and to have knowledge enough to interpret data of real life contexts, and to think and discuss students' concepts. So, textbooks should introduce actual concepts at real life context to make use of exact orthography and to let teachers be reeducated for development of professionalism.

A sample size calibration approach for the p-value problem in huge samples

  • Park, Yousung;Jeon, Saebom;Kwon, Tae Yeon
    • Communications for Statistical Applications and Methods
    • /
    • v.25 no.5
    • /
    • pp.545-557
    • /
    • 2018
  • The inclusion of covariates in the model often affects not only the estimates of meaningful variables of interest but also its statistical significance. Such gap between statistical and subject-matter significance is a critical issue in huge sample studies. A popular huge sample study, the sample cohort data from Korean National Health Insurance Service, showed such gap of significance in the inference for the effect of obesity on cause of mortality, requiring careful consideration. In this regard, this paper proposes a sample size calibration method based on a Monte Carlo t (or z)-test approach without Monte Carlo simulation, and also proposes a test procedure for subject-matter significance using this calibration method in order to complement the deflated p-value in the huge sample size. Our calibration method shows no subject-matter significance of the obesity paradox regardless of race, sex, and age groups, unlike traditional statistical suggestions based on p-values.

Bayesian Test of Quasi-Independence in a Sparse Two-Way Contingency Table

  • Kwak, Sang-Gyu;Kim, Dal-Ho
    • Communications for Statistical Applications and Methods
    • /
    • v.19 no.3
    • /
    • pp.495-500
    • /
    • 2012
  • We consider a Bayesian test of independence in a two-way contingency table that has some zero cells. To do this, we take a three-stage hierarchical Bayesian model under each hypothesis. For prior, we use Dirichlet density to model the marginal cell and each cell probabilities. Our method does not require complicated computation such as a Metropolis-Hastings algorithm to draw samples from each posterior density of parameters. We draw samples using a Gibbs sampler with a grid method. For complicated posterior formulas, we apply the Monte-Carlo integration and the sampling important resampling algorithm. We compare the values of the Bayes factor with the results of a chi-square test and the likelihood ratio test.

WLDF: Effective Statistical Shape Feature for Cracked Tongue Recognition

  • Li, Xiao-qiang;Wang, Dan;Cui, Qing
    • Journal of Electrical Engineering and Technology
    • /
    • v.12 no.1
    • /
    • pp.420-427
    • /
    • 2017
  • This paper proposes a new method using Wide Line Detector based statistical shape Feature (WLDF) to identify whether or not a tongue is cracked; a cracked tongue is one of the most frequently used visible features for diagnosis in traditional Chinese Medicine (TCM). We first detected a wide line in the tongue image, and then extracted WLDF, such as the maximum length of each detected region, and the ratio between maximum length and the area of the detected region. We trained a binary support vector machine (SVM) based on the WLDF to build a classifier for cracked tongues. We conducted an experiment based on our proposed scheme, using 196 samples of cracked tongues and 245 samples of non-cracked tongues. The results of the experiment indicate that the recognition accuracy of the proposed method is greater than 95%. In addition, we provide an analysis of the results of this experiment with different parameters, demonstrating the feasibility and effectiveness of the proposed scheme.

To study of optimal subgroup size for estimating variance on autocorrelated small samples (소표본 자기상관 자료의 분산 추정을 위한 최적 부분군 크기에 대한 연구)

  • Lee, Jong-Seon;Lee, Jae-Jun;Bae, Soon-Hee
    • Proceedings of the Korean Society for Quality Management Conference
    • /
    • 2007.04a
    • /
    • pp.302-309
    • /
    • 2007
  • To conduct statistical process control needs the assumption that the process data are independent. However, most of chemical processes, like a semi-conduct processes do not satisfy the assumption because of autocorrelation. It causes abnormal out of control signal in the process control and misleading process capability. In this study, we introduce that Shore's method to solve the problem and to find the optimal subgroup size to estimate variance for AR(l) model. Especially, we focus on finding an actual subgroup size for small samples using simulation. It may be very useful for statistical process control to analyze process capability and to make a Shewhart chart properly.

  • PDF

Estimation for the Exponentiated Exponential Distribution Based on Multiply Type-II Censored Samples

  • Kang Suk-Bok;Park Sun-Mi
    • Communications for Statistical Applications and Methods
    • /
    • v.12 no.3
    • /
    • pp.643-652
    • /
    • 2005
  • It has been known that the exponentiated exponential distribution can be used as a possible alternative to the gamma distribution or the Weibull distribution in many situations. But the maximum likelihood method does not admit explicit solutions when the sample is multiply censored. So we derive the approximate maximum likelihood estimators for the location and scale parameters in the exponentiated exponential distribution that are explicit function of order statistics. We also compare the proposed estimators in the sense of the mean squared error for various censored samples.

PERFORMANCE EVALUATION VIA MONTE CARLO IMPORTANCE SAMPLING IN SINGLE USER DIGITAL COMMUNICATION SYSTEMS

  • Oh Man-Suk
    • Journal of the Korean Statistical Society
    • /
    • v.35 no.2
    • /
    • pp.157-166
    • /
    • 2006
  • This research proposes an efficient Monte Carlo algorithm for computing error probability in high performance digital communication st stems. It characterizes special features of the problem and suggests an importance sampling algorithm specially designed to handle the problem. It uses a shifted exponential density as the importance sampling density, and shows an adaptive way of choosing the rate and the origin of the shifted exponential density. Instead of equal allocation, an intelligent allocation of the samples is proposed so that more samples are allocated to more important part of the error probability. The algorithm uses the nested feature of the error space and avoids redundancy in estimating the probability. The algorithm is applied to an example data set and shows a great improvement in accuracy of the error probability estimation.

Estimation for the Half Logistic Distribution under Progressive Type-II Censoring

  • Kang, Suk-Bok;Cho, Young-Seuk;Han, Jun-Tae
    • Communications for Statistical Applications and Methods
    • /
    • v.15 no.6
    • /
    • pp.815-823
    • /
    • 2008
  • In this paper, we derive the approximate maximum likelihood estimators(AMLEs) and maximum likelihood estimator of the scale parameter in a half-logistic distribution based on progressive Type-II censored samples. We compare the proposed estimators in the sense of the mean squared error for various censored samples. We also obtain the approximate maximum likelihood estimators of the reliability function using the proposed estimators. We compare the proposed estimators in the sense of the mean squared error.

A View on the Validity of Central Limit Theorem: An Empirical Study Using Random Samples from Uniform Distribution

  • Lee, Chanmi;Kim, Seungah;Jeong, Jaesik
    • Communications for Statistical Applications and Methods
    • /
    • v.21 no.6
    • /
    • pp.539-559
    • /
    • 2014
  • We derive the exact distribution of summation for random samples from uniform distribution and then compare the exact distribution with the approximated normal distribution obtained by the central limit theorem. To check the similarity between two distributions, we consider five existing normality tests based on the difference between the target normal distribution and empirical distribution: Anderson-Darling test, Kolmogorov-Smirnov test, Cramer-von Mises test, Shapiro-Wilk test and Shaprio-Francia test. For the purpose of comparison, those normality tests are applied to the simulated data. It can sometimes be difficult to derive an exact distribution. Thus, we try two different transformations to find out which transform is easier to get the exact distribution in terms of calculation complexity. We compare two transformations and comment on the advantages and disadvantages for each transformation.

A Study of Non-parametric Statistical Tests to Quantify the Change of Water Quality (수질변화의 계량화를 위한 비모수적 통계 준거에 관한 연구)

  • Lee, Sang-Hoon
    • Journal of Environmental Impact Assessment
    • /
    • v.6 no.1
    • /
    • pp.111-119
    • /
    • 1997
  • This study was carried out to suggest the best statistical test which may be used to quantify the change of water quality between two groups. Traditional t-test may not be used in cases where the normality of underlying population distribution is not assured. Three non-parametric tests which are based on the relative order of the measurements, were studied to find out the applicability in water quality data analysis. The sign test is based on the sign of the deviation of the measurement from the median value, and the binomial distribution table is used. The signed rank test utilizes not only the sign but also the magnitude of the deviation. The Wilcoxon rank-sum test which is basically same as Mann-Whitney test, tests the mean difference between two independent samples which may have missing data. Among the three non-parametric tests studied, the singed rank test was found out to be applicable in the quantification of the change of water quality between two samples.

  • PDF