• Title/Summary/Keyword: Statistical analysis

Search Result 17,648, Processing Time 0.049 seconds

Inappropriate Survey Design Analysis of the Korean National Health and Nutrition Examination Survey May Produce Biased Results

  • Kim, Yangho;Park, Sunmin;Kim, Nam-Soo;Lee, Byung-Kook
    • Journal of Preventive Medicine and Public Health
    • /
    • v.46 no.2
    • /
    • pp.96-104
    • /
    • 2013
  • Objectives: The inherent nature of the Korean National Health and Nutrition Examination Survey (KNHANES) design requires special analysis by incorporating sample weights, stratification, and clustering not used in ordinary statistical procedures. Methods: This study investigated the proportion of research papers that have used an appropriate statistical methodology out of the research papers analyzing the KNHANES cited in the PubMed online system from 2007 to 2012. We also compared differences in mean and regression estimates between the ordinary statistical data analyses without sampling weight and design-based data analyses using the KNHANES 2008 to 2010. Results: Of the 247 research articles cited in PubMed, only 19.8% of all articles used survey design analysis, compared with 80.2% of articles that used ordinary statistical analysis, treating KNHANES data as if it were collected using a simple random sampling method. Means and standard errors differed between the ordinary statistical data analyses and design-based analyses, and the standard errors in the design-based analyses tended to be larger than those in the ordinary statistical data analyses. Conclusions: Ignoring complex survey design can result in biased estimates and overstated significance levels. Sample weights, stratification, and clustering of the design must be incorporated into analyses to ensure the development of appropriate estimates and standard errors of these estimates.

Analysis of Articles Published in the Journal of Korea Institute of Oriental Medicine - from 2010 to 2012 (최근 3년간(2010-2012) 한국한의학연구원논문집 게재 논문의 통계기법에 관한 연구)

  • Kang, Kyungwon;Lee, Minhee;Kim, Jungeun;Lee, Sang-Hun;Choi, Sunmi
    • Korean Journal of Oriental Medicine
    • /
    • v.18 no.3
    • /
    • pp.127-132
    • /
    • 2012
  • Background and Purpose : This study was to investigate statistical validities and trends of previously reported papers that used various statistical techniques such as t-test and analysis of variance. Methods : To analyze the statistical procedures, 38 original articles using those statistical methods were selected from Journal of Korea Institute of Oriental Medicine(JKIOM) published from 2010 to 2012. Results : Analysis of variance and t-test were used in 20 papers (38.5%), 16 papers (30.8%) of 52 papers. Four articles(10.5%) did not report ${\alpha}$ values and nineteen papers(50.0%) of 38 ones were not tested for normal distribution. Five papers (13.2%) misused t-test and 3 papers (7.9%) did not carry out the multiple comparison. Conclusions : To improve the quality of JKIOM, The participation of statisticians in research design will reduce the significant errors in statistical interpretation of the results.

Implementation of Statistical Significance and Practical Significance Using Research Hypothesis and Statistical Hypothesis in the Six Sigma Projects (식스시그마 프로젝트에서 연구가설과 통계가설에 의한 통계적 유의성 및 실무적 유의성의 적용방안)

  • Choi, Sung-Woon
    • Journal of the Korea Safety Management & Science
    • /
    • v.15 no.1
    • /
    • pp.283-292
    • /
    • 2013
  • This paper aims to propose a new steps of hypothesis testing using analysis process and improvement process in the six sigma DMAIC. The six sigma implementation models proposed in this paper consist of six steps. The first step is to establish a research hypothesis by specification directionality and FBP(Falsibility By Popper). The second step is to translate the research hypothesis such as RHAT(Research Hypothesis Absent Type) and RHPT(Research Hypothesis Present Type) into statistical hypothesis such as $H_0$(Null Hypothesis) and $H_1$(Alternative Hypothesis). The third step is to implement statistical hypothesis testing by PBC(Proof By Contradiction) and proper sample size. The fourth step is to interpret the result of statistical hypothesis test. The fifth step is to establish the best conditions of product and process conditions by experimental optimization and interval estimation. The sixth step is to draw a conclusion by considering practical significance and statistical significance. Important for both quality practitioners and academicians, case analysis on six sigma projects with implementation guidelines are provided.

Teaching Statistical Graphics using R (R에 의한 통계그래픽스 : 강의 내용 및 방법의 논의)

  • Park, Dong-Ryeon
    • The Korean Journal of Applied Statistics
    • /
    • v.20 no.3
    • /
    • pp.619-634
    • /
    • 2007
  • It is well known that graphical display is critical to data analysis. A lot of research for data visualization has been done, so many effective graphical tools are now available. With the proper use of these graphical tools, we can penetrate the complex structure of data set easily. To enjoy the benefit of the powerful graphical display, the choice of the statistical software is very crucial. R is a popular open source software tool for statistical analysis and graphics, and can provide the very powerful graphics facilities. Moreover, many researchers believe that R is the best software for statistical graphics. In this paper, we would like to discuss what we teach and how we teach in statistical graphics course using R.

Nonlinear damage detection using higher statistical moments of structural responses

  • Yu, Ling;Zhu, Jun-Hua
    • Structural Engineering and Mechanics
    • /
    • v.54 no.2
    • /
    • pp.221-237
    • /
    • 2015
  • An integrated method is proposed for structural nonlinear damage detection based on time series analysis and the higher statistical moments of structural responses in this study. It combines the time series analysis, the higher statistical moments of AR model residual errors and the fuzzy c-means (FCM) clustering techniques. A few comprehensive damage indexes are developed in the arithmetic and geometric mean of the higher statistical moments, and are classified by using the FCM clustering method to achieve nonlinear damage detection. A series of the measured response data, downloaded from the web site of the Los Alamos National Laboratory (LANL) USA, from a three-storey building structure considering the environmental variety as well as different nonlinear damage cases, are analyzed and used to assess the performance of the new nonlinear damage detection method. The effectiveness and robustness of the new proposed method are finally analyzed and concluded.

An Introduction to Data Analysis (자료 분석의 기초)

  • Pak, Son-Il;Lee, Young-Won
    • Journal of Veterinary Clinics
    • /
    • v.26 no.3
    • /
    • pp.189-199
    • /
    • 2009
  • With the growing importance of evidence-based medicine, clinical or biomedical research relies critically on the validity and reliability of data, and the subsequent statistical inferences for medical decision-making may lead to valid conclusion. Despite widespread use of analytical techniques in papers published in the Journal of Veterinary Clinics statistical errors particularly in design of experiments, research methodology or data analysis methods are commonly encountered. These flaws often leading to misinterpretation of the data, thereby, subjected to inappropriate conclusions. This article is the first in a series of nontechnical introduction designed not to systemic review of medical statistics but intended to provide the journal readers with an understanding of common statistical concepts, including data scale, selection of appropriate statistical methods, descriptive statistics, data transformation, confidence interval, the principles of hypothesis testing, sampling distribution, and interpretation of results.

A Study on the Validity of the Statistical Collection and Analysis in Gwangju and Chonnam (통계자료의 수집 및 분석의 타당성에 관한 연구- 광주,전남지역을 중심으로 -)

  • 이화영
    • The Korean Journal of Applied Statistics
    • /
    • v.6 no.2
    • /
    • pp.443-452
    • /
    • 1993
  • A check list which includes the items that are to be considered in the process of the statistical data collection and analysis by non-scientific organizations is proposed. Based on the suggested check list, the output resulting from the statistical survey conducted by private organizations, banks, organs of expression and enterprises in Gwangju and Chonnam are examined about the validity of data collection and statistical analysis.

  • PDF

Application of data mining and statistical measurement of agricultural high-quality development

  • Yan Zhou
    • Advances in nano research
    • /
    • v.14 no.3
    • /
    • pp.225-234
    • /
    • 2023
  • In this study, we aim to use big data resources and statistical analysis to obtain a reliable instruction to reach high-quality and high yield agricultural yields. In this regard, soil type data, raining and temperature data as well as wheat production in each year are collected for a specific region. Using statistical methodology, the acquired data was cleaned to remove incomplete and defective data. Afterwards, using several classification methods in machine learning we tried to distinguish between different factors and their influence on the final crop yields. Comparing the proposed models' prediction using statistical quantities correlation factor and mean squared error between predicted values of the crop yield and actual values the efficacy of machine learning methods is discussed. The results of the analysis show high accuracy of machine learning methods in the prediction of the crop yields. Moreover, it is indicated that the random forest (RF) classification approach provides best results among other classification methods utilized in this study.

Predictive analysis in insurance: An application of generalized linear mixed models

  • Rosy Oh;Nayoung Woo;Jae Keun Yoo;Jae Youn Ahn
    • Communications for Statistical Applications and Methods
    • /
    • v.30 no.5
    • /
    • pp.437-451
    • /
    • 2023
  • Generalized linear models and generalized linear mixed models (GLMMs) are fundamental tools for predictive analyses. In insurance, GLMMs are particularly important, because they provide not only a tool for prediction but also a theoretical justification for setting premiums. Although thousands of resources are available for introducing GLMMs as a classical and fundamental tool in statistical analysis, few resources seem to be available for the insurance industry. This study targets insurance professionals already familiar with basic actuarial mathematics and explains GLMMs and their linkage with classical actuarial pricing tools, such as the Buhlmann premium method. Focus of the study is mainly on the modeling aspect of GLMMs and their application to pricing, while avoiding technical issues related to statistical estimation, which can be automatically handled by most statistical software.

Patterns of Data Analysis\ulcorner

  • Unwin, Antony
    • Journal of the Korean Statistical Society
    • /
    • v.30 no.2
    • /
    • pp.219-230
    • /
    • 2001
  • How do you carry out data analysis\ulcorner There are few texts and little theory. One approach could be to use a pattern language, an idea which has been successful in field as diverse as town planning and software engineering. Patterns for data analysis are defined and discussed, illustrated with examples.

  • PDF