• Title/Summary/Keyword: Complex Sample Design

Search Result 153, Processing Time 0.024 seconds

Measurement Error Variance Estimation Based on Complex Survey Data with Subsample Re-Measurements

  • Heo, Sunyeong;Eltinge, John L.
    • Communications for Statistical Applications and Methods
    • /
    • v.10 no.2
    • /
    • pp.553-566
    • /
    • 2003
  • In many cases, the measurement error variances may be functions of the unknown true values or related covariates. This paper considers design-based estimators of the parameters of these variance functions based on the within-unit sample variances. This paper devotes to: (1) define an error scale factor $\delta$; (2) develop estimators of the parameters of the linear measurement error variance function of the true values under large-sample and small-error conditions; (3) use propensity methods to adjust survey weights to account for possible selection effects at the replicate level. The proposed methods are applied to medical examination data from the U.S. Third National Health and Nutrition Examination Survey (NHANES III).

A Study on Diagnostics Method for Categorical Data (범주형 자료의 진단방법에 관한 연구)

  • 이선규;조범석
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.18 no.33
    • /
    • pp.93-102
    • /
    • 1995
  • In this study we are concerned with the diagnostics method of cross-classified categorical data using logistic regression model of binary response models for cell proportions. under this model, we could examine the goodness-of-fit of the models using Pearson's $x^2$test statistic and likelihood ratio statistic. Under this model, these statistics are assumed that sample survey schemes are with replacement sampling model. But these statistics are often inappropriate for analysing contingency tables consists of complex sampling schemes obtained sample survey data. In this study we are examined diagnostics procedures detecting any outlying cell proportions and influential observations on design space in logistic regression modeltake account of the survey design effects.

  • PDF

Error cause analysis of Pearson test statistics for k-population homogeneity test (k-모집단 동질성검정에서 피어슨검정의 오차성분 분석에 관한 연구)

  • Heo, Sunyeong
    • Journal of the Korean Data and Information Science Society
    • /
    • v.24 no.4
    • /
    • pp.815-824
    • /
    • 2013
  • Traditional Pearson chi-squared test is not appropriate for the data collected by the complex sample design. When one uses the traditional Pearson chi-squared test to the complex sample categorical data, it may give wrong test results, and the error may occur not only due to the biased variance estimators but also due to the biased point estimators of cell proportions. In this study, the design based consistent Wald test statistics was derived for k-population homogeneity test, and the traditional Pearson chi-squared test statistics was partitioned into three parts according to the causes of error; the error due to the bias of variance estimator, the error due to the bias of cell proportion estimator, and the unseparated error due to the both bias of variance estimator and bias of cell proportion estimator. An analysis was conducted for empirical results of the relative size of each error component to the Pearson chi-squared test statistics. The second year data from the fourth Korean national health and nutrition examination survey (KNHANES, IV-2) was used for the analysis. The empirical results show that the relative size of error from the bias of variance estimator was relatively larger than the size of error from the bias of cell proportion estimator, but its degrees were different variable by variable.

Optical Detection of Red Blood Cell Aggregation under vibration (진동장에서의 적혈구 응집성을 측정하는 광학적 방법)

  • Jang, Ju-Hee;Park, Myung-Su;Ku, Yun-Hee;Shin, Se-Hyun
    • Proceedings of the KSME Conference
    • /
    • 2004.11a
    • /
    • pp.1510-1515
    • /
    • 2004
  • Aggregability of red blood cells (RBCs) was determined by a laser backscattering light analysis in a microfluidic channel. Available techniques for RBC aggregation often adopt a rotational Couette-flow using bob-and-cup system for disaggregating RBCs, which causes the system to be complex and expensive. A disposable microfluidic channel and vibration generating mechanism were used in the proposed new detection system for RBC aggregation. Prior to measurement, RBC aggregates in a blood sample were completely disaggregated by applying vibration-induced shear. With the present apparatus, the aggregation indexes of RBCs can be easily measured with small quantities of blood sample. The measurements with the present aggregometer were compared with those of LORCA and showed a strong correlation between them. The aggregability of the defibrinogenated blood RBCs is markedly lower than that of the normal RBCs. The noble feature of this design is the vibration-induced disaggregation mechanism, which enables to incorporate disposable element that holds the blood sample.

  • PDF

A System Decomposition Technique Using A Multi-Objective Genetic Algorithm (다목적 유전알고리듬을 이용한 시스템 분해 기법)

  • Park, Hyung-Wook;Kim, Min-Soo;Choi, Dong-Hoon
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.27 no.4
    • /
    • pp.499-506
    • /
    • 2003
  • The design cycle associated with large engineering systems requires an initial decomposition of the complex system into design processes which are coupled through the transference of output data. Some of these design processes may be grouped into iterative subcycles. In analyzing or optimizing such a coupled system, it is essential to determine the best order of the processes within these subcycles to reduce design cycle time and cost. This is accomplished by decomposing large multidisciplinary problems into several sub design structure matrices (DSMs) and processing them in parallel This paper proposes a new method for parallel decomposition of multidisciplinary problems to improve design efficiency by using the multi-objective genetic algorithm and two sample test cases are presented to show the effect of the suggested decomposition method.

Optimized Design of Dioxin Analysis for Water Sample

  • Choi, Jaewon;Lee, Jaehee;Kim, Kyoungsim;Kim, Sunheong;Bae, Kyunghee
    • Journal of Korean Society on Water Environment
    • /
    • v.21 no.3
    • /
    • pp.219-229
    • /
    • 2005
  • The analytical methods for dioxins in water sample from wastewater to tap water were reviewed. For extraction method, liquid-liquid extraction (LLE) has been widely used, however, this process needs too much time and man power. New approach including solid phase extraction (SPE) is now applicable to large volume of water sample with high extraction efficiency. Column clean up in classical analytical methods were very complex and time consuming procedures during decade. Modifications were tried to decrease solvent and reagents volume. Moreover, use of column connection method has been demonstrated in the environmental matrices. Instrumental configurations also have been improved, in which GC/MS/MS with large volume injection approach can analyze picogram levels. Absolute sensitivities of HRMS increased compared to old versions of double focusing sector type mass spectrometers. Based on these analytical evolutions during last 10 years, we tried to optimize the analytical method for dioxins in water sample from sample extraction to instrumental analysis.

Weight Function-based Sequential Maximin Distance Design to Enhance Accuracy and Robustness of Surrogate Model (대체모델의 정확성 및 강건성 향상을 위한 가중함수 기반 순차 최소거리최대화계획)

  • Jang, Junyong;Cho, Su-Gil;Lee, Tae Hee
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.39 no.4
    • /
    • pp.369-374
    • /
    • 2015
  • In order to efficiently optimize the problem involving complex computer codes or computationally expensive simulation, surrogate models are widely used. Because their accuracy significantly depends on sample points, many experimental designs have been proposed. One approach is the sequential design of experiments that consider existing information of responses. In earlier research, the correlation coefficients of the kriging surrogate model are introduced as weight parameters to define the scaled distance between sample points. However, if existing information is incorrect or lacking, new sample points can be misleading. Thus, our goal in this paper is to propose a weight function derived from correlation coefficients to generate new points robustly. To verify the performance of the proposed method, several existing sequential design methods are compared for use as mathematical examples.

Design-Based Properties of Least Square Estimators of Panel Regression Coefficients Based on Complex Panel Data (복합패널 데이터에 기초한 최소제곱 패널회귀추정량의 설계기반 성질)

  • Kim, Kyu-Seong
    • Communications for Statistical Applications and Methods
    • /
    • v.17 no.4
    • /
    • pp.515-525
    • /
    • 2010
  • We investigated design-based properties of the ordinary least square estimator(OLSE) and the weighted least square estimator(WLSE) in a panel regression model. Given a complex data we derive the magnitude of the design-based bias of two estimators and show that the bias of WLSE is smaller than that of OLSE. We also conducted a simulation study using Korean welfare panel data in order to compare design-based properties of two estimators numerically. In the study we found the followings. First, the relative bias of OLSE is nearly two times larger than that of WLSE and the bias ratio of OLSE is greater than that of WLSE. Also the relative bias of OLSE remains steady but that of WLSE becomes smaller as the sample size increases. Next, both the variance and mean square error(MSE) of two estimators decrease when the sample size increases. Also there is a tendency that the proportion of squared bias in MSE of OLSE increases as the sample size increase, but that of WLSE decreases. Finally, the variance of OLSE is smaller than that of WLSE in almost all cases and the MSE of OLSE is smaller in many cases. However, the number of cases of larger MSE of OLSE increases when the sample size increases.

System Decomposition Technique using Multiple Objective Genetic Algorithm (다목적 유전알고리듬을 이용한 시스템 분해 기법)

  • Park, Hyung-Wook;Kim, Min-Soo;Choi, Dong-Hoon
    • Proceedings of the KSME Conference
    • /
    • 2001.06c
    • /
    • pp.170-175
    • /
    • 2001
  • The design cycle associated with large engineering systems requires an initial decomposition of the complex system into design processes which are coupled through the transference of output data. Some of these design processes may be grouped into iterative subcycles. In analyzing or optimizing such a coupled system, it is essential to determine the best order of the processes within these subcycles to reduce design cycle time and cost. This is accomplished by decomposing large multidisciplinary problems into several multidisciplinary analysis subsystems (MDASS) and processing it in parallel. This paper proposes new strategy for parallel decomposition of multidisciplinary problems to improve design efficiency by using the multiple objective genetic algorithm (MOGA), and a sample test case is presented to show the effects of optimizing the sequence with MOGA.

  • PDF

Sample Size and Statistical Power Calculation in Genetic Association Studies

  • Hong, Eun-Pyo;Park, Ji-Wan
    • Genomics & Informatics
    • /
    • v.10 no.2
    • /
    • pp.117-122
    • /
    • 2012
  • A sample size with sufficient statistical power is critical to the success of genetic association studies to detect causal genes of human complex diseases. Genome-wide association studies require much larger sample sizes to achieve an adequate statistical power. We estimated the statistical power with increasing numbers of markers analyzed and compared the sample sizes that were required in case-control studies and case-parent studies. We computed the effective sample size and statistical power using Genetic Power Calculator. An analysis using a larger number of markers requires a larger sample size. Testing a single-nucleotide polymorphism (SNP) marker requires 248 cases, while testing 500,000 SNPs and 1 million markers requires 1,206 cases and 1,255 cases, respectively, under the assumption of an odds ratio of 2, 5% disease prevalence, 5% minor allele frequency, complete linkage disequilibrium (LD), 1:1 case/control ratio, and a 5% error rate in an allelic test. Under a dominant model, a smaller sample size is required to achieve 80% power than other genetic models. We found that a much lower sample size was required with a strong effect size, common SNP, and increased LD. In addition, studying a common disease in a case-control study of a 1:4 case-control ratio is one way to achieve higher statistical power. We also found that case-parent studies require more samples than case-control studies. Although we have not covered all plausible cases in study design, the estimates of sample size and statistical power computed under various assumptions in this study may be useful to determine the sample size in designing a population-based genetic association study.