• Title/Summary/Keyword: Sampling-Based Algorithm

Search Result 477, Processing Time 0.026 seconds

Multivariate Stratification Method for the Multipurpose Sample Survey : A Case Study of the Sample Design for Fisher Production Survey (다목적 표본조사를 위한 다변량 층화 : 어업비계통생산량조사를 위한 표본설계 사례)

  • Park, Jin-Woo;Kim, Young-Won;Lee, Seok-Hoon;Shin, Ji-Eun
    • Survey Research
    • /
    • v.9 no.1
    • /
    • pp.69-85
    • /
    • 2008
  • Stratification is a feature of the majority of field sample design. This paper considers the multivariate stratification strategy for multipurpose sample survey with several auxiliary variables. In a multipurpose survey, stratification procedure is very complicated because we have to simultaneously consider the efficiencies of stratification for several variables of interest. We propose stratification strategy based on factor analysis and cluster analysis using several stratification variables. To improve the efficiency of stratification, we first select the stratification variables by factor analysis, and then apply the K-means clustering algorithm to the formation of strata. An application of the stratification strategy in the sampling design for the Fisher Production Survey is discussed, and it turns out that the variances of estimators are significantly less than those obtained by simple random sampling.

  • PDF

STUDY ON APPLICATION OF NEURO-COMPUTER TO NONLINEAR FACTORS FOR TRAVEL OF AGRICULTURAL CRAWLER VEHICLES

  • Inaba, S.;Takase, A.;Inoue, E.;Yada, K.;Hashiguchi, K.
    • Proceedings of the Korean Society for Agricultural Machinery Conference
    • /
    • 2000.11b
    • /
    • pp.124-131
    • /
    • 2000
  • In this study, the NEURAL NETWORK (hereinafter referred to as NN) was applied to control of the nonlinear factors for turning movement of the crawler vehicle and experiment was carried out using a small model of crawler vehicle in order to inspect an application of NN. Furthermore, CHAOS NEURAL NETWORK (hereinafter referred to as CNN) was also applied to this control so as to compare with conventional NN. CNN is especially effective for plane in many variables with local minimum which conventional NN is apt to fall into, and it is relatively useful to nonlinear factors. Experiment of turning on the slope of crawler vehicle was performed in order to estimate an adaptability of nonlinear problems by NN and CNN. The inclination angles of the road surface which the vehicles travel on, were respectively 4deg, 8deg, 12deg. These field conditions were selected by the object for changing nonlinear magnitude in turning phenomenon of vehicle. Learning of NN and CNN was carried out by referring to positioning data obtained from measurement at every 15deg in turning. After learning, the sampling data at every 15deg were interpolated based on the constructed learning system of NN and CNN. Learning and simulation programs of NN and CNN were made by C language ("Association of research for algorithm of calculating machine (1992)"). As a result, conventional NN and CNN were available for interpolation of sampling data. Moreover, when nonlinear intensity is not so large under the field condition of small slope, interpolation performance of CNN was a little not so better than NN. However, when nonlinear intensity is large under the field condition of large slope, interpolation performance of CNN was relatively better than NN.

  • PDF

An Approximated Model of the Coefficients for Interchannel Interference of OFDM System with Frequency Offset (주파수 오프셋이 있는 OFDM시스템에서 채널간간섭의 간섭계수 근사화 모델)

  • Li, Shuang;Kwon, Hyeock-Chan;Kang, Seog-Geun
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.13 no.5
    • /
    • pp.917-922
    • /
    • 2018
  • In the conventional interchannel interference self-cancellation (ICI-SC) schemes, the length of sampling window is the same as the symbol length of orthogonal frequency division multiplexing (OFDM). Thus, the number of complex operations to compute the interference coefficient of each subchannel is significantly increased. To solve this problem, we present an approximated mathematical model for the coefficients of ICI-SC schemes. Based on the proposed approximation, we analyze mean squared error (MSE) and computational complexity of the ICI-SC schemes with the length of sampling window. As a result, the presented approximation has an error of less than 0.01% on the MSE compared to the original equation. When the number of subchannels is 1024, the number of complex computations for the interference coefficients is reduced by 98% or more. Since the computational complexity can be remarkably reduced without sacrificing the self-cancellation capability, it is considered that the proposed approximation is very useful to develop an algorithm for the ICI-SC scheme.

한 인구학도의 회고

  • 김택일
    • Korea journal of population studies
    • /
    • v.11 no.1
    • /
    • pp.1-13
    • /
    • 1988
  • This study examines the sampling bias that may have resulted from the large number of missing observations. Despite well-designed and reliable sampling procedures, the observed sample values in DSFH(Demographic Survey on Changes in Family and Household Structure, Japan) included many missing observations. The head administerd survey method of DSFH resulted in a large number of missing observations regarding characteristics of elderly non-head parents and their children. In addition, the response probability of a particular item in DSFH significantly differs by characteristics of elderly parents and their children. Furthermore, missing observations of many items occurred simultaneously. This complex pattern of missing observations critically limits the ability to produce an unbiased analysis. First, the large number of missing observations is likely to cause a misleading estimate of the standard error. Even worse, the possible dependency of missing observations on their latent values is likely to produce biased estimates of covariates. Two models are employed to solve the possible inference biases. First, EM algorithm is used to infer the missing values based on the knowledge of the association between the observed values and other covariates. Second, a selection model was employed given the suspicion that the probability of missing observations of proximity depends on its unobserved outcome.

  • PDF

Performance Analysis of Matching Cost Functions of Stereo Matching Algorithm for Making 3D Contents (3D 콘텐츠 생성에서의 스테레오 매칭 알고리즘에 대한 매칭 비용 함수 성능 분석)

  • Hong, Gwang-Soo;Jeong, Yeon-Kyu;Kim, Byung-Gyu
    • Convergence Security Journal
    • /
    • v.13 no.3
    • /
    • pp.9-15
    • /
    • 2013
  • Calculating of matching cost is an important for efficient stereo matching. To investigate the performance of matching process, the concepts of the existing methods are introduced. Also we analyze the performance and merits of them. The simplest matching costs assume constant intensities at matching image locations. We consider matching cost functions which can be distinguished between pixel-based and window-based approaches. The Pixel-based approach includes absolute differences (AD) and sampling-intensitive absolute differences (BT). The window-based approach includes the sum of the absolute differences, the sum of squared differences, the normalized cross-correlation, zero-mean normalized cross-correlation, census transform, and the absolute differences census transform (AD-Census). We evaluate matching cost functions in terms of accuracy and time complexity. In terms of the accuracy, AD-Census method shows the lowest matching error ratio (the best solution). The ZNCC method shows the lowest matching error ratio in non-occlusion and all evaluation part. But it performs high matching error ratio at the discontinuities evaluation part due to blurring effect in the boundary. The pixel-based AD method shows a low complexity in terms of time complexity.

Assessment of System Reliability and Capacity-Rating of Composite Steel Box-Girder Highway Bridges (합성 강 상자형 도로교의 체계신뢰성 해석 및 안전도평가)

  • Cho, Hyo Nam;Lee, Seung Jae;Kang, Kyoung Koo
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.13 no.2
    • /
    • pp.51-59
    • /
    • 1993
  • This paper develops practical and realistic reliability models and methods for the evaluation of system-reliability and system reliability-based rating of various types of box-girder bridge superstructures. The strength limit state model for box-girder bridges suggested in the paper are based on not only the basic flexural strength but also the strength interaction equations which simultaneously take into account flexure, shear and torsion. And the system reliability problem of box-girder superstructure is formulated as parallel-series models obtained from the FMA(Failure Mode Approach) based on major failure mechanisms or critical failure states of each girder. In the paper, an improved IST(Importance Sampling Technique) simulation algorithm is used for the system reliability analysis of the proposed models. This paper proposes a practical but rational approach for the evaluation of capacity rating in terms of the equivalent system-capacity rating corresponding to the estimated system-reliability index which is derived based on the concept of the equivalent FOSM(First Order Second Moment) form of system reliability index. The results of the reliability evaluation and rating of existing bridges indicate that the reserved reliability and capacity rating at system level are significantly different from those of element reliability or conventional methods especially in the case of highly redundant box-girder bridges.

  • PDF

Comparison of Univariate Kriging Algorithms for GIS-based Thematic Mapping with Ground Survey Data (현장 조사 자료를 이용한 GIS 기반 주제도 작성을 위한 단변량 크리깅 기법의 비교)

  • Park, No-Wook
    • Korean Journal of Remote Sensing
    • /
    • v.25 no.4
    • /
    • pp.321-338
    • /
    • 2009
  • The objective of this paper is to compare spatial prediction capabilities of univariate kriging algorithms for generating GIS-based thematic maps from ground survey data with asymmetric distributions. Four univariate kriging algorithms including traditional ordinary kriging, three non-linear transform-based kriging algorithms such as log-normal kriging, multi-Gaussian kriging and indicator kriging are applied for spatial interpolation of geochemical As and Pb elements. Cross validation based on a leave-one-out approach is applied and then prediction errors are computed. The impact of the sampling density of the ground survey data on the prediction errors are also investigated. Through the case study, indicator kriging showed the smallest prediction errors and superior prediction capabilities of very low and very high values. Other non-linear transform based kriging algorithms yielded better prediction capabilities than traditional ordinary kriging. Log-normal kriging which has been widely applied, however, produced biased estimation results (overall, overestimation). It is expected that such quantitative comparison results would be effectively used for the selection of an optimal kriging algorithm for spatial interpolation of ground survey data with asymmetric distributions.

EM Algorithm and Two Stage Model for Incomplete Data (불완전한 자료에 대한 보완기법(EM 알고리듬과 2단계(Two Stage) 모델))

  • 박경숙
    • Korea journal of population studies
    • /
    • v.21 no.1
    • /
    • pp.162-183
    • /
    • 1998
  • This study examines the sampling bias that may have resulted from the large number of missing observations. Despite well-designed and reliable sampling procedures, the observed sample values in DSFH(Demographic Survey on Changes in Family and Household Structure, Japan) included many missing observations. The head administerd survey method of DSFH resulted in a large number of missing observations regarding characteristics of elderly non-head parents and their children. In addition, the response probability of a particular item in DSFH significantly differs by characteristics of elderly parents and their children. Furthermore, missing observations of many items occurred simultaneously. This complex pattern of missing observations critically limits the ability to produce an unbiased analysis. First, the large number of missing observations is likely to cause a misleading estimate of the standard error. Even worse, the possible dependency of missing observations on their latent values is likely to produce biased estimates of covariates. Two models are employed to solve the possible inference biases. First, EM algorithm is used to infer the missing values based on the knowledge of the association between the observed values and other covariates. Second, a selection model was employed given the suspicion that the probability of missing observations of proximity depends on its unobserved outcome.

  • PDF

Performance Characteristics of an Ensemble Machine Learning Model for Turbidity Prediction With Improved Data Imbalance (데이터 불균형 개선에 따른 탁도 예측 앙상블 머신러닝 모형의 성능 특성)

  • HyunSeok Yang;Jungsu Park
    • Ecology and Resilient Infrastructure
    • /
    • v.10 no.4
    • /
    • pp.107-115
    • /
    • 2023
  • High turbidity in source water can have adverse effects on water treatment plant operations and aquatic ecosystems, necessitating turbidity management. Consequently, research aimed at predicting river turbidity continues. This study developed a multi-class classification model for prediction of turbidity using LightGBM (Light Gradient Boosting Machine), a representative ensemble machine learning algorithm. The model utilized data that was classified into four classes ranging from 1 to 4 based on turbidity, from low to high. The number of input data points used for analysis varied among classes, with 945, 763, 95, and 25 data points for classes 1 to 4, respectively. The developed model exhibited precisions of 0.85, 0.71, 0.26, and 0.30, as well as recalls of 0.82, 0.76, 0.19, and 0.60 for classes 1 to 4, respectively. The model tended to perform less effectively in the minority classes due to the limited data available for these classes. To address data imbalance, the SMOTE (Synthetic Minority Over-sampling Technique) algorithm was applied, resulting in improved model performance. For classes 1 to 4, the Precision and Recall of the improved model were 0.88, 0.71, 0.26, 0.25 and 0.79, 0.76, 0.38, 0.60, respectively. This demonstrated that alleviating data imbalance led to a significant enhancement in Recall of the model. Furthermore, to analyze the impact of differences in input data composition addressing the input data imbalance, input data was constructed with various ratios for each class, and the model performances were compared. The results indicate that an appropriate composition ratio for model input data improves the performance of the machine learning model.

Improved VFM Method for High Accuracy Flight Simulation (고정밀 비행 시뮬레이션을 위한 개선 VFM 기법 연구)

  • Lee, Chiho;Kim, Mukyeom;Lee, Jae-Lyun;Jeon, Kwon-Su;Tyan, Maxim;Lee, Jae-Woo
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.49 no.9
    • /
    • pp.709-719
    • /
    • 2021
  • Recent progress in analysis and flight simulation methods enables wider use of a virtual certification and reduces number of certification flight tests. Aerodynamic database (AeroDB) is one of the most important components for the flight simulation. It is composed of aerodynamic coefficients at a range of flight conditions and control deflections. This paper proposes and efficient method for construction of AeroDB that combines Gaussian Process based Variable Fidelity Modeling with adaptive sampling algorithm. A case study of virtual certification of a F-16 fighter is presented. Four AeroDB were constructed using different number and distribution of high-fidelity data points. The constructed database is then used to simulate gliding, short pitch, and roll response. Compliance with certification regulations is then checked. The case study demonstrates that the proposed method can significantly reduce number of high-fidelity data points while maintaining high accuracy of the simulation.