• Title/Summary/Keyword: 근사 알고리듬

Search Result 110, Processing Time 0.022 seconds

Optimization of Data Recovery using Non-Linear Equalizer in Cellular Mobile Channel (셀룰라 이동통신 채널에서 비선형 등화기를 이용한 최적의 데이터 복원)

  • Choi, Sang-Ho;Ho, Kwang-Chun;Kim, Yung-Kwon
    • Journal of IKEEE
    • /
    • v.5 no.1 s.8
    • /
    • pp.1-7
    • /
    • 2001
  • In this paper, we have investigated the CDMA(Code Division Multiple Access) Cellular System with non-linear equalizer in reverse link channel. In general, due to unknown characteristics of channel in the wireless communication, the distribution of the observables cannot be specified by a finite set of parameters; instead, we partitioned the m-dimensional sample space Into a finite number of disjointed regions by using quantiles and a vector quantizer based on training samples. The algorithm proposed is based on a piecewise approximation to regression function based on quantiles and conditional partition moments which are estimated by Robbins Monro Stochastic Approximation (RMSA) algorithm. The resulting equalizers and detectors are robust in the sense that they are insensitive to variations in noise distributions. The main idea is that the robust equalizers and robust partition detectors yield better performance in equiprobably partitioned subspace of observations than the conventional equalizer in unpartitioned observation space under any condition. And also, we apply this idea to the CDMA system and analyze the BER performance.

  • PDF

Automatic Generation of 3D Face Model from Trinocular Images (Trinocular 영상을 이용한 3D 얼굴 모델 자동 생성)

  • Yi, Kwang-Do;Ahn, Sang-Chul;Kwon, Yong-Moo;Ko, Han-Seok;Kim, Hyoung-Gon
    • Journal of the Korean Institute of Telematics and Electronics S
    • /
    • v.36S no.7
    • /
    • pp.104-115
    • /
    • 1999
  • This paper proposes an efficient method for 3D modeling of a human face from trinocular images by reconstructing face surface using range data. By using a trinocular camera system, we mitigated the tradeoff between the occlusion problem and the range resolution limitation which is the critical limitation in binocular camera system. We also propose an MPC_MBS (Matching Pixel Count Multiple Baseline Stereo) area-based matching method to reduce boundary overreach phenomenon and to improve both of accuracy and precision in matching. In this method, the computing time can be reduced significantly by removing the redundancies. In the model generation sub-pixel accurate surface data are achieved by 2D interpolation of disparity values, and are sampled to make regular triangular meshes. The data size of the triangular mesh model can be controlled by merging the vertices that lie on the same plane within user defined error threshold.

  • PDF

Design Optimization for Loop Heat Pipe Using Tabu Search (Tabu Search를 이용한 Loop Heat Pipe의 최적설계에 관한 연구)

  • Park, Yong-Jin;Yun, Su-Hwan;Ku, Yo-Cheun;Lee, Dong-Ho
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.37 no.8
    • /
    • pp.737-743
    • /
    • 2009
  • Design optimization process and results of Loop Heat Pipe(LHP) using Tabu Search have been presented in this study. An objective of optimization is to reduce a mass of the LHP with satisfying operating temperature of a Lithium Ion battery onboard an aircraft. The battery is assumed to be used as power supply of air borne high energy laser system because of its high specific energy. The analytical models are based on a steady state mathematical model and the design optimization is performed using a Meta Model and Tabu Search. As an optimization results, the Tabu search algorithm guarantees global optimum with small computation time. Due to searching by random numbers, initial value is dominant factor to search global optimum. The optimization process could reduce the mass of the LHP which express the same performance as an published LHP.

An analysis of BER performance of LDPC decoder for WiMAX (WiMAX용 LDPC 복호기의 비트오율 성능 분석)

  • Kim, Hae-Ju;Shin, Kyung-Wook
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2010.05a
    • /
    • pp.771-774
    • /
    • 2010
  • In this paper, BER performance of LDPC(Low-Density Parity-Check) decoder for WiMAX is analyzed, and optimal design conditions of LDPC decoder are derived. The min-sum LDPC decoding algorithm which is based on an approximation of LLR sum-product algorithm is modeled and simulated by Matlab, and it is analyzed that the effects of LLR approximation bit-width and maximum iteration cycles on the bit error rate(BER) performance of LDCP decoder. The parity check matrix for IEEE 802.16e standard which has block length of 2304 and code rate of 1/2 is used, and AWGN channel with QPSK modulation is assumed. The simulation results show that optimal BER performance is achieved for 7 iteration cycles and LLR bit-width of (8,6).

  • PDF

Dose Verification Study of Brachytherapy Plans Using Monte Carlo Methods and CT Images (CT 영상 및 몬테칼로 계산에 기반한 근접 방사선치료계획의 선량분포 평가 방법 연구)

  • Cheong, Kwang-Ho;Lee, Me-Yeon;Kang, Sei-Kwon;Bae, Hoon-Sik;Park, So-Ah;Kim, Kyoung-Joo;Hwang, Tae-Jin;Oh, Do-Hoon
    • Progress in Medical Physics
    • /
    • v.21 no.3
    • /
    • pp.253-260
    • /
    • 2010
  • Most brachytherapy treatment planning systems employ a dosimetry formalism based on the AAPM TG-43 report which does not appropriately consider tissue heterogeneity. In this study we aimed to set up a simple Monte Carlo-based intracavitary high-dose-rate brachytherapy (IC-HDRB) plan verification platform, focusing particularly on the robustness of the direct Monte Carlo dose calculation using material and density information derived from CT images. CT images of slab phantoms and a uterine cervical cancer patient were used for brachytherapy plans based on the Plato (Nucletron, Netherlands) brachytherapy planning system. Monte Carlo simulations were implemented using the parameters from the Plato system and compared with the EBT film dosimetry and conventional dose computations. EGSnrc based DOSXYZnrc code was used for Monte Carlo simulations. Each $^{192}Ir$ source of the afterloader was approximately modeled as a parallel-piped shape inside the converted CT data set whose voxel size was $2{\times}2{\times}2\;mm^3$. Bracytherapy dose calculations based on the TG-43 showed good agreement with the Monte Carlo results in a homogeneous media whose density was close to water, but there were significant errors in high-density materials. For a patient case, A and B point dose differences were less than 3%, while the mean dose discrepancy was as much as 5%. Conventional dose computation methods might underdose the targets by not accounting for the effects of high-density materials. The proposed platform was shown to be feasible and to have good dose calculation accuracy. One should be careful when confirming the plan using a conventional brachytherapy dose computation method, and moreover, an independent dose verification system as developed in this study might be helpful.

Development of Intelligent ATP System Using Genetic Algorithm (유전 알고리듬을 적용한 지능형 ATP 시스템 개발)

  • Kim, Tai-Young
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.4
    • /
    • pp.131-145
    • /
    • 2010
  • The framework for making a coordinated decision for large-scale facilities has become an important issue in supply chain(SC) management research. The competitive business environment requires companies to continuously search for the ways to achieve high efficiency and lower operational costs. In the areas of production/distribution planning, many researchers and practitioners have developedand evaluated the deterministic models to coordinate important and interrelated logistic decisions such as capacity management, inventory allocation, and vehicle routing. They initially have investigated the various process of SC separately and later become more interested in such problems encompassing the whole SC system. The accurate quotation of ATP(Available-To-Promise) plays a very important role in enhancing customer satisfaction and fill rate maximization. The complexity for intelligent manufacturing system, which includes all the linkages among procurement, production, and distribution, makes the accurate quotation of ATP be a quite difficult job. In addition to, many researchers assumed ATP model with integer time. However, in industry practices, integer times are very rare and the model developed using integer times is therefore approximating the real system. Various alternative models for an ATP system with time lags have been developed and evaluated. In most cases, these models have assumed that the time lags are integer multiples of a unit time grid. However, integer time lags are very rare in practices, and therefore models developed using integer time lags only approximate real systems. The differences occurring by this approximation frequently result in significant accuracy degradations. To introduce the ATP model with time lags, we first introduce the dynamic production function. Hackman and Leachman's dynamic production function in initiated research directly related to the topic of this paper. They propose a modeling framework for a system with non-integer time lags and show how to apply the framework to a variety of systems including continues time series, manufacturing resource planning and critical path method. Their formulation requires no additional variables or constraints and is capable of representing real world systems more accurately. Previously, to cope with non-integer time lags, they usually model a concerned system either by rounding lags to the nearest integers or by subdividing the time grid to make the lags become integer multiples of the grid. But each approach has a critical weakness: the first approach underestimates, potentially leading to infeasibilities or overestimates lead times, potentially resulting in excessive work-inprocesses. The second approach drastically inflates the problem size. We consider an optimized ATP system with non-integer time lag in supply chain management. We focus on a worldwide headquarter, distribution centers, and manufacturing facilities are globally networked. We develop a mixed integer programming(MIP) model for ATP process, which has the definition of required data flow. The illustrative ATP module shows the proposed system is largely affected inSCM. The system we are concerned is composed of a multiple production facility with multiple products, multiple distribution centers and multiple customers. For the system, we consider an ATP scheduling and capacity allocationproblem. In this study, we proposed the model for the ATP system in SCM using the dynamic production function considering the non-integer time lags. The model is developed under the framework suitable for the non-integer lags and, therefore, is more accurate than the models we usually encounter. We developed intelligent ATP System for this model using genetic algorithm. We focus on a capacitated production planning and capacity allocation problem, develop a mixed integer programming model, and propose an efficient heuristic procedure using an evolutionary system to solve it efficiently. This method makes it possible for the population to reach the approximate solution easily. Moreover, we designed and utilized a representation scheme that allows the proposed models to represent real variables. The proposed regeneration procedures, which evaluate each infeasible chromosome, makes the solutions converge to the optimum quickly.

Numerical Simulation of Normal Logging Measurements in the Proximity of Earth Surface (지표 부근에서의 노멀전기검층 수치 모델링)

  • Nam, Myung-Jin;Hwang, Se-Ho
    • Economic and Environmental Geology
    • /
    • v.43 no.3
    • /
    • pp.259-267
    • /
    • 2010
  • Resistivity logging instruments were designed to measure electrical resistivity of formation, which can be directly interpreted to provide water-saturation profile. Short and long normal logging measurements are made under groundwater level. In some investigation sites, groundwater level reaches to a depth of a few meters. It has come to attention that the proximity of groundwater level might distort short and long normal logging readings, when the measurements are made near groundwater level, owing to the proximity of an insulating air. This study investigates the effects of the proximity of groundwater level (and also the proximity of earth surface) on the normal by simulating normal logging measurements near groundwater level. In the simulation, we consider all the details of real logging situation, i.e., the presence of wellbore, the tool mandrel with current and potential electrodes, and currentreturn and reference-potential electrodes. We also model the air to include the earth’'s surface in the simulation rather than the customary choice of imposing a boundary condition. To obtain apparent resistivity, we compute the voltage, i.e., potential difference between monitoring and reference electrodes. For the simulation, we use a twodimensional, goal-oriented and high-order self-adaptive hp finite element refinement strategy (h denotes the element size and p the polynomial order of approximation within each element) to obtain accurate simulation results. Numerical results indicate that distortion on the normal logging is greater when the reference potential electrode is closer to the borehole and distortions on long normal logging are larger than those on short normal logging.

A Small-area Hardware Implementation of EGML-based Moving Object Detection Processor (EGML 기반 이동객체 검출 프로세서의 저면적 하드웨어 구현)

  • Sung, Mi-ji;Shin, Kyung-wook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.21 no.12
    • /
    • pp.2213-2220
    • /
    • 2017
  • This paper proposes an efficient approach for hardware implementation of moving object detection (MOD) processor using effective Gaussian mixture learning (EGML)-based background subtraction method. Arithmetic units used in background generation were implemented using LUT-based approximation to reduce hardware complexity. Hardware resources used for both background subtraction and Gaussian probability density calculation were shared. The MOD processor was verified by FPGA-in-the-loop simulation using MATLAB/Simulink. The MOD performance was evaluated by using six types of video defined in IEEE CDW-2014 dataset, which resulted the average of recall value of 0.7700, the average of precision value of 0.7170, and the average of F-measure value of 0.7293. The MOD processor was implemented with 882 slices and block RAM of $146{\times}36kbits$ on Virtex5 FPGA, resulting in 60% hardware reduction compared to conventional design based on EGML. It was estimated that the MOD processor could operate with 75 MHz clock, resulting in real-time processing of $800{\times}600$ video with a frame rate of 39 fps.

Review on the Three-Dimensional Inversion of Magnetotelluric Date (MT 자료의 3차원 역산 개관)

  • Kim Hee Joon;Nam Myung Jin;Han Nuree;Choi Jihyang;Lee Tae Jong;Song Yoonho;Suh Jung Hee
    • Geophysics and Geophysical Exploration
    • /
    • v.7 no.3
    • /
    • pp.207-212
    • /
    • 2004
  • This article reviews recent developments in three-dimensional (3-D) magntotelluric (MT) imaging. The inversion of MT data is fundamentally ill-posed, and therefore the resultant solution is non-unique. A regularizing scheme must be involved to reduce the non-uniqueness while retaining certain a priori information in the solution. The standard approach to nonlinear inversion in geophysis has been the Gauss-Newton method, which solves a sequence of linearized inverse problems. When running to convergence, the algorithm minimizes an objective function over the space of models and in the sense produces an optimal solution of the inverse problem. The general usefulness of iterative, linearized inversion algorithms, however is greatly limited in 3-D MT applications by the requirement of computing the Jacobian(partial derivative, sensitivity) matrix of the forward problem. The difficulty may be relaxed using conjugate gradients(CG) methods. A linear CG technique is used to solve each step of Gauss-Newton iterations incompletely, while the method of nonlinear CG is applied directly to the minimization of the objective function. These CG techniques replace computation of jacobian matrix and solution of a large linear system with computations equivalent to only three forward problems per inversion iteration. Consequently, the algorithms are efficient in computational speed and memory requirement, making 3-D inversion feasible.

Impact of Lambertian Cloud Top Pressure Error on Ozone Profile Retrieval Using OMI (램버시안 구름 모델의 운정기압 오차가 OMI 오존 프로파일 산출에 미치는 영향)

  • Nam, Hyeonshik;Kim, Jae Hawn;Shin, Daegeun;Baek, Kanghyun
    • Korean Journal of Remote Sensing
    • /
    • v.35 no.3
    • /
    • pp.347-358
    • /
    • 2019
  • Lambertian cloud model (Lambertian Cloud Model) is the simplified cloud model which is used to effectively retrieve the vertical ozone distribution of the atmosphere where the clouds exist. By using the Lambertian cloud model, the optical characteristics of clouds required for radiative transfer simulation are parametrized by Optical Centroid Cloud Pressure (OCCP) and Effective Cloud Fraction (ECF), and the accuracy of each parameter greatly affects the radiation simulation accuracy. However, it is very difficult to generalize the vertical ozone error due to the OCCP error because it varies depending on the radiation environment and algorithm setting. In addition, it is also difficult to analyze the effect of OCCP error because it is mixed with other errors that occur in the vertical ozone calculation process. This study analyzed the ozone retrieval error due to OCCP error using two methods. First, we simulated the impact of OCCP error on ozone retrieval based on Optimal Estimation. Using LIDORT radiation model, the radiation error due to the OCCP error is calculated. In order to convert the radiation error to the ozone calculation error, the radiation error is assigned to the conversion equation of the optimal estimation method. The results show that when the OCCP error occurs by 100 hPa, the total ozone is overestimated by 2.7%. Second, a case analysis is carried out to find the ozone retrieval error due to OCCP error. For the case analysis, the ozone retrieval error is simulated assuming OCCP error and compared with the ozone error in the case of PROFOZ 2005-2006, an OMI ozone profile product. In order to define the ozone error in the case, we assumed an ideal assumption. Considering albedo, and the horizontal change of ozone for satisfying the assumption, the 49 cases are selected. As a result, 27 out of 49 cases(about 55%)showed a correlation of 0.5 or more. This result show that the error of OCCP has a significant influence on the accuracy of ozone profile calculation.