• Title/Summary/Keyword: Gradient descent

Search Result 339, Processing Time 0.022 seconds

Time-domain Seismic Waveform Inversion for Anisotropic media (이방성을 고려한 탄성매질에서의 시간영역 파형역산)

  • Lee, Ho-Yong;Min, Dong-Joo;Kwon, Byung-Doo;Yoo, Hai-Soo
    • 한국지구물리탐사학회:학술대회논문집
    • /
    • 2008.10a
    • /
    • pp.51-56
    • /
    • 2008
  • The waveform inversion for isotropic media has ever been studied since the 1980s, but there has been few studies for anisotropic media. We present a seismic waveform inversion algorithm for 2-D heterogeneous transversely isotropic structures. A cell-based finite difference algorithm for anisotropic media in time domain is adopted. The steepest descent during the non-linear iterative inversion approach is obtained by backpropagating residual errors using a reverse time migration technique. For scaling the gradient of a misfit function, we use the pseudo Hessian matrix which is assumed to neglect the zero-lag auto-correlation terms of impulse responses in the approximate Hessian matrix of the Gauss-Newton method. We demonstrate the use of these waveform inversion algorithm by applying them to a two layer model and the anisotropic Marmousi model data. With numerical examples, we show that it's difficult to converge to the true model when we assumed that anisotropic media are isotropic. Therefore, it is expected that our waveform inversion algorithm for anisotropic media is adequate to interpret real seismic exploration data.

  • PDF

Development of Gait Distance Measurement System Based on Inertial Measurement Units (관성측정장치를 이용한 보행거리 측정 시스템 개발)

  • Lee, K.H.;Kang, S.I.;Cho, J.S.;Lim, D.H.;Lee, J.S.;Kim, I.Y.
    • Journal of rehabilitation welfare engineering & assistive technology
    • /
    • v.9 no.2
    • /
    • pp.161-168
    • /
    • 2015
  • In this paper, we present an inertial sensor-based gait distance measurement system using accelerometer, gyroscope, and magnetometer. To minimize offset and gain error of inertial sensors, we performed the calibration using the self-made calibration jig with 9 degrees of freedom. For measuring accurate gait distance, we used gradient descent algorithm to remove gravity error and used analysis of gait pattern to remove drift error. Finally, we measured a gait distance by double-integration of the error-removed acceleration data. To evaluate the performance of our system, we walked 10m in a straight line indoors to observe the improvement of removing error which compared un-calibrated to calibrated data. Also, the gait distance measured by the system was compared to the measurement of the Vicon motion capture system. The evaluation resulted in the improvement of $31.4{\pm}14.38%$(mean${\pm}$S.D.), $78.64{\pm}10.84%$ and $69.71{\pm}26.25%$ for x, y and z axis, respectively when walked in a straight line, and a root mean square error of 0.10m, 0.16m, and 0.12m for x, y and z axis, respectively when compared to the Vicon motion capture system.

  • PDF

Application of Text-Classification Based Machine Learning in Predicting Psychiatric Diagnosis (텍스트 분류 기반 기계학습의 정신과 진단 예측 적용)

  • Pak, Doohyun;Hwang, Mingyu;Lee, Minji;Woo, Sung-Il;Hahn, Sang-Woo;Lee, Yeon Jung;Hwang, Jaeuk
    • Korean Journal of Biological Psychiatry
    • /
    • v.27 no.1
    • /
    • pp.18-26
    • /
    • 2020
  • Objectives The aim was to find effective vectorization and classification models to predict a psychiatric diagnosis from text-based medical records. Methods Electronic medical records (n = 494) of present illness were collected retrospectively in inpatient admission notes with three diagnoses of major depressive disorder, type 1 bipolar disorder, and schizophrenia. Data were split into 400 training data and 94 independent validation data. Data were vectorized by two different models such as term frequency-inverse document frequency (TF-IDF) and Doc2vec. Machine learning models for classification including stochastic gradient descent, logistic regression, support vector classification, and deep learning (DL) were applied to predict three psychiatric diagnoses. Five-fold cross-validation was used to find an effective model. Metrics such as accuracy, precision, recall, and F1-score were measured for comparison between the models. Results Five-fold cross-validation in training data showed DL model with Doc2vec was the most effective model to predict the diagnosis (accuracy = 0.87, F1-score = 0.87). However, these metrics have been reduced in independent test data set with final working DL models (accuracy = 0.79, F1-score = 0.79), while the model of logistic regression and support vector machine with Doc2vec showed slightly better performance (accuracy = 0.80, F1-score = 0.80) than the DL models with Doc2vec and others with TF-IDF. Conclusions The current results suggest that the vectorization may have more impact on the performance of classification than the machine learning model. However, data set had a number of limitations including small sample size, imbalance among the category, and its generalizability. With this regard, the need for research with multi-sites and large samples is suggested to improve the machine learning models.

A Fast-Loaming Algorithm for MLP in Pattern Recognition (패턴인식의 MLP 고속학습 알고리즘)

  • Lee, Tae-Seung;Choi, Ho-Jin
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.3
    • /
    • pp.344-355
    • /
    • 2002
  • Having a variety of good characteristics against other pattern recognition techniques, Multilayer Perceptron (MLP) has been used in wide applications. But, it is known that Error Backpropagation (EBP) algorithm which MLP uses in learning has a defect that requires relatively long leaning time. Because learning data in pattern recognition contain abundant redundancies, in order to increase learning speed it is very effective to use online-based teaming methods, which update parameters of MLP pattern by pattern. Typical online EBP algorithm applies fixed learning rate for each update of parameters. Though a large amount of speedup with online EBP can be obtained by choosing an appropriate fixed rate, fixing the rate leads to the problem that the algorithm cannot respond effectively to different leaning phases as the phases change and the learning pattern areas vary. To solve this problem, this paper defines learning as three phases and proposes a Instant Learning by Varying Rate and Skipping (ILVRS) method to reflect only necessary patterns when learning phases change. The basic concept of ILVRS is as follows. To discriminate and use necessary patterns which change as learning proceeds, (1) ILVRS uses a variable learning rate which is an error calculated from each pattern and is suppressed within a proper range, and (2) ILVRS bypasses unnecessary patterns in loaming phases. In this paper, an experimentation is conducted for speaker verification as an application of pattern recognition, and the results are presented to verify the performance of ILVRS.

Generalized Sigmidal Basis Function for Improving the Learning Performance fo Multilayer Perceptrons (다층 퍼셉트론의 학습 성능 개선을 위한 일반화된 시그모이드 베이시스 함수)

  • Park, Hye-Yeong;Lee, Gwan-Yong;Lee, Il-Byeong;Byeon, Hye-Ran
    • Journal of KIISE:Software and Applications
    • /
    • v.26 no.11
    • /
    • pp.1261-1269
    • /
    • 1999
  • 다층 퍼셉트론은 다양한 응용 분야에 성공적으로 적용되고 있는 대표적인 신경회로망 모델이다. 그러나 다층 퍼셉트론의 학습에서 나타나는 플라토에 기인한 느린 학습 속도와 지역 극소는 실제 응용문제에 적용함에 있어서 가장 큰 문제로 지적되어왔다. 이 문제를 해결하기 위해 여러 가지 다양한 학습알고리즘들이 개발되어 왔으나, 계산의 비효율성으로 인해 실제 문제에는 적용하기 힘든 예가 많은 등, 현재까지 만족할 만한 해결책은 제시되지 못하고 있다. 본 논문에서는 다층퍼셉트론의 베이시스 함수로 사용되는 시그모이드 함수를 보다 일반화된 형태로 정의하여 사용함으로써 학습에 있어서의 플라토를 완화하고, 지역극소에 빠지는 것을 줄이는 접근방법을 소개한다. 본 방법은 기존의 변형된 가중치 수정식을 사용한 학습 속도 향상의 방법들과는 다른 접근 방법을 택함으로써 기존의 방법들과 함께 사용하는 것이 가능하다는 특징을 갖고 있다. 제안하는 방법의 성능을 확인하기 위하여 간단한 패턴 인식 문제들에의 적용 실험 및 기존의 학습 속도 향상 방법을 함께 사용하여 시계열 예측 문제에 적용한 실험을 수행하였고, 그 결과로부터 제안안 방법의 효율성을 확인할 수 있었다. Abstract A multilayer perceptron is the most well-known neural network model which has been successfully applied to various fields of application. Its slow learning caused by plateau and local minima of gradient descent learning, however, have been pointed as the biggest problems in its practical use. To solve such a problem, a number of researches on learning algorithms have been conducted, but it can be said that none of satisfying solutions have been presented so far because the problems such as computational inefficiency have still been existed in these algorithms. In this paper, we propose a new learning approach to minimize the effect of plateau and reduce the possibility of getting trapped in local minima by generalizing the sigmoidal function which is used as the basis function of a multilayer perceptron. Adapting a new approach that differs from the conventional methods with revised updating equation, the proposed method can be used together with the existing methods to improve the learning performance. We conducted some experiments to test the proposed method on simple problems of pattern recognition and a problem of time series prediction, compared our results with the results of the existing methods, and confirmed that the proposed method is efficient enough to apply to the real problems.

Numerical Study of SPGD-based Phase Control of Coherent Beam Combining under Various Turbulent Atmospheric Conditions (대기외란에 따른 SPGD 기반 결맞음 빔결합 시스템 위상제어 동작성능 분석)

  • Kim, Hansol;Na, Jeongkyun;Jeong, Yoonchan
    • Korean Journal of Optics and Photonics
    • /
    • v.31 no.6
    • /
    • pp.247-258
    • /
    • 2020
  • In this paper, based on a stochastic parallel gradient descent (SPGD) algorithm we study phase control of a coherent-beam-combining system under turbulent atmospheric conditions. Based on the statistical theory of atmospheric turbulence, we carry out the analysis of the phase and wavefront distortion of a laser beam propagating through a turbulent atmospheric medium. We also conduct numerical simulations of a coherent-beam-combining system with 7- and 19-channel laser beams distorted by atmospheric turbulence. Through numerical simulations, we characterize the phase-control characteristics and efficiency of the coherent-beam-combining system under various degrees of atmospheric turbulence. It is verified that the SPGD algorithm is capable of realizing 7-channel coherent beam combining with a beam-combining efficiency of more than 90%, even under the turbulent atmospheric conditions up to cn2 of 10-13 m-2/3. In the case of 19-channel coherent beam combining, it is shown that the same turbulent atmospheric conditions result in a drastic reduction of the beam-combining efficiency down to 60%, due to the elevated impact of the corresponding refractive-index inhomogeneity. In addition, by putting together the number of iterations of the SPGD algorithm required for phase locking under atmospheric turbulence and the time intervals of atmospheric phenomena, which typically are of the order of ㎲, it is estimated that hundreds of MHz to a few GHz of computing bandwidth of SPGD-based phase control may be required for a coherent-beam-combining system to confront such turbulent atmospheric conditions. We expect the results of this paper to be useful for quantitatively analyzing and predicting the effects of atmospheric turbulence on the SPGD-based phase-control performance of a coherent-beam-combining system.

Hierarchical Particle Swarm Optimization for Multi UAV Waypoints Planning Under Various Threats (다양한 위협 하에서 복수 무인기의 경로점 계획을 위한 계층적 입자 군집 최적화)

  • Chung, Wonmo;Kim, Myunggun;Lee, Sanha;Lee, Sang-Pill;Park, Chun-Shin;Son, Hungsun
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.50 no.6
    • /
    • pp.385-391
    • /
    • 2022
  • This paper presents to develop a path planning algorithm combining gradient descent-based path planning (GBPP) and particle swarm optimization (PSO) for considering prohibited flight areas, terrain information, and characteristics of fixed-wing unmmaned aerial vehicle (UAV) in 3D space. Path can be generated fast using GBPP, but it is often happened that an unsafe path can be generated by converging to a local minimum depending on the initial path. Bio-inspired swarm intelligence algorithms, such as Genetic algorithm (GA) and PSO, can avoid the local minima problem by sampling several paths. However, if the number of optimal variable increases due to an increase in the number of UAVs and waypoints, it requires heavy computation time and efforts due to increasing the number of particles accordingly. To solve the disadvantages of the two algorithms, hierarchical path planning algorithm associated with hierarchical particle swarm optimization (HPSO) is developed by defining the initial path, which is the input of GBPP, as two variables including particles variables. Feasibility of the proposed algorithm is verified by software-in-the-loop simulation (SILS) of flight control computer (FCC) for UAVs.

Development of new artificial neural network optimizer to improve water quality index prediction performance (수질 지수 예측성능 향상을 위한 새로운 인공신경망 옵티마이저의 개발)

  • Ryu, Yong Min;Kim, Young Nam;Lee, Dae Won;Lee, Eui Hoon
    • Journal of Korea Water Resources Association
    • /
    • v.57 no.2
    • /
    • pp.73-85
    • /
    • 2024
  • Predicting water quality of rivers and reservoirs is necessary for the management of water resources. Artificial Neural Networks (ANNs) have been used in many studies to predict water quality with high accuracy. Previous studies have used Gradient Descent (GD)-based optimizers as an optimizer, an operator of ANN that searches parameters. However, GD-based optimizers have the disadvantages of the possibility of local optimal convergence and absence of a solution storage and comparison structure. This study developed improved optimizers to overcome the disadvantages of GD-based optimizers. Proposed optimizers are optimizers that combine adaptive moments (Adam) and Nesterov-accelerated adaptive moments (Nadam), which have low learning errors among GD-based optimizers, with Harmony Search (HS) or Novel Self-adaptive Harmony Search (NSHS). To evaluate the performance of Long Short-Term Memory (LSTM) using improved optimizers, the water quality data from the Dasan water quality monitoring station were used for training and prediction. Comparing the learning results, Mean Squared Error (MSE) of LSTM using Nadam combined with NSHS (NadamNSHS) was the lowest at 0.002921. In addition, the prediction rankings according to MSE and R2 for the four water quality indices for each optimizer were compared. Comparing the average of ranking for each optimizer, it was confirmed that LSTM using NadamNSHS was the highest at 2.25.

The Pattern Analysis of Financial Distress for Non-audited Firms using Data Mining (데이터마이닝 기법을 활용한 비외감기업의 부실화 유형 분석)

  • Lee, Su Hyun;Park, Jung Min;Lee, Hyoung Yong
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.4
    • /
    • pp.111-131
    • /
    • 2015
  • There are only a handful number of research conducted on pattern analysis of corporate distress as compared with research for bankruptcy prediction. The few that exists mainly focus on audited firms because financial data collection is easier for these firms. But in reality, corporate financial distress is a far more common and critical phenomenon for non-audited firms which are mainly comprised of small and medium sized firms. The purpose of this paper is to classify non-audited firms under distress according to their financial ratio using data mining; Self-Organizing Map (SOM). SOM is a type of artificial neural network that is trained using unsupervised learning to produce a lower dimensional discretized representation of the input space of the training samples, called a map. SOM is different from other artificial neural networks as it applies competitive learning as opposed to error-correction learning such as backpropagation with gradient descent, and in the sense that it uses a neighborhood function to preserve the topological properties of the input space. It is one of the popular and successful clustering algorithm. In this study, we classify types of financial distress firms, specially, non-audited firms. In the empirical test, we collect 10 financial ratios of 100 non-audited firms under distress in 2004 for the previous two years (2002 and 2003). Using these financial ratios and the SOM algorithm, five distinct patterns were distinguished. In pattern 1, financial distress was very serious in almost all financial ratios. 12% of the firms are included in these patterns. In pattern 2, financial distress was weak in almost financial ratios. 14% of the firms are included in pattern 2. In pattern 3, growth ratio was the worst among all patterns. It is speculated that the firms of this pattern may be under distress due to severe competition in their industries. Approximately 30% of the firms fell into this group. In pattern 4, the growth ratio was higher than any other pattern but the cash ratio and profitability ratio were not at the level of the growth ratio. It is concluded that the firms of this pattern were under distress in pursuit of expanding their business. About 25% of the firms were in this pattern. Last, pattern 5 encompassed very solvent firms. Perhaps firms of this pattern were distressed due to a bad short-term strategic decision or due to problems with the enterpriser of the firms. Approximately 18% of the firms were under this pattern. This study has the academic and empirical contribution. In the perspectives of the academic contribution, non-audited companies that tend to be easily bankrupt and have the unstructured or easily manipulated financial data are classified by the data mining technology (Self-Organizing Map) rather than big sized audited firms that have the well prepared and reliable financial data. In the perspectives of the empirical one, even though the financial data of the non-audited firms are conducted to analyze, it is useful for find out the first order symptom of financial distress, which makes us to forecast the prediction of bankruptcy of the firms and to manage the early warning and alert signal. These are the academic and empirical contribution of this study. The limitation of this research is to analyze only 100 corporates due to the difficulty of collecting the financial data of the non-audited firms, which make us to be hard to proceed to the analysis by the category or size difference. Also, non-financial qualitative data is crucial for the analysis of bankruptcy. Thus, the non-financial qualitative factor is taken into account for the next study. This study sheds some light on the non-audited small and medium sized firms' distress prediction in the future.