• 제목/요약/키워드: Algorithm optimization

검색결과 5,708건 처리시간 0.037초

Application of the optimal fuzzy-based system on bearing capacity of concrete pile

  • Kun Zhang;Yonghua Zhang;Behnaz Razzaghzadeh
    • Steel and Composite Structures
    • /
    • 제51권1호
    • /
    • pp.25-41
    • /
    • 2024
  • The measurement of pile bearing capacity is crucial for the design of pile foundations, where in-situ tests could be costly and time needed. The primary objective of this research was to investigate the potential use of fuzzy-based techniques to anticipate the maximum weight that concrete driven piles might bear. Despite the existence of several suggested designs, there is a scarcity of specialized studies on the exploration of adaptive neuro-fuzzy inference systems (ANFIS) for the estimation of pile bearing capacity. This paper presents the introduction and validation of a novel technique that integrates the fire hawk optimizer (FHO) and equilibrium optimizer (EO) with the ANFIS, referred to as ANFISFHO and ANFISEO, respectively. A comprehensive compilation of 472 static load test results for driven piles was located within the database. The recommended framework was built, validated, and tested using the training set (70%), validation set (15%), and testing set (15%) of the dataset, accordingly. Moreover, the sensitivity analysis is performed in order to determine the impact of each input on the output. The results show that ANFISFHO and ANFISEO both have amazing potential for precisely calculating pile bearing capacity. The R2 values obtained for ANFISFHO were 0.9817, 0.9753, and 0.9823 for the training, validating, and testing phases. The findings of the examination of uncertainty showed that the ANFISFHO system had less uncertainty than the ANFISEO model. The research found that the ANFISFHO model provides a more satisfactory estimation of the bearing capacity of concrete driven piles when considering various performance evaluations and comparing it with existing literature.

Image-Based Machine Learning Model for Malware Detection on LLVM IR (LLVM IR 대상 악성코드 탐지를 위한 이미지 기반 머신러닝 모델)

  • Kyung-bin Park;Yo-seob Yoon;Baasantogtokh Duulga;Kang-bin Yim
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • 제34권1호
    • /
    • pp.31-40
    • /
    • 2024
  • Recently, static analysis-based signature and pattern detection technologies have limitations due to the advanced IT technologies. Moreover, It is a compatibility problem of multiple architectures and an inherent problem of signature and pattern detection. Malicious codes use obfuscation and packing techniques to hide their identity, and they also avoid existing static analysis-based signature and pattern detection techniques such as code rearrangement, register modification, and branching statement addition. In this paper, We propose an LLVM IR image-based automated static analysis of malicious code technology using machine learning to solve the problems mentioned above. Whether binary is obfuscated or packed, it's decompiled into LLVM IR, which is an intermediate representation dedicated to static analysis and optimization. "Therefore, the LLVM IR code is converted into an image before being fed to the CNN-based transfer learning algorithm ResNet50v2 supported by Keras". As a result, we present a model for image-based detection of malicious code.

Development of the Performance-Based Bridge Maintenance System to Generate Optimum Maintenance Strategy Considering Life-Cycle Cost (생애주기비용을 고려한 성능기반 교량 최적 유지관리 전략 수립 시스템 개발)

  • Park, Kyung-Hoon;Lee, Sang-Yoon;Hwang, Yoon-Koog;Kong, Jung-Sik;Lim, Jong-Kwon
    • Journal of the Korea institute for structural maintenance and inspection
    • /
    • 제11권4호
    • /
    • pp.109-120
    • /
    • 2007
  • In this study, a bridge maintenance system is developed to generate performance-based optimum maintenance strategy by considering the life-cycle cost. A multi-objective combinatorial optimization problem is formulated to generate a tradeoff maintenance scenarios which satisfies the balance among the conflicting objectives such as the performance and cost during the bridge lifetime and a genetic algorithm is applied to the system. By using the developed program, this study proposes a process of optimum maintenance scenario applying to the steel girder bridge of national road. The developed system improves the current methods of establishing the bridge maintenance strategy and can be utilized as an efficient tool to provide the optimum bridge maintenance scenario corresponding to the various constraints and requirements of bridge agency.

Nonlinear intelligent control systems subjected to earthquakes by fuzzy tracking theory

  • Z.Y. Chen;Y.M. Meng;Ruei-Yuan Wang;Timothy Chen
    • Smart Structures and Systems
    • /
    • 제33권4호
    • /
    • pp.291-300
    • /
    • 2024
  • Uncertainty of the model, system delay and drive dynamics can be considered as normal uncertainties, and the main source of uncertainty in the seismic control system is related to the nature of the simulated seismic error. In this case, optimizing the management strategy for one particular seismic record will not yield the best results for another. In this article, we propose a framework for online management of active structural management systems with seismic uncertainty. For this purpose, the concept of reinforcement learning is used for online optimization of active crowd management software. The controller consists of a differential controller, an unplanned gain ratio, the gain of which is enhanced using an online reinforcement learning algorithm. In addition, the proposed controller includes a dynamic status forecaster to solve the delay problem. To evaluate the performance of the proposed controllers, thousands of ground motion data sets were processed and grouped according to their spectrum using fuzzy clustering techniques with spatial hazard estimation. Finally, the controller is implemented in a laboratory scale configuration and its operation is simulated on a vibration table using cluster location and some actual seismic data. The test results show that the proposed controller effectively withstands strong seismic interference with delay. The goals of this paper are towards access to adequate, safe and affordable housing and basic services, promotion of inclusive and sustainable urbanization and participation, implementation of sustainable and disaster-resilient buildings, sustainable human settlement planning and manage. Simulation results is believed to achieved in the near future by the ongoing development of AI and control theory.

Life prediction of IGBT module for nuclear power plant rod position indicating and rod control system based on SDAE-LSTM

  • Zhi Chen;Miaoxin Dai;Jie Liu;Wei Jiang;Yuan Min
    • Nuclear Engineering and Technology
    • /
    • 제56권9호
    • /
    • pp.3740-3749
    • /
    • 2024
  • To reduce the losses caused by aging failure of insulation gate bipolar transistor (IGBT), which is the core components of nuclear power plant rod position indicating and rod control (RPC) system. It is necessary to conduct studies on its life prediction. The selection of IGBT failure characteristic parameters in existing research relies heavily on failure principles and expert experience. Moreover, the analysis and learning of time-domain degradation data have not been fully conducted, resulting in low prediction efficiency as the monotonicity, time correlation, and poor anti-interference ability of extracted degradation features. This paper utilizes the advantages of the stacked denoising autoencoder(SDAE) network in adaptive feature extraction and denoising capabilities to perform adaptive feature extraction on IGBT time-domain degradation data; establishes a long-short-term memory (LSTM) prediction model, and optimizes the learning rate, number of nodes in the hidden layer, and number of hidden layers using the Gray Wolf Optimization (GWO) algorithm; conducts verification experiments on the IGBT accelerated aging dataset provided by NASA PCoE Research Center, and selects performance evaluation indicators to compare and analyze the prediction results of the SDAE-LSTM model, PSOLSTM model, and BP model. The results show that the SDAE-LSTM model can achieve more accurate and stable IGBT life prediction.

Steel Plate Faults Diagnosis with S-MTS (S-MTS를 이용한 강판의 표면 결함 진단)

  • Kim, Joon-Young;Cha, Jae-Min;Shin, Junguk;Yeom, Choongsub
    • Journal of Intelligence and Information Systems
    • /
    • 제23권1호
    • /
    • pp.47-67
    • /
    • 2017
  • Steel plate faults is one of important factors to affect the quality and price of the steel plates. So far many steelmakers generally have used visual inspection method that could be based on an inspector's intuition or experience. Specifically, the inspector checks the steel plate faults by looking the surface of the steel plates. However, the accuracy of this method is critically low that it can cause errors above 30% in judgment. Therefore, accurate steel plate faults diagnosis system has been continuously required in the industry. In order to meet the needs, this study proposed a new steel plate faults diagnosis system using Simultaneous MTS (S-MTS), which is an advanced Mahalanobis Taguchi System (MTS) algorithm, to classify various surface defects of the steel plates. MTS has generally been used to solve binary classification problems in various fields, but MTS was not used for multiclass classification due to its low accuracy. The reason is that only one mahalanobis space is established in the MTS. In contrast, S-MTS is suitable for multi-class classification. That is, S-MTS establishes individual mahalanobis space for each class. 'Simultaneous' implies comparing mahalanobis distances at the same time. The proposed steel plate faults diagnosis system was developed in four main stages. In the first stage, after various reference groups and related variables are defined, data of the steel plate faults is collected and used to establish the individual mahalanobis space per the reference groups and construct the full measurement scale. In the second stage, the mahalanobis distances of test groups is calculated based on the established mahalanobis spaces of the reference groups. Then, appropriateness of the spaces is verified by examining the separability of the mahalanobis diatances. In the third stage, orthogonal arrays and Signal-to-Noise (SN) ratio of dynamic type are applied for variable optimization. Also, Overall SN ratio gain is derived from the SN ratio and SN ratio gain. If the derived overall SN ratio gain is negative, it means that the variable should be removed. However, the variable with the positive gain may be considered as worth keeping. Finally, in the fourth stage, the measurement scale that is composed of selected useful variables is reconstructed. Next, an experimental test should be implemented to verify the ability of multi-class classification and thus the accuracy of the classification is acquired. If the accuracy is acceptable, this diagnosis system can be used for future applications. Also, this study compared the accuracy of the proposed steel plate faults diagnosis system with that of other popular classification algorithms including Decision Tree, Multi Perception Neural Network (MLPNN), Logistic Regression (LR), Support Vector Machine (SVM), Tree Bagger Random Forest, Grid Search (GS), Genetic Algorithm (GA) and Particle Swarm Optimization (PSO). The steel plates faults dataset used in the study is taken from the University of California at Irvine (UCI) machine learning repository. As a result, the proposed steel plate faults diagnosis system based on S-MTS shows 90.79% of classification accuracy. The accuracy of the proposed diagnosis system is 6-27% higher than MLPNN, LR, GS, GA and PSO. Based on the fact that the accuracy of commercial systems is only about 75-80%, it means that the proposed system has enough classification performance to be applied in the industry. In addition, the proposed system can reduce the number of measurement sensors that are installed in the fields because of variable optimization process. These results show that the proposed system not only can have a good ability on the steel plate faults diagnosis but also reduce operation and maintenance cost. For our future work, it will be applied in the fields to validate actual effectiveness of the proposed system and plan to improve the accuracy based on the results.

Direct Reconstruction of Displaced Subdivision Mesh from Unorganized 3D Points (연결정보가 없는 3차원 점으로부터 차이분할메쉬 직접 복원)

  • Jung, Won-Ki;Kim, Chang-Heon
    • Journal of KIISE:Computer Systems and Theory
    • /
    • 제29권6호
    • /
    • pp.307-317
    • /
    • 2002
  • In this paper we propose a new mesh reconstruction scheme that produces a displaced subdivision surface directly from unorganized points. The displaced subdivision surface is a new mesh representation that defines a detailed mesh with a displacement map over a smooth domain surface, but original displaced subdivision surface algorithm needs an explicit polygonal mesh since it is not a mesh reconstruction algorithm but a mesh conversion (remeshing) algorithm. The main idea of our approach is that we sample surface detail from unorganized points without any topological information. For this, we predict a virtual triangular face from unorganized points for each sampling ray from a parameteric domain surface. Direct displaced subdivision surface reconstruction from unorganized points has much importance since the output of this algorithm has several important properties: It has compact mesh representation since most vertices can be represented by only a scalar value. Underlying structure of it is piecewise regular so it ran be easily transformed into a multiresolution mesh. Smoothness after mesh deformation is automatically preserved. We avoid time-consuming global energy optimization by employing the input data dependant mesh smoothing, so we can get a good quality displaced subdivision surface quickly.

Development of a Stock Trading System Using M & W Wave Patterns and Genetic Algorithms (M&W 파동 패턴과 유전자 알고리즘을 이용한 주식 매매 시스템 개발)

  • Yang, Hoonseok;Kim, Sunwoong;Choi, Heung Sik
    • Journal of Intelligence and Information Systems
    • /
    • 제25권1호
    • /
    • pp.63-83
    • /
    • 2019
  • Investors prefer to look for trading points based on the graph shown in the chart rather than complex analysis, such as corporate intrinsic value analysis and technical auxiliary index analysis. However, the pattern analysis technique is difficult and computerized less than the needs of users. In recent years, there have been many cases of studying stock price patterns using various machine learning techniques including neural networks in the field of artificial intelligence(AI). In particular, the development of IT technology has made it easier to analyze a huge number of chart data to find patterns that can predict stock prices. Although short-term forecasting power of prices has increased in terms of performance so far, long-term forecasting power is limited and is used in short-term trading rather than long-term investment. Other studies have focused on mechanically and accurately identifying patterns that were not recognized by past technology, but it can be vulnerable in practical areas because it is a separate matter whether the patterns found are suitable for trading. When they find a meaningful pattern, they find a point that matches the pattern. They then measure their performance after n days, assuming that they have bought at that point in time. Since this approach is to calculate virtual revenues, there can be many disparities with reality. The existing research method tries to find a pattern with stock price prediction power, but this study proposes to define the patterns first and to trade when the pattern with high success probability appears. The M & W wave pattern published by Merrill(1980) is simple because we can distinguish it by five turning points. Despite the report that some patterns have price predictability, there were no performance reports used in the actual market. The simplicity of a pattern consisting of five turning points has the advantage of reducing the cost of increasing pattern recognition accuracy. In this study, 16 patterns of up conversion and 16 patterns of down conversion are reclassified into ten groups so that they can be easily implemented by the system. Only one pattern with high success rate per group is selected for trading. Patterns that had a high probability of success in the past are likely to succeed in the future. So we trade when such a pattern occurs. It is a real situation because it is measured assuming that both the buy and sell have been executed. We tested three ways to calculate the turning point. The first method, the minimum change rate zig-zag method, removes price movements below a certain percentage and calculates the vertex. In the second method, high-low line zig-zag, the high price that meets the n-day high price line is calculated at the peak price, and the low price that meets the n-day low price line is calculated at the valley price. In the third method, the swing wave method, the high price in the center higher than n high prices on the left and right is calculated as the peak price. If the central low price is lower than the n low price on the left and right, it is calculated as valley price. The swing wave method was superior to the other methods in the test results. It is interpreted that the transaction after checking the completion of the pattern is more effective than the transaction in the unfinished state of the pattern. Genetic algorithms(GA) were the most suitable solution, although it was virtually impossible to find patterns with high success rates because the number of cases was too large in this simulation. We also performed the simulation using the Walk-forward Analysis(WFA) method, which tests the test section and the application section separately. So we were able to respond appropriately to market changes. In this study, we optimize the stock portfolio because there is a risk of over-optimized if we implement the variable optimality for each individual stock. Therefore, we selected the number of constituent stocks as 20 to increase the effect of diversified investment while avoiding optimization. We tested the KOSPI market by dividing it into six categories. In the results, the portfolio of small cap stock was the most successful and the high vol stock portfolio was the second best. This shows that patterns need to have some price volatility in order for patterns to be shaped, but volatility is not the best.

Dose Verification Using Pelvic Phantom in High Dose Rate (HDR) Brachytherapy (자궁경부암용 팬톰을 이용한 HDR (High dose rate) 근접치료의 선량 평가)

  • 장지나;허순녕;김회남;윤세철;최보영;이형구;서태석
    • Progress in Medical Physics
    • /
    • 제14권1호
    • /
    • pp.15-19
    • /
    • 2003
  • High dose rate (HDR) brachytherapy for treating a cervix carcinoma has become popular, because it eliminates many of the problems associated with conventional brachytherapy. In order to improve the clinical effectiveness with HDR brachytherapy, a dose calculation algorithm, optimization procedures, and image registrations need to be verified by comparing the dose distributions from a planning computer and those from a phantom. In this study, the phantom was fabricated in order to verify the absolute doses and the relative dose distributions. The measured doses from the phantom were then compared with the treatment planning system for the dose verification. The phantom needs to be designed such that the dose distributions can be quantitatively evaluated by utilizing the dosimeters with a high spatial resolution. Therefore, the small size of the thermoluminescent dosimeter (TLD) chips with a dimension of <1/8"and film dosimetry with a spatial resolution of <1mm used to measure the radiation dosages in the phantom. The phantom called a pelvic phantom was made from water and the tissue-equivalent acrylic plates. In order to firmly hold the HDR applicators in the water phantom, the applicators were inserted into the grooves of the applicator holder. The dose distributions around the applicators, such as Point A and B, were measured by placing a series of TLD chips (TLD-to-TLD distance: 5mm) in the three TLD holders, and placing three verification films in the orthogonal planes. This study used a Nucletron Plato treatment planning system and a Microselectron Ir-192 source unit. The results showed good agreement between the treatment plan and measurement. The comparisons of the absolute dose showed agreement within $\pm$4.0 % of the dose at point A and B, and the bladder and rectum point. In addition, the relative dose distributions by film dosimetry and those calculated by the planning computer show good agreement. This pelvic phantom could be a useful to verify the dose calculation algorithm and the accuracy of the image localization algorithm in the high dose rate (HDR) planning computer. The dose verification with film dosimetry and TLD as quality assurance (QA) tools are currently being undertaken in the Catholic University, Seoul, Korea.

  • PDF

Video Analysis System for Action and Emotion Detection by Object with Hierarchical Clustering based Re-ID (계층적 군집화 기반 Re-ID를 활용한 객체별 행동 및 표정 검출용 영상 분석 시스템)

  • Lee, Sang-Hyun;Yang, Seong-Hun;Oh, Seung-Jin;Kang, Jinbeom
    • Journal of Intelligence and Information Systems
    • /
    • 제28권1호
    • /
    • pp.89-106
    • /
    • 2022
  • Recently, the amount of video data collected from smartphones, CCTVs, black boxes, and high-definition cameras has increased rapidly. According to the increasing video data, the requirements for analysis and utilization are increasing. Due to the lack of skilled manpower to analyze videos in many industries, machine learning and artificial intelligence are actively used to assist manpower. In this situation, the demand for various computer vision technologies such as object detection and tracking, action detection, emotion detection, and Re-ID also increased rapidly. However, the object detection and tracking technology has many difficulties that degrade performance, such as re-appearance after the object's departure from the video recording location, and occlusion. Accordingly, action and emotion detection models based on object detection and tracking models also have difficulties in extracting data for each object. In addition, deep learning architectures consist of various models suffer from performance degradation due to bottlenects and lack of optimization. In this study, we propose an video analysis system consists of YOLOv5 based DeepSORT object tracking model, SlowFast based action recognition model, Torchreid based Re-ID model, and AWS Rekognition which is emotion recognition service. Proposed model uses single-linkage hierarchical clustering based Re-ID and some processing method which maximize hardware throughput. It has higher accuracy than the performance of the re-identification model using simple metrics, near real-time processing performance, and prevents tracking failure due to object departure and re-emergence, occlusion, etc. By continuously linking the action and facial emotion detection results of each object to the same object, it is possible to efficiently analyze videos. The re-identification model extracts a feature vector from the bounding box of object image detected by the object tracking model for each frame, and applies the single-linkage hierarchical clustering from the past frame using the extracted feature vectors to identify the same object that failed to track. Through the above process, it is possible to re-track the same object that has failed to tracking in the case of re-appearance or occlusion after leaving the video location. As a result, action and facial emotion detection results of the newly recognized object due to the tracking fails can be linked to those of the object that appeared in the past. On the other hand, as a way to improve processing performance, we introduce Bounding Box Queue by Object and Feature Queue method that can reduce RAM memory requirements while maximizing GPU memory throughput. Also we introduce the IoF(Intersection over Face) algorithm that allows facial emotion recognized through AWS Rekognition to be linked with object tracking information. The academic significance of this study is that the two-stage re-identification model can have real-time performance even in a high-cost environment that performs action and facial emotion detection according to processing techniques without reducing the accuracy by using simple metrics to achieve real-time performance. The practical implication of this study is that in various industrial fields that require action and facial emotion detection but have many difficulties due to the fails in object tracking can analyze videos effectively through proposed model. Proposed model which has high accuracy of retrace and processing performance can be used in various fields such as intelligent monitoring, observation services and behavioral or psychological analysis services where the integration of tracking information and extracted metadata creates greate industrial and business value. In the future, in order to measure the object tracking performance more precisely, there is a need to conduct an experiment using the MOT Challenge dataset, which is data used by many international conferences. We will investigate the problem that the IoF algorithm cannot solve to develop an additional complementary algorithm. In addition, we plan to conduct additional research to apply this model to various fields' dataset related to intelligent video analysis.