• Title/Summary/Keyword: Optimization algorithms

Search Result 1,716, Processing Time 0.036 seconds

Feature-selection algorithm based on genetic algorithms using unstructured data for attack mail identification (공격 메일 식별을 위한 비정형 데이터를 사용한 유전자 알고리즘 기반의 특징선택 알고리즘)

  • Hong, Sung-Sam;Kim, Dong-Wook;Han, Myung-Mook
    • Journal of Internet Computing and Services
    • /
    • v.20 no.1
    • /
    • pp.1-10
    • /
    • 2019
  • Since big-data text mining extracts many features and data, clustering and classification can result in high computational complexity and low reliability of the analysis results. In particular, a term document matrix obtained through text mining represents term-document features, but produces a sparse matrix. We designed an advanced genetic algorithm (GA) to extract features in text mining for detection model. Term frequency inverse document frequency (TF-IDF) is used to reflect the document-term relationships in feature extraction. Through a repetitive process, a predetermined number of features are selected. And, we used the sparsity score to improve the performance of detection model. If a spam mail data set has the high sparsity, detection model have low performance and is difficult to search the optimization detection model. In addition, we find a low sparsity model that have also high TF-IDF score by using s(F) where the numerator in fitness function. We also verified its performance by applying the proposed algorithm to text classification. As a result, we have found that our algorithm shows higher performance (speed and accuracy) in attack mail classification.

A Study on the Optimization Model for the Project Portfolio Manpower Assignment Using Genetic Algorithm (유전자 알고리즘을 이용한 프로젝트 포트폴리오 투입인력 최적화 모델에 관한 연구)

  • Kim, Dong-Wook;Lee, Won-Young
    • Journal of Information Technology Services
    • /
    • v.17 no.4
    • /
    • pp.101-117
    • /
    • 2018
  • Companies are responding appropriately to the rapidly changing business environment and striving to lead those changes. As part of that, we are meeting our strategic goals through IT projects, which increase the number of simultaneous projects and the importance of project portfolio management for successful project execution. It also strives for efficient deployment of human resources that have the greatest impact on project portfolio management. In the early stages of project portfolio management, it is very important to establish a reasonable manpower plan and allocate performance personnel. This problem is a problem that can not be solved by linear programming because it is calculated through the standard deviation of the input ratio of professional manpower considering the uniformity of load allocated to the input development manpower and the importance of each project. In this study, genetic algorithm, one of the heuristic methods, was applied to solve this problem. As the objective function, we used the proper input ratio of projects, the input rate of specialist manpower for important projects, and the equal load of workload by manpower. Constraints were not able to input duplicate manpower, Was used as a condition. We also developed a program for efficient application of genetic algorithms and confirmed the execution results. In addition, the parameters of the genetic algorithm were variously changed and repeated test results were selected through the independent sample t test to select optimal parameters, and the improvement effect of about 31.2% was confirmed.

Optimization of LEA Quantum Circuits to Apply Grover's Algorithm (그루버 알고리즘 적용을 위한 LEA 양자 회로 최적화)

  • Jang, Kyung Bae;Kim, Hyun Jun;Park, Jae Hoon;Song, Gyeung Ju;Seo, Hwa Jeong
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.10 no.4
    • /
    • pp.101-106
    • /
    • 2021
  • Quantum algorithms and quantum computers can break the security of many of the ciphers we currently use. If Grover's algorithm is applied to a symmetric key cipher with n-bit security level, the security level can be lowered to (n/2)-bit. In order to apply Grover's algorithm, it is most important to optimize the target cipher as a quantum circuit because the symmetric key cipher must be implemented as a quantum circuit in the oracle function. Accordingly, researches on implementing AES(Advanced Encryption Standard) or lightweight block ciphers as quantum circuits have been actively conducted in recent years. In this paper, korean lightweight block cipher LEA was optimized and implemented as a quantum circuit. Compared to the previous LEA quantum circuit implementation, quantum gates were used more, but qubits were drastically reduced, and performance evaluation was performed for this tradeoff problem. Finally, we evaluated quantum resources for applying Grover's algorithm to the proposed LEA implementation.

A novel radioactive particle tracking algorithm based on deep rectifier neural network

  • Dam, Roos Sophia de Freitas;dos Santos, Marcelo Carvalho;do Desterro, Filipe Santana Moreira;Salgado, William Luna;Schirru, Roberto;Salgado, Cesar Marques
    • Nuclear Engineering and Technology
    • /
    • v.53 no.7
    • /
    • pp.2334-2340
    • /
    • 2021
  • Radioactive particle tracking (RPT) is a minimally invasive nuclear technique that tracks a radioactive particle inside a volume of interest by means of a mathematical location algorithm. During the past decades, many algorithms have been developed including ones based on artificial intelligence techniques. In this study, RPT technique is applied in a simulated test section that employs a simplified mixer filled with concrete, six scintillator detectors and a137Cs radioactive particle emitting gamma rays of 662 keV. The test section was developed using MCNPX code, which is a mathematical code based on Monte Carlo simulation, and 3516 different radioactive particle positions (x,y,z) were simulated. Novelty of this paper is the use of a location algorithm based on a deep learning model, more specifically a 6-layers deep rectifier neural network (DRNN), in which hyperparameters were defined using a Bayesian optimization method. DRNN is a type of deep feedforward neural network that substitutes the usual sigmoid based activation functions, traditionally used in vanilla Multilayer Perceptron Networks, for rectified activation functions. Results show the great accuracy of the DRNN in a RPT tracking system. Root mean squared error for x, y and coordinates of the radioactive particle is, respectively, 0.03064, 0.02523 and 0.07653.

Investigation on the nonintrusive multi-fidelity reduced-order modeling for PWR rod bundles

  • Kang, Huilun;Tian, Zhaofei;Chen, Guangliang;Li, Lei;Chu, Tianhui
    • Nuclear Engineering and Technology
    • /
    • v.54 no.5
    • /
    • pp.1825-1834
    • /
    • 2022
  • Performing high-fidelity computational fluid dynamics (HF-CFD) to predict the flow and heat transfer state of the coolant in the reactor core is expensive, especially in scenarios that require extensive parameter search, such as uncertainty analysis and design optimization. This work investigated the performance of utilizing a multi-fidelity reduced-order model (MF-ROM) in PWR rod bundles simulation. Firstly, basis vectors and basis vector coefficients of high-fidelity and low-fidelity CFD results are extracted separately by the proper orthogonal decomposition (POD) approach. Secondly, a surrogate model is trained to map the relationship between the extracted coefficients from different fidelity results. In the prediction stage, the coefficients of the low-fidelity data under the new operating conditions are extracted by using the obtained POD basis vectors. Then, the trained surrogate model uses the low-fidelity coefficients to regress the high-fidelity coefficients. The predicted high-fidelity data is reconstructed from the product of extracted basis vectors and the regression coefficients. The effectiveness of the MF-ROM is evaluated on a flow and heat transfer problem in PWR fuel rod bundles. Two data-driven algorithms, the Kriging and artificial neural network (ANN), are trained as surrogate models for the MF-ROM to reconstruct the complex flow and heat transfer field downstream of the mixing vanes. The results show good agreements between the data reconstructed with the trained MF-ROM and the high-fidelity CFD simulation result, while the former only requires to taken the computational burden of low-fidelity simulation. The results also show that the performance of the ANN model is slightly better than the Kriging model when using a high number of POD basis vectors for regression. Moreover, the result presented in this paper demonstrates the suitability of the proposed MF-ROM for high-fidelity fixed value initialization to accelerate complex simulation.

STAR-24K: A Public Dataset for Space Common Target Detection

  • Zhang, Chaoyan;Guo, Baolong;Liao, Nannan;Zhong, Qiuyun;Liu, Hengyan;Li, Cheng;Gong, Jianglei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.2
    • /
    • pp.365-380
    • /
    • 2022
  • The target detection algorithm based on supervised learning is the current mainstream algorithm for target detection. A high-quality dataset is the prerequisite for the target detection algorithm to obtain good detection performance. The larger the number and quality of the dataset, the stronger the generalization ability of the model, that is, the dataset determines the upper limit of the model learning. The convolutional neural network optimizes the network parameters in a strong supervision method. The error is calculated by comparing the predicted frame with the manually labeled real frame, and then the error is passed into the network for continuous optimization. Strongly supervised learning mainly relies on a large number of images as models for continuous learning, so the number and quality of images directly affect the results of learning. This paper proposes a dataset STAR-24K (meaning a dataset for Space TArget Recognition with more than 24,000 images) for detecting common targets in space. Since there is currently no publicly available dataset for space target detection, we extracted some pictures from a series of channels such as pictures and videos released by the official websites of NASA (National Aeronautics and Space Administration) and ESA (The European Space Agency) and expanded them to 24,451 pictures. We evaluate popular object detection algorithms to build a benchmark. Our STAR-24K dataset is publicly available at https://github.com/Zzz-zcy/STAR-24K.

A Bi-objective Game-based Task Scheduling Method in Cloud Computing Environment

  • Guo, Wanwan;Zhao, Mengkai;Cui, Zhihua;Xie, Liping
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.11
    • /
    • pp.3565-3583
    • /
    • 2022
  • The task scheduling problem has received a lot of attention in recent years as a crucial area for research in the cloud environment. However, due to the difference in objectives considered by service providers and users, it has become a major challenge to resolve the conflicting interests of service providers and users while both can still take into account their respective objectives. Therefore, the task scheduling problem as a bi-objective game problem is formulated first, and then a task scheduling model based on the bi-objective game (TSBOG) is constructed. In this model, energy consumption and resource utilization, which are of concern to the service provider, and cost and task completion rate, which are of concern to the user, are calculated simultaneously. Furthermore, a many-objective evolutionary algorithm based on a partitioned collaborative selection strategy (MaOEA-PCS) has been developed to solve the TSBOG. The MaOEA-PCS can find a balance between population convergence and diversity by partitioning the objective space and selecting the best converging individuals from each region into the next generation. To balance the players' multiple objectives, a crossover and mutation operator based on dynamic games is proposed and applied to MaPEA-PCS as a player's strategy update mechanism. Finally, through a series of experiments, not only the effectiveness of the model compared to a normal many-objective model is demonstrated, but also the performance of MaOEA-PCS and the validity of DGame.

A Study on Classification Models for Predicting Bankruptcy Based on XAI (XAI 기반 기업부도예측 분류모델 연구)

  • Jihong Kim;Nammee Moon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.12 no.8
    • /
    • pp.333-340
    • /
    • 2023
  • Efficient prediction of corporate bankruptcy is an important part of making appropriate lending decisions for financial institutions and reducing loan default rates. In many studies, classification models using artificial intelligence technology have been used. In the financial industry, even if the performance of the new predictive models is excellent, it should be accompanied by an intuitive explanation of the basis on which the result was determined. Recently, the US, EU, and South Korea have commonly presented the right to request explanations of algorithms, so transparency in the use of AI in the financial sector must be secured. In this paper, an artificial intelligence-based interpretable classification prediction model was proposed using corporate bankruptcy data that was open to the outside world. First, data preprocessing, 5-fold cross-validation, etc. were performed, and classification performance was compared through optimization of 10 supervised learning classification models such as logistic regression, SVM, XGBoost, and LightGBM. As a result, LightGBM was confirmed as the best performance model, and SHAP, an explainable artificial intelligence technique, was applied to provide a post-explanation of the bankruptcy prediction process.

AI based complex sensor application study for energy management in WTP (정수장에서의 에너지 관리를 위한 AI 기반 복합센서 적용 연구)

  • Hong, Sung-Taek;An, Sang-Byung;Kim, Kuk-Il;Sung, Min-Seok
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.322-323
    • /
    • 2022
  • The most necessary thing for the optimal operation of a water purification plant is to accurately predict the pattern and amount of tap water used by consumers. The required amount of tap water should be delivered to the drain using a pump and stored, and the required flow rate should be supplied in a timely manner using the minimum amount of electrical energy. The short-term demand forecasting required from the point of view of energy optimization operation among water purification plant volume predictions has been made in consideration of seasons, major periods, and regional characteristics using time series analysis, regression analysis, and neural network algorithms. In this paper, we analyzed energy management methods through AI-based complex sensor applicability analysis such as LSTM (Long Short-Term Memory) and GRU (Gated Recurrent Units), which are types of cyclic neural networks.

  • PDF

Improvement in Inefficient Repetition of Gauss Sieve (Gauss Sieve 반복 동작에서의 비효율성 개선)

  • Byeongho Cheon;Changwon Lee;Chanho Jeon;Seokhie Hong;Suhri Kim
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.33 no.2
    • /
    • pp.223-233
    • /
    • 2023
  • Gauss Sieve is an algorithm for solving SVP and requires exponential time and space complexity. The terminationcondition of the Sieve is determined by the size of the constructed list and the number of collisions related to space complexity. The term 'collision' refers to the state in which the sampled vector is reduced to the vector that is already inthe list. if collisions occur more than a certain number of times, the algorithm terminates. When executing previous algorithms, we noticed that unnecessary operations continued even after the shortest vector was found. This means that the existing termination condition is set larger than necessary. In this paper, after identifying the point where unnecessary operations are repeated, optimization is performed on the number of operations required. The tests are conducted by adjusting the threshold of the collision that becomes the termination condition and the distribution in whichthe sample vector is generated. According to the experiments, the operation that occupies the largest proportion decreased by62.6%. The space and time complexity also decreased by 4.3 and 1.6%, respectively.