• Title/Summary/Keyword: 의사결정 알고리즘

Search Result 583, Processing Time 0.028 seconds

A Multi-agent System based on Genetic Algorithm for Integration Planning in a Supply Chain Management (유전 알고리즘에 기반한 동적 공급사슬 통합계획을 위한 멀티 에이전트 시스템)

  • Park, Byung-Joo;Choi, Hyung-Rim;Kang, Moo-Hong
    • Journal of Intelligence and Information Systems
    • /
    • v.13 no.3
    • /
    • pp.47-61
    • /
    • 2007
  • In SCM (supply chain management), companies are pursuing a new approach through which overall functions within the supply chain, ranging from material purchase to production, distribution, and sales are designed, planned, and managed in an integrated way. The core functions among them are production planning and distribution planning. As these problems are mutually related, they should be dealt with simultaneously in an integrated manner. SCM is large-scale and multi-stage problems. Also, its various kinds of internal or external factors can, at any time, dynamically bring a change to the existing plan or situation. Recently, many enterprises are moving toward an open architecture for integrating their activities with their suppliers, customers and other partners within the supply chain. Agent-based technology provides an effective approach in such environments. Multi-agent systems have been proven suitable to represent domains such as supply chain networks which involve interactions among manufacturing organization, their customers, suppliers, etc. with different individual goals and propriety information. In this paper, we propose a multi-agent system based on the genetic algorithm that make it possible to integrate the production and distribution planning on a real-time basis in SCM. The proposed genetic algorithm produced near optimal solution and we checked that there is a great difference in the results between integrated planning and non-integrated planning.

  • PDF

Development of Vehicle Arrival Time Prediction Algorithm Based on a Demand Volume (교통수요 기반의 도착예정시간 산출 알고리즘 개발)

  • Kim, Ji-Hong;Lee, Gyeong-Sun;Kim, Yeong-Ho;Lee, Seong-Mo
    • Journal of Korean Society of Transportation
    • /
    • v.23 no.2
    • /
    • pp.107-116
    • /
    • 2005
  • The information on travel time in providing the information of traffic to drivers is one of the most important data to control a traffic congestion efficiently. Especially, this information is the major element of route choice of drivers, and based on the premise that it has the high degree of confidence in real situation. This study developed a vehicle arrival time prediction algorithm called as "VAT-DV" for 6 corridors in total 6.1Km of "Nam-san area trffic information system" in order to give an information of congestion to drivers using VMS, ARS, and WEB. The spatial scope of this study is 2.5km~3km sections of each corridor, but there are various situations of traffic flow in a short period because they have signalized intersections in a departure point and an arrival point of each corridor, so they have almost characteristics of interrupted and uninterrupted traffic flow. The algorithm uses the information on a demand volume and a queue length. The demand volume is estimated from density of each points based on the Greenburg model, and the queue length is from the density and speed of each point. In order to settle the variation of the unit time, the result of this algorithm is strategically regulated by importing the AVI(Automatic Vehicle Identification), one of the number plate matching methods. In this study, the AVI travel time information is composed by Hybrid Model in order to use it as the basic parameter to make one travel time in a day using ILD to classify the characteristics of the traffic flow along the queue length. According to the result of this study, in congestion situation, this algorithm has about more than 84% degree of accuracy. Specially, the result of providing the information of "Nam-san area traffic information system" shows that 72.6% of drivers are available.

The Design of Feature Selection Classifier based on Physiological Signal for Emotion Detection (감성판별을 위한 생체신호기반 특징선택 분류기 설계)

  • Lee, JeeEun;Yoo, Sun K.
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.11
    • /
    • pp.206-216
    • /
    • 2013
  • The emotion plays a critical role in human's daily life including learning, action, decision and communication. In this paper, emotion discrimination classifier is designed to reduce system complexity through reduced selection of dominant features from biosignals. The photoplethysmography(PPG), skin temperature, skin conductance, fontal and parietal electroencephalography(EEG) signals were measured during 4 types of movie watching associated with the induction of neutral, sad, fear joy emotions. The genetic algorithm with support vector machine(SVM) based fitness function was designed to determine dominant features among 24 parameters extracted from measured biosignals. It shows maximum classification accuracy of 96.4%, which is 17% higher than that of SVM alone. The minimum error features selected are the mean and NN50 of heart rate variability from PPG signal, the mean of PPG induced pulse transit time, the mean of skin resistance, and ${\delta}$ and ${\beta}$ frequency band powers of parietal EEG. The combination of parietal EEG, PPG, and skin resistance is recommendable in high accuracy instrumentation, while the combinational use of PPG and skin conductance(79% accuracy) is affordable in simplified instrumentation.

Committee Learning Classifier based on Attribute Value Frequency (속성 값 빈도 기반의 전문가 다수결 분류기)

  • Lee, Chang-Hwan;Jung, In-Chul;Kwon, Young-S.
    • Journal of KIISE:Databases
    • /
    • v.37 no.4
    • /
    • pp.177-184
    • /
    • 2010
  • In these day, many data including sensor, delivery, credit and stock data are generated continuously in massive quantity. It is difficult to learn from these data because they are large in volume and changing fast in their concepts. To handle these problems, learning methods based in sliding window methods over time have been used. But these approaches have a problem of rebuilding models every time new data arrive, which requires a lot of time and cost. Therefore we need very simple incremental learning methods. Bayesian method is an example of these methods but it has a disadvantage which it requries the prior knowledge(probabiltiy) of data. In this study, we propose a learning method based on attribute values. In the proposed method, even though we don't know the prior knowledge(probability) of data, we can apply our new method to data. The main concept of this method is that each attribute value is regarded as an expert learner, summing up the expert learners lead to better results. Experimental results show our learning method learns from data very fast and performs well when compared to current learning methods(decision tree and bayesian).

Study on Detection for Cochlodinium polykrikoides Red Tide using the GOCI image and Machine Learning Technique (GOCI 영상과 기계학습 기법을 이용한 Cochlodinium polykrikoides 적조 탐지 기법 연구)

  • Unuzaya, Enkhjargal;Bak, Su-Ho;Hwang, Do-Hyun;Jeong, Min-Ji;Kim, Na-Kyeong;Yoon, Hong-Joo
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.15 no.6
    • /
    • pp.1089-1098
    • /
    • 2020
  • In this study, we propose a method to detect red tide Cochlodinium Polykrikoide using by machine learning and geostationary marine satellite images. To learn the machine learning model, GOCI Level 2 data were used, and the red tide location data of the National Fisheries Research and Development Institute was used. The machine learning model used logistic regression model, decision tree model, and random forest model. As a result of the performance evaluation, compared to the traditional GOCI image-based red tide detection algorithm without machine learning (Son et al., 2012) (75%), it was confirmed that the accuracy was improved by about 13~22%p (88~98%). In addition, as a result of comparing and analyzing the detection performance between machine learning models, the random forest model (98%) showed the highest detection accuracy.It is believed that this machine learning-based red tide detection algorithm can be used to detect red tide early in the future and track and monitor its movement and spread.

Prediction of Salinity of Nakdong River Estuary Using Deep Learning Algorithm (LSTM) for Time Series Analysis (시계열 분석 딥러닝 알고리즘을 적용한 낙동강 하굿둑 염분 예측)

  • Woo, Joung Woon;Kim, Yeon Joong;Yoon, Jong Sung
    • Journal of Korean Society of Coastal and Ocean Engineers
    • /
    • v.34 no.4
    • /
    • pp.128-134
    • /
    • 2022
  • Nakdong river estuary is being operated with the goal of expanding the period of seawater inflow from this year to 2022 every month and creating a brackish water area within 15 km of the upstream of the river bank. In this study, the deep learning algorithm Long Short-Term Memory (LSTM) was applied to predict the salinity of the Nakdong Bridge (about 5 km upstream of the river bank) for the purpose of rapid decision making for the target brackish water zone and prevention of salt water damage. Input data were constructed to reflect the temporal and spatial characteristics of the Nakdong River estuary, such as the amount of discharge from Changnyeong and Hamanbo, and an optimal model was constructed in consideration of the hydraulic characteristics of the Nakdong River Estuary by changing the degree according to the sequence length. For prediction accuracy, statistical analysis was performed using the coefficient of determination (R-squred) and RMSE (root mean square error). When the sequence length was 12, the R-squred 0.997 and RMSE 0.122 were the highest, and the prior prediction time showed a high degree of R-squred 0.93 or more until the 12-hour interval.

Establishing meteorological drought severity considering the level of emergency water supply (비상급수의 규모를 고려한 기상학적 가뭄 강도 수립)

  • Lee, Seungmin;Wang, Wonjoon;Kim, Donghyun;Han, Heechan;Kim, Soojun;Kim, Hung Soo
    • Journal of Korea Water Resources Association
    • /
    • v.56 no.10
    • /
    • pp.619-629
    • /
    • 2023
  • Recent intensification of climate change has led to an increase in damages caused by droughts. Currently, in Korea, the Standardized Precipitation Index (SPI) is used as a criterion to classify the intensity of droughts. Based on the accumulated precipitation over the past six months (SPI-6), meteorological drought intensities are classified into four categories: concern, caution, alert, and severe. However, there is a limitation in classifying drought intensity solely based on precipitation. To overcome the limitations of the meteorological drought warning criteria based on SPI, this study collected emergency water supply damage data from the National Drought Information Portal (NDIP) to classify drought intensity. Factors of SPI, such as precipitation, and factors used to calculate evapotranspiration, such as temperature and humidity, were indexed using min-max normalization. Coefficients for each factor were determined based on the Genetic Algorithm (GA). The drought intensity based on emergency water supply was used as the dependent variable, and the coefficients of each meteorological factor determined by GA were used as coefficients to derive a new Drought Severity Classification Index (DSCI). After deriving the DSCI, cumulative distribution functions were used to present intensity stage classification boundaries. It is anticipated that using the proposed DSCI in this study will allow for more accurate drought intensity classification than the traditional SPI, supporting decision-making for disaster management personnel.

Game Theory Based Co-Evolutionary Algorithm (GCEA) (게임 이론에 기반한 공진화 알고리즘)

  • Sim, Kwee-Bo;Kim, Ji-Youn;Lee, Dong-Wook
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.14 no.3
    • /
    • pp.253-261
    • /
    • 2004
  • Game theory is mathematical analysis developed to study involved in making decisions. In 1928, Von Neumann proved that every two-person, zero-sum game with finitely many pure strategies for each player is deterministic. As well, in the early 50's, Nash presented another concept as the basis for a generalization of Von Neumann's theorem. Another central achievement of game theory is the introduction of evolutionary game theory, by which agents can play optimal strategies in the absence of rationality. Not the rationality but through the process of Darwinian selection, a population of agents can evolve to an Evolutionary Stable Strategy (ESS) introduced by Maynard Smith. Keeping pace with these game theoretical studies, the first computer simulation of co-evolution was tried out by Hillis in 1991. Moreover, Kauffman proposed NK model to analyze co-evolutionary dynamics between different species. He showed how co-evolutionary phenomenon reaches static states and that these states are Nash equilibrium or ESS introduced in game theory. Since the studies about co-evolutionary phenomenon were started, however many other researchers have developed co-evolutionary algorithms, in this paper we propose Game theory based Co-Evolutionary Algorithm (GCEA) and confirm that this algorithm can be a solution of evolutionary problems by searching the ESS.To evaluate newly designed GCEA approach, we solve several test Multi-objective Optimization Problems (MOPs). From the results of these evaluations, we confirm that evolutionary game can be embodied by co-evolutionary algorithm and analyze optimization performance of GCEA by comparing experimental results using GCEA with the results using other evolutionary optimization algorithms.

Data-Driven Exploration for Transient Association Rules (한시적 연관규칙을 위한 데이타 주도 탐사 기법)

  • Cho, Ll-Rae;Kim, Jong-Deok;Lee, Do-Heon
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.4
    • /
    • pp.895-907
    • /
    • 1997
  • The mining of assciation rules disovers the trndency of events ocuring simultaneously in large databases. Previous announced research on association rules deals with associations with associations with respect to the whole transaction. However, xome association rules could have very high confidence in a sub-range of the time domain, even though they do not have quite high confidence in the whole time domain. Such kind of association rules are ecpected to be very usdful in various decion making problems.In this paper, we define transient association rule, as an association with high cimfidence worthy of special attention in a partial time interval, and propose an dfficeint algorithm wich finds out the time intervals appropriate to transient association rules from large-databases.We propose the data-driven retrival method excluding unecessary interval search, and design an effective data structure manageable in main memory obtined by one scanning of database, which offers the necessary information to next retrieval phase. In addition, our simulation shows that the suggested algorithm has reliable performance at the time cost acceptable in application areas.

  • PDF

A study on integrating and discovery of semantic based knowledge model (의미 기반의 지식모델 통합과 탐색에 관한 연구)

  • Chun, Seung-Su
    • Journal of Internet Computing and Services
    • /
    • v.15 no.6
    • /
    • pp.99-106
    • /
    • 2014
  • Generation and analysis methods have been proposed in recent years, such as using a natural language and formal language processing, artificial intelligence algorithms based knowledge model is effective meaning. its semantic based knowledge model has been used effective decision making tree and problem solving about specific context. and it was based on static generation and regression analysis, trend analysis with behavioral model, simulation support for macroeconomic forecasting mode on especially in a variety of complex systems and social network analysis. In this study, in this sense, integrating knowledge-based models, This paper propose a text mining derived from the inter-Topic model Integrated formal methods and Algorithms. First, a method for converting automatically knowledge map is derived from text mining keyword map and integrate it into the semantic knowledge model for this purpose. This paper propose an algorithm to derive a method of projecting a significant topic map from the map and the keyword semantically equivalent model. Integrated semantic-based knowledge model is available.