• Title/Summary/Keyword: Action Decision

Search Result 412, Processing Time 0.024 seconds

An Action Decision and Execution Method of Robotic Soccer System based on Neural Networks (신경회로망을 이용한 로봇축구 시스템의 행동결정 및 행동실행 방법)

  • Lee, Kyoung-Tae;Kim, Hak-Il;Kim, Choon-Woo
    • Proceedings of the KIEE Conference
    • /
    • 1998.11b
    • /
    • pp.543-545
    • /
    • 1998
  • Robotic soccer is multi-agent system playing soccer game under given rule. This system consists of three mobile robots, vision sensor, action decision module, action execution module and communication module. This paper presents new action decision method using multi-layer neural networks.

  • PDF

On the Optimal Selection of Smart Phone by Analytic Hierarchy Process (AHP를 이용한 스마트폰의 최적선정에 관한 연구)

  • Chung, Soon-Suk;Kim, Kwang-Soo
    • Proceedings of the Safety Management and Science Conference
    • /
    • 2010.11a
    • /
    • pp.199-207
    • /
    • 2010
  • Decision analysis has becomes an important technique for decision making in the face of uncertainty. It is characterized by enumerating all the available courses of action, identifying the payoffs for all possible outcomes, and quantifying the subjective probabilities for the all possible random events. When the data are available, decision analysis becomes a powerful tool for determining an optimal course of action. In this paper, we use the analytic hierarchy process in weights calculating. For the purpose of making optimal decision, the data of three different smart phones models are used.

  • PDF

On the Multi-attribute Decision Making by Entropy Methods (엔트로피 방법에 의한 다 요소 의사결정에 관한 연구)

  • 정순석
    • Journal of the Korea Safety Management & Science
    • /
    • v.6 no.2
    • /
    • pp.177-186
    • /
    • 2004
  • Decision analysis has becomes an important technique for decision making in the face of uncertainty. It is characterized by enumerating all the available courses of action, identifying the payoffs for all possible outcomes, and quantifying the subjective probabilities for the all possible random events. When the data are available, decision analysis becomes a powerful tool for determining an optimal course of action. We study the multi-attribute decision making in a compensatory models. In this paper, we use the entropy methods in weights calculating. For the purpose of making optimal decision, the data of five different car models are used. For computing, we used Visual Numerica Version 1.0 software package.

A Bayesian Decision Model for a Deteriorating Repairable System (열화시스템의 수리를 위한 베이지안 의사결정 모형의 개발)

  • Kim, Taeksang;Ahn, Suneung
    • Journal of Korean Institute of Industrial Engineers
    • /
    • v.32 no.2
    • /
    • pp.141-152
    • /
    • 2006
  • This paper presents the development of a decision model to examine the optimal repair action for a deteriorating system. In order to make a reasonable decision, it is necessary to perform an analysis of the uncertainties embedded in deterioration and to evaluate the repair actions based on the expected future cost. Focusing on the power law failure model, the uncertainties related to deterioration are analyzed based on the Bayesian approach. In addition, we develop a decision model for the optimal repair action by applying a repair cost function. A case study is given to illustrate a decision-making process by analyzing the loss incurred due to deterioration.

A Simulation Sample Accumulation Method for Efficient Simulation-based Policy Improvement in Markov Decision Process (마르코프 결정 과정에서 시뮬레이션 기반 정책 개선의 효율성 향상을 위한 시뮬레이션 샘플 누적 방법 연구)

  • Huang, Xi-Lang;Choi, Seon Han
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.7
    • /
    • pp.830-839
    • /
    • 2020
  • As a popular mathematical framework for modeling decision making, Markov decision process (MDP) has been widely used to solve problem in many engineering fields. MDP consists of a set of discrete states, a finite set of actions, and rewards received after reaching a new state by taking action from the previous state. The objective of MDP is to find an optimal policy, that is, to find the best action to be taken in each state to maximize the expected discounted reward of policy (EDR). In practice, MDP is typically unknown, so simulation-based policy improvement (SBPI), which improves a given base policy sequentially by selecting the best action in each state depending on rewards observed via simulation, can be a practical way to find the optimal policy. However, the efficiency of SBPI is still a concern since many simulation samples are required to precisely estimate EDR for each action in each state. In this paper, we propose a method to select the best action accurately in each state using a small number of simulation samples, thereby improving the efficiency of SBPI. The proposed method accumulates the simulation samples observed in the previous states, so it is possible to precisely estimate EDR even with a small number of samples in the current state. The results of comparative experiments on the existing method demonstrate that the proposed method can improve the efficiency of SBPI.

Partially Observable Markov Decision Processes (POMDPs) and Wireless Body Area Networks (WBAN): A Survey

  • Mohammed, Yahaya Onimisi;Baroudi, Uthman A.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.7 no.5
    • /
    • pp.1036-1057
    • /
    • 2013
  • Wireless body area network (WBAN) is a promising candidate for future health monitoring system. Nevertheless, the path to mature solutions is still facing a lot of challenges that need to be overcome. Energy efficient scheduling is one of these challenges given the scarcity of available energy of biosensors and the lack of portability. Therefore, researchers from academia, industry and health sectors are working together to realize practical solutions for these challenges. The main difficulty in WBAN is the uncertainty in the state of the monitored system. Intelligent learning approaches such as a Markov Decision Process (MDP) were proposed to tackle this issue. A Markov Decision Process (MDP) is a form of Markov Chain in which the transition matrix depends on the action taken by the decision maker (agent) at each time step. The agent receives a reward, which depends on the action and the state. The goal is to find a function, called a policy, which specifies which action to take in each state, so as to maximize some utility functions (e.g., the mean or expected discounted sum) of the sequence of rewards. A partially Observable Markov Decision Processes (POMDP) is a generalization of Markov decision processes that allows for the incomplete information regarding the state of the system. In this case, the state is not visible to the agent. This has many applications in operations research and artificial intelligence. Due to incomplete knowledge of the system, this uncertainty makes formulating and solving POMDP models mathematically complex and computationally expensive. Limited progress has been made in terms of applying POMPD to real applications. In this paper, we surveyed the existing methods and algorithms for solving POMDP in the general domain and in particular in Wireless body area network (WBAN). In addition, the papers discussed recent real implementation of POMDP on practical problems of WBAN. We believe that this work will provide valuable insights for the newcomers who would like to pursue related research in the domain of WBAN.

NPC Control Model for Defense in Soccer Game Applying the Decision Tree Learning Algorithm (결정트리 학습 알고리즘을 활용한 축구 게임 수비 NPC 제어 방법)

  • Cho, Dal-Ho;Lee, Yong-Ho;Kim, Jin-Hyung;Park, So-Young;Rhee, Dae-Woong
    • Journal of Korea Game Society
    • /
    • v.11 no.6
    • /
    • pp.61-70
    • /
    • 2011
  • In this paper, we propose a defense NPC control model in the soccer game by applying the Decision Tree learning algorithm. The proposed model extracts the direction patterns and the action patterns generated by many soccer game users, and applies these patterns to the Decision Tree learning algorithm. Then, the proposed model decides the direction and the action according to the learned Decision Tree. Experimental results show that the proposed model takes some time to learn the Decision Tree while the proposed model takes 0.001-0.003 milliseconds to decide the direction and the action based on the learned Decision Tree. Therefore, the proposed model can control NPC in the soccer game system in real time. Also, the proposed model achieves higher accuracy than a previous model (Letia98); because the proposed model can utilize current state information, its analyzed information, and previous state information.

Intelligent Fault Diagnosis System Using Hybrid Data Mining (하이브리드 데이터마이닝을 이용한 지능형 이상 진단 시스템)

  • Baek, Jun-Geol;Heo, Jun
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 2005.05a
    • /
    • pp.960-968
    • /
    • 2005
  • The high cost in maintaining complex manufacturing process makes it necessary to enhance an efficient maintenance system. For the effective maintenance of manufacturing process, precise fault diagnosis should be performed and an appropriate maintenance action should be executed. This paper suggests an intelligent fault diagnosis system using hybrid data mining. In this system, the rules for the fault diagnosis are generated by hybrid decision tree/genetic algorithm and the most effective maintenance action is selected by decision network and AHP. To verify the proposed intelligent fault diagnosis system, we compared the accuracy of the hybrid decision tree/genetic algorithm with one of the general decision tree learning algorithm(C4.5) by data collected from a coil-spring manufacturing process.

  • PDF

A Study on the Action decision by Changing of Condition of Time-Space (시·공간의 환경변화에 따른 행태 결정에 관한 연구)

  • Kim, Bo-Ra;Hong, Il-Tae
    • Korean Institute of Interior Design Journal
    • /
    • v.22 no.6
    • /
    • pp.98-107
    • /
    • 2013
  • Space evolves from the concept of deterministic static location to a dynamic, connected area through the interference of the user. While this does incorporate physical changes of the space, it also reflects the changes of the program or characteristics of the space through the actions and changes of the user. Therefore, in this study we plan to review the characteristic of time appearing within space, thereby discussing the impact of changing of condition in time-space to the decision making of the user. Further, we plan to analyze the specific causes, and subsequently introduce a new perspective over space. In order to achieve this, we need to first understand the reason why the attribute of time needs to be discussed in space, and perform a fundamental analysis of factors for the changes of the users' actions following changes in space-time condition. This means that space is not limited to merely satisfying its innate objective as an area, but may have a basis for modifying its role to help the decision making of the users caused by changes in space-time conditions. Accordingly, we analyze the factors for change of environment that can appear in space following the flow of time caused by correlation in space-time, as well as psychological factors and variables for decision making by the users. Based on this, we analyze cases to study the influence of condition changes in time-space on the action decision judgment of the users. Through this, we propose that the actions of the users can be determined following changes in time-space conditions, and discuss the need for changes in our perspective of space.

A Design of the Decision Maker of ECG Using the Intellegent Control System (지능 제어 시스템을 이용한 심전도 판단자 설계)

  • 김민수;김상득;구자헌;서희돈
    • Proceedings of the IEEK Conference
    • /
    • 2001.06e
    • /
    • pp.207-210
    • /
    • 2001
  • This Paper presents a design of the fuzzy decision maker analyzable of output result of ECG signals. The fuzzy decision maker proposed are divided into two groups whose functions are different each other. The one rules when decision of heart rates, The other decision values for an interval of each points of waveform using of which static state values and abnormal values. We have chosen several variable used for composing condition and action part by knowledge of an Expert The result of outputs with fuzzy rules suggested was a proved of satisfied with by classify ECG arrythmia signals

  • PDF