• Title/Summary/Keyword: Learning Agent

Search Result 448, Processing Time 0.02 seconds

Designing a Reinforcement Learning-Based 3D Object Reconstruction Data Acquisition Simulation (강화학습 기반 3D 객체복원 데이터 획득 시뮬레이션 설계)

  • Young-Hoon Jin
    • Journal of Internet of Things and Convergence
    • /
    • v.9 no.6
    • /
    • pp.11-16
    • /
    • 2023
  • The technology of 3D reconstruction, primarily relying on point cloud data, is essential for digitizing objects or spaces. This paper aims to utilize reinforcement learning to achieve the acquisition of point clouds in a given environment. To accomplish this, a simulation environment is constructed using Unity, and reinforcement learning is implemented using the Unity package known as ML-Agents. The process of point cloud acquisition involves initially setting a goal and calculating a traversable path around the goal. The traversal path is segmented at regular intervals, with rewards assigned at each step. To prevent the agent from deviating from the path, rewards are increased. Additionally, rewards are granted each time the agent fixates on the goal during traversal, facilitating the learning of optimal points for point cloud acquisition at each traversal step. Experimental results demonstrate that despite the variability in traversal paths, the approach enables the acquisition of relatively accurate point clouds.

Process Chain-Based Information Systems Development and Agent-Based Microworld Simulation As Enablers of the Learning & Agile Organization (학습, 민활 조직 실현을 위한 프로세스 사슬 기반 정보시스템 개발과 에이전트 기반 소세계 시뮬레이션)

  • Park, Kwang-Ho
    • Asia pacific journal of information systems
    • /
    • v.9 no.3
    • /
    • pp.177-194
    • /
    • 1999
  • Identifying knowledge as the single most important asset ultimately defining organizational competitiveness, enterprises are trying to move towards knowledge-oriented practices. Such practices have given rise to learning and agile organization, This paper presents applied information technologies to realize the learning and agile organization, focusing on systems thinking. Firstly, in order to establish a framework for the systems thinking, an information systems development method based on process chain is proposed. Then, an agent-based microworld simulation approach is presented. The approaches provide visible and analytical information to knowledge workers so that they can have systems thinking capabilities eventually. Various microworlds on the top of the information system can be constructed with agents and simulated for possible business events. All decision makings are dynamic in nature. To let knowledge workers look ahead the possible outcomes of the whole relevant processes is the core capability of the approaches. Through watching, the knowledge workers would be able to acquire new insights or problem solving knowledge for the problem in hand.

  • PDF

Multagent Control Strategy Using Reinforcement Learning (강화학습을 이용한 다중 에이전트 제어 전략)

  • Lee, Hyong-Ill;Kim, Byung-Cheon
    • The KIPS Transactions:PartB
    • /
    • v.10B no.3
    • /
    • pp.249-256
    • /
    • 2003
  • The most important problems in the multi-agent system are to accomplish a goal through the efficient coordination of several agents and to prevent collision with other agents. In this paper, we propose a new control strategy for succeeding the goal of the prey pursuit problem efficiently. Our control method uses reinforcement learning to control the multi-agent system and consider the distance as well as the space relationship between the agents in the state space of the prey pursuit problem.

Visual Analysis of Deep Q-network

  • Seng, Dewen;Zhang, Jiaming;Shi, Xiaoying
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.3
    • /
    • pp.853-873
    • /
    • 2021
  • In recent years, deep reinforcement learning (DRL) models are enjoying great interest as their success in a variety of challenging tasks. Deep Q-Network (DQN) is a widely used deep reinforcement learning model, which trains an intelligent agent that executes optimal actions while interacting with an environment. This model is well known for its ability to surpass skilled human players across many Atari 2600 games. Although DQN has achieved excellent performance in practice, there lacks a clear understanding of why the model works. In this paper, we present a visual analytics system for understanding deep Q-network in a non-blind matter. Based on the stored data generated from the training and testing process, four coordinated views are designed to expose the internal execution mechanism of DQN from different perspectives. We report the system performance and demonstrate its effectiveness through two case studies. By using our system, users can learn the relationship between states and Q-values, the function of convolutional layers, the strategies learned by DQN and the rationality of decisions made by the agent.

L-CAA : An Architecture for Behavior-Based Reinforcement Learning (L-CAA : 행위 기반 강화학습 에이전트 구조)

  • Hwang, Jong-Geun;Kim, In-Cheol
    • Journal of Intelligence and Information Systems
    • /
    • v.14 no.3
    • /
    • pp.59-76
    • /
    • 2008
  • In this paper, we propose an agent architecture called L-CAA that is quite effective in real-time dynamic environments. L-CAA is an extension of CAA, the behavior-based agent architecture which was also developed by our research group. In order to improve adaptability to the changing environment, it is extended by adding reinforcement learning capability. To obtain stable performance, however, behavior selection and execution in the L-CAA architecture do not entirely rely on learning. In L-CAA, learning is utilized merely as a complimentary means for behavior selection and execution. Behavior selection mechanism in this architecture consists of two phases. In the first phase, the behaviors are extracted from the behavior library by checking the user-defined applicable conditions and utility of each behavior. If multiple behaviors are extracted in the first phase, the single behavior is selected to execute in the help of reinforcement learning in the second phase. That is, the behavior with the highest expected reward is selected by comparing Q values of individual behaviors updated through reinforcement learning. L-CAA can monitor the maintainable conditions of the executing behavior and stop immediately the behavior when some of the conditions fail due to dynamic change of the environment. Additionally, L-CAA can suspend and then resume the current behavior whenever it encounters a higher utility behavior. In order to analyze effectiveness of the L-CAA architecture, we implement an L-CAA-enabled agent autonomously playing in an Unreal Tournament game that is a well-known dynamic virtual environment, and then conduct several experiments using it.

  • PDF

Grouping System for e-Learning Community(GSE): based on Intelligent Personalized Agent (온라인 학습공동체 그룹핑 시스템 개발: 지능적 에이전트 활용)

  • Kim, Myung Sook;Cho, Young Im
    • The Journal of Korean Association of Computer Education
    • /
    • v.7 no.6
    • /
    • pp.117-128
    • /
    • 2004
  • Compared with traditional face-to-face instruction, online learning causes learners to experience more severe feeling of isolation and results in higher dropout rate. This is due to the lack of interaction, sense of belonging, membership, interdependency, cooperation among members and social environment that enables persistence in online learning. Therefore, it is very important for grouping e-learning community to lower the dropout rate and eliminate feeling of isolation. In this paper, the research has been done on the inclination test list to be applied for grouping the desirable learning community. And on the basis of this research, the grouping system for e-learning community(GSE) based on intelligent multi agents for an inclination test using homogeneous and heterogeneous items has been developed. GSE system has such properties that construct a personalized user profile by an agent, and then make groupings according to users' inclination. When this system was evaluated, about 88% of learners were satisfied, and they wanted the group not to be disorganized but to be maintained.

  • PDF

Reinforcement Learning Using State Space Compression (상태 공간 압축을 이용한 강화학습)

  • Kim, Byeong-Cheon;Yun, Byeong-Ju
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.3
    • /
    • pp.633-640
    • /
    • 1999
  • Reinforcement learning performs learning through interacting with trial-and-error in dynamic environment. Therefore, in dynamic environment, reinforcement learning method like Q-learning and TD(Temporal Difference)-learning are faster in learning than the conventional stochastic learning method. However, because many of the proposed reinforcement learning algorithms are given the reinforcement value only when the learning agent has reached its goal state, most of the reinforcement algorithms converge to the optimal solution too slowly. In this paper, we present COMREL(COMpressed REinforcement Learning) algorithm for finding the shortest path fast in a maze environment, select the candidate states that can guide the shortest path in compressed maze environment, and learn only the candidate states to find the shortest path. After comparing COMREL algorithm with the already existing Q-learning and Priortized Sweeping algorithm, we could see that the learning time shortened very much.

  • PDF

FuzzyQ-Learning to Process the Vague Goals of Intelligent Agent (지능형 에이전트의 모호한 목적을 처리하기 위한 FuzzyQ-Learning)

  • 서호섭;윤소정;오경환
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2000.04b
    • /
    • pp.271-273
    • /
    • 2000
  • 일반적으로, 지능형 에이전트는 사용자의 목적과 주위 환경으로부터 최적의 행동을 스스로 찾아낼 수 있어야 한다. 만약 에이전트의 목적이나 주위 환경이 불확실성을 포함하는 경우, 에이전트는 적절한 행동을 선택하기 어렵다. 그러나, 사용자의 목적이 인간 지식의 불확실성을 포함하는 언어값으로 표현되었을 경우, 이를 처리하려는 연구는 없었다. 본 논문에서는 모호한 사용자의 의도를 퍼지 목적으로 나타내고, 에이전트가 인지하는 불확실한 환경을 퍼지 상태로 표현하는 방법을 제안한다. 또, 퍼지 목적과 상태를 이용하여 확장한 펴지 강화 함수와를 이용하여, 기존 강화 학습 알고리즘 중 하나인 Q-Learning을 FuzzyQ-Learning으로 확장하고, 이에 대한 타당성을 검증한다.

  • PDF

Effects of Red Ginseng Extract Including Vitamin B Groups on Learning and Memory in Mice (비타민 B군이 함유된 홍삼 추출물이 학습 및 기억에 미치는 영향)

  • 김학성;장춘곤
    • Journal of Ginseng Research
    • /
    • v.20 no.3
    • /
    • pp.226-232
    • /
    • 1996
  • This study was performed to investigate the effect of red ginseng extract including some vitamin B groups as test drug on learning and memory in mice. Single and repeated administrations of the test drug improved the acquisition and the process of consolidation in the tests using step-through and step-down apparatus, indicating this test drug improved learning and memory. However, the test drug did not improve scopolamine-induced amnesia. These results suggest that test drug may be useful as a nootropic agent.

  • PDF

KubEVC-Agent : Kubernetes Edge Vision Cluster Agent for Optimal DNN Inference and Operation (KubEVC-Agent : 머신러닝 추론 엣지 컴퓨팅 클러스터 관리 자동화 시스템)

  • Moohyun Song;Kyumin Kim;Jihun Moon;Yurim Kim;Chaewon Nam;Jongbin Park;Kyungyong Lee
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.6
    • /
    • pp.293-301
    • /
    • 2023
  • With the advancement of artificial intelligence and its various use cases, accessing it through edge computing environments is gaining traction. However, due to the nature of edge computing environments, efficient management and optimization of clusters distributed in different geographical locations is considered a major challenge. To address these issues, this paper proposes a centralization and automation tool called KubEVC-Agent based on Kubernetes. KubEVC-Agent centralizes the deployment, operation, and management of edge clusters and presents a use case of the data transformation for optimizing intra-cluster communication. This paper describes the components of KubEVC-Agent, its working principle, and experimental results to verify its effectiveness.