• Title/Summary/Keyword: Robot-based Learning

Search Result 479, Processing Time 0.03 seconds

A reinforcement learning-based method for the cooperative control of mobile robots (강화 학습에 의한 소형 자율 이동 로봇의 협동 알고리즘 구현)

  • 김재희;조재승;권인소
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1997.10a
    • /
    • pp.648-651
    • /
    • 1997
  • This paper proposes methods for the cooperative control of multiple mobile robots and constructs a robotic soccer system in which the cooperation will be implemented as a pass play of two robots. To play a soccer game, elementary actions such as shooting and moving have been designed, and Q-learning, which is one of the popular methods for reinforcement learning, is used to determine what actions to take. Through simulation, learning is successful in case of deliberate initial arrangements of ball and robots, thereby cooperative work can be accomplished.

  • PDF

Map-Based Obstacle Avoidance Algorithm for Mobile Robot Using Deep Reinforcement Learning (심층 강화학습을 이용한 모바일 로봇의 맵 기반 장애물 회피 알고리즘)

  • Sunwoo, Yung-Min;Lee, Won-Chang
    • Journal of IKEEE
    • /
    • v.25 no.2
    • /
    • pp.337-343
    • /
    • 2021
  • Deep reinforcement learning is an artificial intelligence algorithm that enables learners to select optimal behavior based on raw and, high-dimensional input data. A lot of research using this is being conducted to create an optimal movement path of a mobile robot in an environment in which obstacles exist. In this paper, we selected the Dueling Double DQN (D3QN) algorithm that uses the prioritized experience replay to create the moving path of mobile robot from the image of the complex surrounding environment. The virtual environment is implemented using Webots, a robot simulator, and through simulation, it is confirmed that the mobile robot grasped the position of the obstacle in real time and avoided it to reach the destination.

Intelligent Robot Design: Intelligent Agent Based Approach (지능로봇: 지능 에이전트를 기초로 한 접근방법)

  • Kang, Jin-Shig
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.14 no.4
    • /
    • pp.457-467
    • /
    • 2004
  • In this paper, a robot is considered as an agent, a structure of robot is presented which consisted by multi-subagents and they have diverse capacity such as perception, intelligence, action etc., required for robot. Also, subagents are consisted by micro-agent($\mu$agent) charged for elementary action required. The structure of robot control have two sub-agents, the one is behavior based reactive controller and action selection sub agent, and action selection sub-agent select a action based on the high label action and high performance, and which have a learning mechanism based on the reinforcement learning. For presented robot structure, it is easy to give intelligence to each element of action and a new approach of multi robot control. Presented robot is simulated for two goals: chaotic exploration and obstacle avoidance, and fabricated by using 8bit microcontroller, and experimented.

Energy-Efficient DNN Processor on Embedded Systems for Spontaneous Human-Robot Interaction

  • Kim, Changhyeon;Yoo, Hoi-Jun
    • Journal of Semiconductor Engineering
    • /
    • v.2 no.2
    • /
    • pp.130-135
    • /
    • 2021
  • Recently, deep neural networks (DNNs) are actively used for action control so that an autonomous system, such as the robot, can perform human-like behaviors and operations. Unlike recognition tasks, the real-time operation is essential in action control, and it is too slow to use remote learning on a server communicating through a network. New learning techniques, such as reinforcement learning (RL), are needed to determine and select the correct robot behavior locally. In this paper, we propose an energy-efficient DNN processor with a LUT-based processing engine and near-zero skipper. A CNN-based facial emotion recognition and an RNN-based emotional dialogue generation model is integrated for natural HRI system and tested with the proposed processor. It supports 1b to 16b variable weight bit precision with and 57.6% and 28.5% lower energy consumption than conventional MAC arithmetic units for 1b and 16b weight precision. Also, the near-zero skipper reduces 36% of MAC operation and consumes 28% lower energy consumption for facial emotion recognition tasks. Implemented in 65nm CMOS process, the proposed processor occupies 1784×1784 um2 areas and dissipates 0.28 mW and 34.4 mW at 1fps and 30fps facial emotion recognition tasks.

Robot Control via SGA-based Reinforcement Learning Algorithms (SGA 기반 강화학습 알고리즘을 이용한 로봇 제어)

  • 박주영;김종호;신호근
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2004.10a
    • /
    • pp.63-66
    • /
    • 2004
  • The SGA(stochastic gradient ascent) algorithm is one of the most important tools in the area of reinforcement learning, and has been applied to a wide range of practical problems. In particular, this learning method was successfully applied by Kimura et a1. [1] to the control of a simple creeping robot which has finite number of control input choices. In this paper, we considered the application of the SGA algorithm to Kimura's robot control problem for the case that the control input is not confined to a finite set but can be chosen from a infinite subset of the real numbers. We also developed a MATLAB-based robot animation program, which showed the effectiveness of the training algorithms vividly.

  • PDF

Development of Runway Cleaning Robot Based on Deep Learning (딥러닝 기반 활주로 청소 로봇 개발)

  • Park, Ga-Gyeong;Kim, Ji-Yong;Keum, Jae-Yeong;Lee, Sang Soon
    • Journal of the Semiconductor & Display Technology
    • /
    • v.20 no.3
    • /
    • pp.140-145
    • /
    • 2021
  • This paper deals with the development of a deep-learning-based runway cleaning robot using an optical camera. A suitable model to realize real-time object detection was investigated, and the differences between the selected YOLOv3 and other deep learning models were analyzed. In order to check whether the proposed system is applicable to the actual runway, an experiment was conducted by making a prototype of the robot and a runway model. As a result, it was confirmed that the robot was well developed because the detection rate of FOD (Foreign Object Debris) and cracks was high, and the collection of foreign substances was carried out smoothly.

Development of a Robot Personality based on Cultural Paradigm using Fuzzy Logic (퍼지 로직을 이용한 문화 패러다임 기반의 로봇 성격 개발)

  • Qureshi, Favad Ahmed;Kim, Eun-Tai;Park, Mi-Gnon
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2008.04a
    • /
    • pp.385-391
    • /
    • 2008
  • Robotics has emerged as an important field for the future. It is our vision that robots in future will be able to transcend these precincts and work side by side humans for the greater good of mankind. We developed a face robot for this purpose. However, Life like robots demands a certain level of intelligence. Some scientists have proposed an event based learning approach, in which the robot can be taken as a small child and through learning from surrounding entities develops its own personality. In fact some scientists have proposed an entire new personality of the robot itself in which robot can have its own internal states, intentions, beliefs, desires and feelings. Our approach should not only be to develop a robot personality model but also to understand human behavior and incorporate it into the robot model. Human's personality is very complex and rests on many factors like its physical surrounding, its social surrounding, and internal states and beliefs etc. This paper discusses the development of this platform to evaluate this and develop a standard by a society based approach including the cultural paradigm. For this purpose the fuzzy control theory is used. Since the fuzzy theory is very near human analytical thinking it provides a very good platform to develop such a model.

  • PDF

DNN-Based Adaptive Optimal Learning Controller for Uncertain Robot Systems (동적 신경망에 기초한 불확실한 로봇 시스템의 적응 최적 학습제어기)

  • 정재욱;국태용;이택종
    • Journal of the Korean Institute of Telematics and Electronics S
    • /
    • v.34S no.6
    • /
    • pp.1-10
    • /
    • 1997
  • This paper presents an adaptive optimal learning controller for uncertian robot systems which makes use fo simple DNN(dynamic neural network) units to estimate uncertain parameters and learn the unknown desired optimal input. With the aid of a lyapunov function, it is shown that all that error signals in the system are bounded and the robot trajectory converges to the desired one globally exponentially. The effectiveness of the proposed controller is hsown by applying the controller to a 2-DOF robot manipulator.

  • PDF

Control of Walking Robot based on Reinforcement Learning and Manifold Control (강화학습과 메니폴드 제어기법을 이용한 걷는 로봇의 제어)

  • Mun, Yeong-Jun;Park, Ju-Yeong
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2008.04a
    • /
    • pp.135-138
    • /
    • 2008
  • 최근 인간을 모방하는 휴머노이드 로봇(Humanoid robot)에 대한 관심이 증가함에 따라, 기계공학, 생체공학, 제어이론 등 여러 분야에서 관련 연구가 활발히 진행되고 있다. 이에 본 논문에서는 액츄에이터(Actuator)가 없이 경사진 지면을 걸을 수 있는 두 발을 가진 패시브 로봇(Passive robot)을 대상으로 강화학습과 메니폴드(Manifold control) 기법을 사용하여 안정적으로 걸을 수 있도록 제어기(Controller)를 설계하는 방안을 고려한다.

  • PDF

Convergence Technologies by a Long-term Case Study on Telepresence Robot-assisted Learning (텔레프리젠스 로봇보조학습 사례 연구를 통한 융합기술)

  • Lim, Mi-Suk;Han, Jeong-Hye
    • Journal of Convergence for Information Technology
    • /
    • v.9 no.7
    • /
    • pp.106-113
    • /
    • 2019
  • The purpose of this paper is aimed to derive suggestions for convergence technology for effective management of distance education by analyzing a long-term case. The experiment was designed with notebook, smartphone or tablet based robot controlled by a remote instructor and a learner, who have experience of distance learning including robot assisted learning. The tablet based robot has the display system of feedback to speakers. During five months, three types of experiments were conducted randomly and a participant was interviewed thoroughly. The result, like the previous research, demonstrates that the task performance of the learner in telepresence robot-assisted learning was better than that in the notebook, and smartphone based. However, it is believed to be necessary to adjust the system for eye-contact and voice transmission for the remote instructor. The instructor required an additional sight by supplementing an extra camera and automatic direction control to source of sound.