• Title/Summary/Keyword: Robot Vision

Search Result 877, Processing Time 0.034 seconds

A vision based people tracking and following for mobile robots using CAMSHIFT and KLT feature tracker (캠시프트와 KLT특징 추적 알고리즘을 융합한 모바일 로봇의 영상기반 사람추적 및 추종)

  • Lee, S.J.;Won, Mooncheol
    • Journal of Korea Multimedia Society
    • /
    • v.17 no.7
    • /
    • pp.787-796
    • /
    • 2014
  • Many mobile robot navigation methods utilize laser scanners, ultrasonic sensors, vision camera, and so on for detecting obstacles and path following. However, human utilizes only vision(e.g. eye) information for navigation. In this paper, we study a mobile robot control method based on only the camera vision. The Gaussian Mixture Model and a shadow removal technology are used to divide the foreground and the background from the camera image. The mobile robot uses a combined CAMSHIFT and KLT feature tracker algorithms based on the information of the foreground to follow a person. The algorithm is verified by experiments where a person is tracked and followed by a robot in a hallway.

Design of a service robot with dual manipulators and stereo vision (Dual Manipulator와 Stereo Vision을 이용한 서비스 로봇)

  • Lee, Dae-Hui;Lee, Hui-Guk;U, Gyeong-Seok;Ham, Sang-Hwa;Park, Ju-Hyeon;Lee, Seok-Gyu
    • Proceedings of the KIEE Conference
    • /
    • 2003.11c
    • /
    • pp.743-746
    • /
    • 2003
  • The service robot, with stereo vision system and dual manipulator of four degree of freedom, has been designed. A fuzzy controller has been implemented for effectively actuating the manipulator of the robot. The fuzzy controller determines operation mode(single or dual manipulators) and orientation from the information of object position and distance. Through actual experimentation, we have confirmed that the robot system with human-like movement of grabber has been executed a rapid and effective motion.

  • PDF

Automatic Extraction of Stable Visual Landmarks for a Mobile Robot under Uncertainty (이동로봇의 불확실성을 고려한 시각 랜드마크의 자동 추출)

  • 문인혁;조강현;윤형로
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2000.10a
    • /
    • pp.264-264
    • /
    • 2000
  • In this paper, we propose a method to automatically extract stable visual landmarks from observed data for a mobile robot with stereo vision system. The robot selects as stable landmarks vertical line segments which are distinct and on planar surfaces, because they are expected to be observed reliably from various view-points. When the robot moves, it uses several, less uncertain landmarks for estimating its motion. Experimental results in real scenes show the validity of the proposed method.

  • PDF

Self-Localization of Autonomous Mobile Robot using Multiple Landmarks (다중 표식을 이용한 자율이동로봇의 자기위치측정)

  • 강현덕;조강현
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.10 no.1
    • /
    • pp.81-86
    • /
    • 2004
  • This paper describes self-localization of a mobile robot from the multiple candidates of landmarks in outdoor environment. Our robot uses omnidirectional vision system for efficient self-localization. This vision system acquires the visible information of all direction views. The robot uses feature of landmarks whose size is bigger than that of others in image such as building, sculptures, placard etc. Robot uses vertical edges and those merged regions as the feature. In our previous work, we found the problem that landmark matching is difficult when selected candidates of landmarks belonging to region of repeating the vertical edges in image. To overcome these problems, robot uses the merged region of vertical edges. If interval of vertical edges is short then robot bundles them regarding as the same region. Thus, these features are selected as candidates of landmarks. Therefore, the extracted merged region of vertical edge reduces the ambiguity of landmark matching. Robot compares with the candidates of landmark between previous and current image. Then, robot is able to find the same landmark between image sequences using the proposed feature and method. We achieved the efficient self-localization result using robust landmark matching method through the experiments implemented in our campus.

Robot Vision to Audio Description Based on Deep Learning for Effective Human-Robot Interaction (효과적인 인간-로봇 상호작용을 위한 딥러닝 기반 로봇 비전 자연어 설명문 생성 및 발화 기술)

  • Park, Dongkeon;Kang, Kyeong-Min;Bae, Jin-Woo;Han, Ji-Hyeong
    • The Journal of Korea Robotics Society
    • /
    • v.14 no.1
    • /
    • pp.22-30
    • /
    • 2019
  • For effective human-robot interaction, robots need to understand the current situation context well, but also the robots need to transfer its understanding to the human participant in efficient way. The most convenient way to deliver robot's understanding to the human participant is that the robot expresses its understanding using voice and natural language. Recently, the artificial intelligence for video understanding and natural language process has been developed very rapidly especially based on deep learning. Thus, this paper proposes robot vision to audio description method using deep learning. The applied deep learning model is a pipeline of two deep learning models for generating natural language sentence from robot vision and generating voice from the generated natural language sentence. Also, we conduct the real robot experiment to show the effectiveness of our method in human-robot interaction.

Development of Vision Control Scheme of Extended Kalman filtering for Robot's Position Control (실시간 로봇 위치 제어를 위한 확장 칼만 필터링의 비젼 저어 기법 개발)

  • Jang, W.S.;Kim, K.S.;Park, S.I.;Kim, K.Y.
    • Journal of the Korean Society for Nondestructive Testing
    • /
    • v.23 no.1
    • /
    • pp.21-29
    • /
    • 2003
  • It is very important to reduce the computational time in estimating the parameters of vision control algorithm for robot's position control in real time. Unfortunately, the batch estimation commonly used requires too murk computational time because it is iteration method. So, the batch estimation has difficulty for robot's position control in real time. On the other hand, the Extended Kalman Filtering(EKF) has many advantages to calculate the parameters of vision system in that it is a simple and efficient recursive procedures. Thus, this study is to develop the EKF algorithm for the robot's vision control in real time. The vision system model used in this study involves six parameters to account for the inner(orientation, focal length etc) and outer (the relative location between robot and camera) parameters of camera. Then, EKF has been first applied to estimate these parameters, and then with these estimated parameters, also to estimate the robot's joint angles used for robot's operation. finally, the practicality of vision control scheme based on the EKF has been experimentally verified by performing the robot's position control.

Path Planning and Obstacle Avoidance for Mobile Robot with Vision System Using Fuzzy Rules (비전과 퍼지 규칙을 이용한 이동로봇의 경로계획과 장애물회피)

  • 배봉규;채양범;이원창;강근택
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.11 no.6
    • /
    • pp.470-476
    • /
    • 2001
  • This paper presents a new algorithm of path planning and obstacle avoidance for autonomous mobile robots with vision system that is working in unknown environments. Distance variation technique is used in path planning to approach the target and avoid obstacles in work space as well . In this approach, the Sobel operator is employed to detect edges of obstacles and the distances between the mobile robot and the obstacles are measured. Fuzzy rules are used for trajectory planning and obstacle avoidance to improve the autonomy of mobile robots. It is shown by computer simulation that the proposed algorithm is superior to the vector field approach which sometimes traps the mobile robot into some local obstacles. An autonomous mobile robot with single vision is developed for experiments. We also show that the developed mobile robot with the proposed algorithm is navigating very well in complex unknown environments.

  • PDF

A Study on the Environment Recognition System of Biped Robot for Stable Walking (안정적 보행을 위한 이족 로봇의 환경 인식 시스템 연구)

  • Song, Hee-Jun;Lee, Seon-Gu;Kang, Tae-Gu;Kim, Dong-Won;Park, Gwi-Tae
    • Proceedings of the KIEE Conference
    • /
    • 2006.07d
    • /
    • pp.1977-1978
    • /
    • 2006
  • This paper discusses the method of vision based sensor fusion system for biped robot walking. Most researches on biped walking robot have mostly focused on walking algorithm itself. However, developing vision systems for biped walking robot is an important and urgent issue since biped walking robots are ultimately developed not only for researches but to be utilized in real life. In the research, systems for environment recognition and tele-operation have been developed for task assignment and execution of biped robot as well as for human robot interaction (HRI) system. For carrying out certain tasks, an object tracking system using modified optical flow algorithm and obstacle recognition system using enhanced template matching and hierarchical support vector machine algorithm by wireless vision camera are implemented with sensor fusion system using other sensors installed in a biped walking robot. Also systems for robot manipulating and communication with user have been developed for robot.

  • PDF