• Title/Summary/Keyword: vision-based control

Search Result 683, Processing Time 0.029 seconds

Vision-Based Robot Manipulator for Grasping Objects (물체 잡기를 위한 비전 기반의 로봇 메뉴플레이터)

  • Baek, Young-Min;Ahn, Ho-Seok;Choi, Jin-Young
    • Proceedings of the KIEE Conference
    • /
    • 2007.04a
    • /
    • pp.331-333
    • /
    • 2007
  • Robot manipulator is one of the important features in service robot area. Until now, there has been a lot of research on robot" manipulator that can imitate the functions of a human being by recognizing and grasping objects. In this paper, we present a robot arm based on the object recognition vision system. We have implemented closed-loop control that use the feedback from visual information, and used a sonar sensor to improve the accuracy. We have placed the web-camera on the top of the hand to recognize objects. We also present some vision-based manipulation issues and our system features.

  • PDF

Development of Vision Control Scheme of Extended Kalman filtering for Robot's Position Control (실시간 로봇 위치 제어를 위한 확장 칼만 필터링의 비젼 저어 기법 개발)

  • Jang, W.S.;Kim, K.S.;Park, S.I.;Kim, K.Y.
    • Journal of the Korean Society for Nondestructive Testing
    • /
    • v.23 no.1
    • /
    • pp.21-29
    • /
    • 2003
  • It is very important to reduce the computational time in estimating the parameters of vision control algorithm for robot's position control in real time. Unfortunately, the batch estimation commonly used requires too murk computational time because it is iteration method. So, the batch estimation has difficulty for robot's position control in real time. On the other hand, the Extended Kalman Filtering(EKF) has many advantages to calculate the parameters of vision system in that it is a simple and efficient recursive procedures. Thus, this study is to develop the EKF algorithm for the robot's vision control in real time. The vision system model used in this study involves six parameters to account for the inner(orientation, focal length etc) and outer (the relative location between robot and camera) parameters of camera. Then, EKF has been first applied to estimate these parameters, and then with these estimated parameters, also to estimate the robot's joint angles used for robot's operation. finally, the practicality of vision control scheme based on the EKF has been experimentally verified by performing the robot's position control.

Robust Defect Size Measuring Method for an Automated Vision Inspection System (영상기반 자동결함 검사시스템에서 재현성 향상을 위한 결함 모델링 및 측정 기법)

  • Joo, Young-Bok;Huh, Kyung-Moo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.19 no.11
    • /
    • pp.974-978
    • /
    • 2013
  • AVI (Automatic Vision Inspection) systems automatically detect defect features and measure their sizes via camera vision. AVI systems usually report different measurements on the same defect with some variations on position or rotation mainly because different images are provided. This is caused by possible variations from the image acquisition process including optical factors, nonuniform illumination, random noises, and so on. For this reason, conventional area based defect measuring methods have problems of robustness and consistency. In this paper, we propose a new defect size measuring method to overcome this problem, utilizing volume information that is completely ignored in the area based defect measuring method. The results show that our proposed method dramatically improves the robustness and consistency of defect size measurement.

The GPU-based Parallel Processing Algorithm for Fast Inspection of Semiconductor Wafers (반도체 웨이퍼 고속 검사를 위한 GPU 기반 병렬처리 알고리즘)

  • Park, Youngdae;Kim, Joon Seek;Joo, Hyonam
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.19 no.12
    • /
    • pp.1072-1080
    • /
    • 2013
  • In a the present day, many vision inspection techniques are used in productive industrial areas. In particular, in the semiconductor industry the vision inspection system for wafers is a very important system. Also, inspection techniques for semiconductor wafer production are required to ensure high precision and fast inspection. In order to achieve these objectives, parallel processing of the inspection algorithm is essentially needed. In this paper, we propose the GPU (Graphical Processing Unit)-based parallel processing algorithm for the fast inspection of semiconductor wafers. The proposed algorithm is implemented on GPU boards made by NVIDIA Company. The defect detection performance of the proposed algorithm implemented on the GPU is the same as if by a single CPU, but the execution time of the proposed method is about 210 times faster than the one with a single CPU.

Visual Servoing of a Mobile Manipulator Based on Stereo Vision

  • Lee, H.J.;Park, M.G.;Lee, M.C.
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2003.10a
    • /
    • pp.767-771
    • /
    • 2003
  • In this study, stereo vision system is applied to a mobile manipulator for effective tasks. The robot can recognize a target and compute the position of the target using a stereo vision system. While a monocular vision system needs properties such as geometric shape of a target, a stereo vision system enables the robot to find the position of a target without additional information. Many algorithms have been studied and developed for an object recognition. However, most of these approaches have a disadvantage of the complexity of computations and they are inadequate for real-time visual servoing. However, color information is useful for simple recognition in real-time visual servoing. In this paper, we refer to about object recognition using colors, stereo matching method, recovery of 3D space and the visual servoing.

  • PDF

Light-Adaptive Vision System for Remote Surveillance Using an Edge Detection Vision Chip

  • Choi, Kyung-Hwa;Jo, Sung-Hyun;Seo, Sang-Ho;Shin, Jang-Kyoo
    • Journal of Sensor Science and Technology
    • /
    • v.20 no.3
    • /
    • pp.162-167
    • /
    • 2011
  • In this paper, we propose a vision system using a field programmable gate array(FPGA) and a smart vision chip. The output of the vision chip is varied by illumination conditions. This chip is suitable as a surveillance system in a dynamic environment. However, because the output swing of a smart vision chip is too small to definitely confirm the warning signal with the FPGA, a modification was needed for a reliable signal. The proposed system is based on a transmission control protocol/internet protocol(TCP/IP) that enables monitoring from a remote place. The warning signal indicates that some objects are too near.

Vision-Based Roadway Sign Recognition

  • Jiang, Gang-Yi;Park, Tae-Young;Hong, Suk-Kyo
    • Transactions on Control, Automation and Systems Engineering
    • /
    • v.2 no.1
    • /
    • pp.47-55
    • /
    • 2000
  • In this paper, a vision-based roadway detection algorithm for an automated vehicle control system, based on roadway sign information on roads, is proposed. First, in order to detect roadway signs, the color scene image is enhanced under hue-invariance. Fuzzy logic is employed to simplify the enhanced color image into a binary image and the binary image is morphologically filtered. Then, an effective algorithm of locating signs based on binary rank order transform (BROT) is utilized to extract signs from the image. This algorithm performs better than those previously presented. Finally, the inner shapes of roadway signs with curving roadway direction information are recognized by neural networks. Experimental results show that the new detection algorithm is simple and robust, and performs well on real sign detection. The results also show that the neural networks used can exactly recognize the inner shapes of signs even for very noisy shapes.

  • PDF

Omni-directional Vision SLAM using a Motion Estimation Method based on Fisheye Image (어안 이미지 기반의 움직임 추정 기법을 이용한 전방향 영상 SLAM)

  • Choi, Yun Won;Choi, Jeong Won;Dai, Yanyan;Lee, Suk Gyu
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.20 no.8
    • /
    • pp.868-874
    • /
    • 2014
  • This paper proposes a novel mapping algorithm in Omni-directional Vision SLAM based on an obstacle's feature extraction using Lucas-Kanade Optical Flow motion detection and images obtained through fish-eye lenses mounted on robots. Omni-directional image sensors have distortion problems because they use a fish-eye lens or mirror, but it is possible in real time image processing for mobile robots because it measured all information around the robot at one time. In previous Omni-Directional Vision SLAM research, feature points in corrected fisheye images were used but the proposed algorithm corrected only the feature point of the obstacle. We obtained faster processing than previous systems through this process. The core of the proposed algorithm may be summarized as follows: First, we capture instantaneous $360^{\circ}$ panoramic images around a robot through fish-eye lenses which are mounted in the bottom direction. Second, we remove the feature points of the floor surface using a histogram filter, and label the candidates of the obstacle extracted. Third, we estimate the location of obstacles based on motion vectors using LKOF. Finally, it estimates the robot position using an Extended Kalman Filter based on the obstacle position obtained by LKOF and creates a map. We will confirm the reliability of the mapping algorithm using motion estimation based on fisheye images through the comparison between maps obtained using the proposed algorithm and real maps.

Implementation of tracking and grasping the moving object using visual feedback (영상궤환을 이용한 이동체의 주적 및 잡기 작업의 구현)

  • Kwon, Chul;Kang, Hyung-Jin;Park, Mig-Non
    • Proceedings of the KIEE Conference
    • /
    • 1995.11a
    • /
    • pp.579-582
    • /
    • 1995
  • Recently, the vision system has the wide and growing' application field on account of the vast information from that visual mechanism. Especially, in the control field, the vision system has been applied to the industrial robot. In this paper, the object tracking and grasping task is accomplished by the robot vision system with a camera in the robot hand. The camera setting method is proposed to implement that task in a simple way. In spite of the calibration error, the stable grasping task is achieved using the tracking control algorithm based on the vision feature.

  • PDF

An Application of Computer Vision System for the Determination of Object Position in the Plane (평면상에 있는 물체 위치 결정을 위한 컴퓨터 비젼 시스템의 응용)

  • 장완식
    • Journal of the Korean Society of Manufacturing Technology Engineers
    • /
    • v.7 no.2
    • /
    • pp.62-68
    • /
    • 1998
  • This paper presents the application of computer vision for the purpose of determining the position of the unknown object in the plane. The presented control method is to estimate the six view parameters representing the relationship between the image plane coordinates and the real physical coordinates. The estimation of six parameters is indispensable for transforming the 2-dimensional camera coordinates to the 3-dimensional spatial coordinates. Then, the position of unknown point is estimated based on the estimated parameters depending on the cameras. The suitability of this control scheme is demonstrated experimentally by determining position of the unknown object in the plane.

  • PDF