• Title/Summary/Keyword: Vision Systems

Search Result 1,716, Processing Time 0.029 seconds

Moving Target Tracking using Vision System for an Omni-directional Wheel Robot (전방향 구동 로봇에서의 비젼을 이용한 이동 물체의 추적)

  • Kim, San;Kim, Dong-Hwan
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.14 no.10
    • /
    • pp.1053-1061
    • /
    • 2008
  • In this paper, a moving target tracking using a binocular vision for an omni-directional mobile robot is addressed. In the binocular vision, three dimensional information on the target is extracted by vision processes including calibration, image correspondence, and 3D reconstruction. The robot controller is constituted with SPI(serial peripheral interface) to communicate effectively between robot master controller and wheel controllers.

EVALUATION OF SPEED AND ACCURACY FOR COMPARISON OF TEXTURE CLASSIFICATION IMPLEMENTATION ON EMBEDDED PLATFORM

  • Tou, Jing Yi;Khoo, Kenny Kuan Yew;Tay, Yong Haur;Lau, Phooi Yee
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2009.01a
    • /
    • pp.89-93
    • /
    • 2009
  • Embedded systems are becoming more popular as many embedded platforms have become more affordable. It offers a compact solution for many different problems including computer vision applications. Texture classification can be used to solve various problems, and implementing it in embedded platforms will help in deploying these applications into the market. This paper proposes to deploy the texture classification algorithms onto the embedded computer vision (ECV) platform. Two algorithms are compared; grey level co-occurrence matrices (GLCM) and Gabor filters. Experimental results show that raw GLCM on MATLAB could achieves 50ms, being the fastest algorithm on the PC platform. Classification speed achieved on PC and ECV platform, in C, is 43ms and 3708ms respectively. Raw GLCM could achieve only 90.86% accuracy compared to the combination feature (GLCM and Gabor filters) at 91.06% accuracy. Overall, evaluating all results in terms of classification speed and accuracy, raw GLCM is more suitable to be implemented onto the ECV platform.

  • PDF

INS/Vision Integrated Navigation System Considering Error Characteristics of Landmark-Based Vision Navigation (랜드마크 기반 비전항법의 오차특성을 고려한 INS/비전 통합 항법시스템)

  • Kim, Youngsun;Hwang, Dong-Hwan
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.19 no.2
    • /
    • pp.95-101
    • /
    • 2013
  • The paper investigates the geometric effect of landmarks to the navigation error in the landmark based 3D vision navigation and introduces the INS/Vision integrated navigation system considering its effect. The integrated system uses the vision navigation results taking into account the dilution of precision for landmark geometry. Also, the integrated system helps the vision navigation to consider it. An indirect filter with feedback structure is designed, in which the position and the attitude errors are measurements of the filter. Performance of the integrated system is evaluated through the computer simulations. Simulation results show that the proposed algorithm works well and that better performance can be expected when the error characteristics of vision navigation are considered.

Image Enhanced Machine Vision System for Smart Factory

  • Kim, ByungJoo
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.13 no.2
    • /
    • pp.7-13
    • /
    • 2021
  • Machine vision is a technology that helps the computer as if a person recognizes and determines things. In recent years, as advanced technologies such as optical systems, artificial intelligence and big data advanced in conventional machine vision system became more accurate quality inspection and it increases the manufacturing efficiency. In machine vision systems using deep learning, the image quality of the input image is very important. However, most images obtained in the industrial field for quality inspection typically contain noise. This noise is a major factor in the performance of the machine vision system. Therefore, in order to improve the performance of the machine vision system, it is necessary to eliminate the noise of the image. There are lots of research being done to remove noise from the image. In this paper, we propose an autoencoder based machine vision system to eliminate noise in the image. Through experiment proposed model showed better performance compared to the basic autoencoder model in denoising and image reconstruction capability for MNIST and fashion MNIST data sets.

Development of a Ubiquitous Vision System for Location-awareness of Multiple Targets by a Matching Technique for the Identity of a Target;a New Approach

  • Kim, Chi-Ho;You, Bum-Jae;Kim, Hag-Bae
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.68-73
    • /
    • 2005
  • Various techniques have been proposed for detection and tracking of targets in order to develop a real-world computer vision system, e.g., visual surveillance systems, intelligent transport systems (ITSs), and so forth. Especially, the idea of distributed vision system is required to realize these techniques in a wide-spread area. In this paper, we develop a ubiquitous vision system for location-awareness of multiple targets. Here, each vision sensor that the system is composed of can perform exact segmentation for a target by color and motion information, and visual tracking for multiple targets in real-time. We construct the ubiquitous vision system as the multiagent system by regarding each vision sensor as the agent (the vision agent). Therefore, we solve matching problem for the identity of a target as handover by protocol-based approach. We propose the identified contract net (ICN) protocol for the approach. The ICN protocol not only is independent of the number of vision agents but also doesn't need calibration between vision agents. Therefore, the ICN protocol raises speed, scalability, and modularity of the system. We adapt the ICN protocol in our ubiquitous vision system that we construct in order to make an experiment. Our ubiquitous vision system shows us reliable results and the ICN protocol is successfully operated through several experiments.

  • PDF

Development and application of unmanned crane system in the warehouse (창고 Crane 무인화 시스템 개발 및 적용)

  • 박남수;김태진
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1996.10b
    • /
    • pp.1079-1082
    • /
    • 1996
  • Automatic control systems for warehouse composed of unmanned crane system and vision system. Unmanned crane system is introduced to reject oscillations of a load suspended from a trolley at a moment of its arrival at its target position. And vision system is applied to find out the coordinates of coils on trucks using image processing.

  • PDF

A Study on the Improvement of Vehicle Recognition Rate of Vision System (Vision 시스템의 차량 인식률 향상에 관한 연구)

  • Oh, Ju-Taek;Lee, Sang-Yong;Lee, Sang-Min;Kim, Young-Sam
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.10 no.3
    • /
    • pp.16-24
    • /
    • 2011
  • The vehicle electronic control system is being developed as the legal and social demand for ensuring driver's safety is rising. The various Driver Assistance Systems with various sensors such as radars, camera, and lasers are in practical use because of the falling price of hardware and the high performance of sensor and processer. In the preceding study of this research, the program was developed to recognize the experiment vehicle's driving lane and the cars nearby or approaching the experiment vehicle throughout the images taken by CCD camera. In addition, the 'dangerous driving analysis program' which is Vision System basis was developed to analyze the cause and consequence of dangerous driving. However, the Vision system developed in the previous studyhad poor recognition rate of lane and vehicles at the time of passing a tunnel, sunrise, or sunset. Therefore, through mounting the brightness response algorithm to the Vision System, the present study is aimed to analyze the causes of driver's dangerous driving clearly by improving the recognition rate of lane and vehicle, regardless of when and where it is.

Feature Extraction for Vision Based Micromanipulation

  • Jang, Min-Soo;Lee, Seok-Joo;Park, Gwi-Tae
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2002.10a
    • /
    • pp.41.5-41
    • /
    • 2002
  • This paper presents a feature extraction algorithm for vision-based micromanipulation. In order to guarantee of the accurate micromanipulation, most of micromanipulation systems use vision sensor. Vision data from an optical microscope or high magnification lens have vast information, however, characteristics of micro image such as emphasized contour, texture, and noise are make it difficult to apply macro image processing algorithms to micro image. Grasping points extraction is very important task in micromanipulation because inaccurate grasping points can cause breakdown of micro gripper or miss of micro objects. To solve those problems and extract grasping points for micromanipulation...

  • PDF

Visual Servoing of a Mobile Manipulator Based on Stereo Vision (스테레오 영상을 이용한 이동형 머니퓰레이터의 시각제어)

  • Lee Hyun Jeong;Park Min Gyu;Lee Min Cheol
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.11 no.5
    • /
    • pp.411-417
    • /
    • 2005
  • In this study, stereo vision system is applied to a mobile manipulator for effective tasks. The robot can recognize a target and compute the potion of the target using a stereo vision system. While a monocular vision system needs properties such as geometric shape of a target, a stereo vision system enables the robot to find the position of a target without additional information. Many algorithms have been studied and developed for an object recognition. However, most of these approaches have a disadvantage of the complexity of computations and they are inadequate for real-time visual servoing. Color information is useful for simple recognition in real-time visual servoing. This paper addresses object recognition using colors, stereo matching method to reduce its calculation time, recovery of 3D space and the visual servoing.

3-D vision sensor system for arc welding robot with coordinated motion by transputer system

  • Ishida, Hirofumi;Kasagami, Fumio;Ishimatsu, Takakazu
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1993.10b
    • /
    • pp.446-450
    • /
    • 1993
  • In this paper we propose an arc welding robot system, where two robots works coordinately and employ the vision sensor. In this system one robot arm holds a welding target as a positioning device, and the other robot moves the welding torch. The vision sensor consists of two laser slit-ray projectors and one CCD TV camera, and is mounted on the top of one robot. The vision sensor detects the 3-dimensional shape of the groove on the target work which needs to be weld. And two robots are moved coordinately to trace the grooves with accuracy. In order to realize fast image processing, totally five sets of high-speed parallel processing units (Transputer) are employed. The teaching tasks of the coordinated motions are simplified considerably due to this vision sensor. Experimental results reveal the applicability of our system.

  • PDF