• Title/Summary/Keyword: machine vision application

Search Result 85, Processing Time 0.025 seconds

Real-Time Eye Detection and Tracking Under Various Light Conditions

  • Park Ho Sik;Nam Kee Hwan;Seol Jeung Bo;Cho Hyeon Seob;Ra Sang Dong;Bae Cheol Soo
    • Proceedings of the IEEK Conference
    • /
    • 2004.08c
    • /
    • pp.862-866
    • /
    • 2004
  • Non-intrusive methods based on active remote IR illumination for eye tracking is important for many applications of vision-based man-machine interaction. One problem that has plagued those methods is their sensitivity to lighting condition change. This tends to significantly limit their scope of application. In this paper, we present a new real-time eye detection and tracking methodology that works under variable and realistic lighting conditions. Based on combining the bright-pupil effect resulted from IR light and the conventional appearance-based object recognition technique, our method can robustly track eyes when the pupils are not very bright due to significant external illumination interferences. The appearance model is incorporated in both eyes detection and tracking via the use of support vector machine and the mean shift tracking. Additional improvement is achieved from modifying the image acquisition apparatus including the illuminator and the camera.

  • PDF

A Development of Video Monitoring System on Real Time (실시간 영상감시 시스템 개발)

  • Cho, Hyun-Seob
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.8 no.2
    • /
    • pp.240-244
    • /
    • 2007
  • Non-intrusive methods based on active remote IR illumination fur eye tracking is important for many applications of vision-based man-machine interaction. One problem that has plagued those methods is their sensitivity to lighting condition change. This tends to significantly limit their scope of application. In this paper, we present a new real-time eye detection and tracking methodology that works under variable and realistic lighting conditions. Based on combining the bright-pupil effect resulted from IR light and the conventional appearance-based object recognition technique, our method can robustly track eyes when the pupils are not very bright due to significant external illumination interferences. The appearance model is incorporated in both eyes detection and tracking via the use of support vector machine and the mean shift tracking. Additional improvement is achieved from modifying the image acquisition apparatus including the illuminator and the camera.

  • PDF

A study on real time inspection of OLED protective film using edge detecting algorithm (Edge Detecting Algorithm을 이용한 OLED 보호 필름의 Real Time Inspection에 대한 연구)

  • Han, Joo-Seok;Han, Bong-Seok;Han, Yu-Jin;Choi, Doo-Sun;Kim, Tae-Min;Ko, Kang-Ho;Park, Jung-Rae;Lim, Dong-Wook
    • Design & Manufacturing
    • /
    • v.14 no.2
    • /
    • pp.14-20
    • /
    • 2020
  • In OLED panel production process, it is necessary to cut a part of protective film as a preprocess for lighting inspection. The current method is to recognize only the fiducial mark of the cut-out panel. Bare Glass Cutting does not compensate for machining cumulative tolerances. Even though process defects still occur, it is necessary to develop technology to solve this problem because only the Align Mark of the panel that has already been cut is used as the reference point for alignment. There is a lot of defective lighting during panel lighting test because the correct protective film is not cut on the panel power and signal application pad position. In laser cutting process to remove the polarizing film / protective film / TSP film of OLED panel, laser processing is not performed immediately after the panel alignment based on the alignment mark only. Therefore, in this paper, we performed real time inspection which minimizes the mechanism tolerance by correcting the laser cutting path of the protective film in real time using Machine Vision. We have studied calibration algorithm of Vision Software coordinate system and real image coordinate system to minimize inspection resolution and position detection error and edge detection algorithm to accurately measure edge of panel.

A Review on Meat Quality Evaluation Methods Based on Non-Destructive Computer Vision and Artificial Intelligence Technologies

  • Shi, Yinyan;Wang, Xiaochan;Borhan, Md Saidul;Young, Jennifer;Newman, David;Berg, Eric;Sun, Xin
    • Food Science of Animal Resources
    • /
    • v.41 no.4
    • /
    • pp.563-588
    • /
    • 2021
  • Increasing meat demand in terms of both quality and quantity in conjunction with feeding a growing population has resulted in regulatory agencies imposing stringent guidelines on meat quality and safety. Objective and accurate rapid non-destructive detection methods and evaluation techniques based on artificial intelligence have become the research hotspot in recent years and have been widely applied in the meat industry. Therefore, this review surveyed the key technologies of non-destructive detection for meat quality, mainly including ultrasonic technology, machine (computer) vision technology, near-infrared spectroscopy technology, hyperspectral technology, Raman spectra technology, and electronic nose/tongue. The technical characteristics and evaluation methods were compared and analyzed; the practical applications of non-destructive detection technologies in meat quality assessment were explored; and the current challenges and future research directions were discussed. The literature presented in this review clearly demonstrate that previous research on non-destructive technologies are of great significance to ensure consumers' urgent demand for high-quality meat by promoting automatic, real-time inspection and quality control in meat production. In the near future, with ever-growing application requirements and research developments, it is a trend to integrate such systems to provide effective solutions for various grain quality evaluation applications.

A Self-Organizing Map Based Hough Transform for Detecting Straight Lines (직선 추출을 위한 자기조직화지도 기반의 허프 변환)

  • Lee, Moon-Kyu
    • Journal of Korean Institute of Industrial Engineers
    • /
    • v.28 no.2
    • /
    • pp.162-170
    • /
    • 2002
  • Detecting straight lines in an image is frequently required for various machine vision applications such as restoring CAD drawings from scanned images and object recognition. The standard Hough transform has been dominantly used to that purpose. However, massive storage requirement and low precision in estimating line parameters due to the quantization of parameter space are the major drawbacks of the Hough transform technique. In this paper, to overcome the drawbacks, an iterative algorithm based on a self-organizing map is presented. The self-organizing map can be adaptively learned such that image points are clustered by prominent lines. Through the procedure of the algorithm, a set of lines are sequentially detected one at a time. The algorithm can produce highly precised estimates of line parameters using very small amount of storage memory. Computational results for synthetically generated images are given. The promise of the algorithm is also demonstrated with its application to two natural images of inserts.

Combining Object Detection and Hand Gesture Recognition for Automatic Lighting System Control

  • Pham, Giao N.;Nguyen, Phong H.;Kwon, Ki-Ryong
    • Journal of Multimedia Information System
    • /
    • v.6 no.4
    • /
    • pp.329-332
    • /
    • 2019
  • Recently, smart lighting systems are the combination between sensors and lights. These systems turn on/off and adjust the brightness of lights based on the motion of object and the brightness of environment. These systems are often applied in places such as buildings, rooms, garages and parking lot. However, these lighting systems are controlled by lighting sensors, motion sensors based on illumination environment and motion detection. In this paper, we propose an automatic lighting control system using one single camera for buildings, rooms and garages. The proposed system is one integration the results of digital image processing as motion detection, hand gesture detection to control and dim the lighting system. The experimental results showed that the proposed system work very well and could consider to apply for automatic lighting spaces.

Utilization of Vision in Off-Line Teaching for assembly robot (조립용 로봇의 오프라인 교시를 위한 영상 정보의 이용에 관한 연구)

  • 안철기
    • Proceedings of the Korean Society of Machine Tool Engineers Conference
    • /
    • 2000.04a
    • /
    • pp.543-548
    • /
    • 2000
  • In this study, an interactive programming method for robot in electronic part assembly task is proposed. Many of industrial robots are still taught and programmed by a teach pendant. The robot is guided by a human operator to the desired application locations. These motions are recorded and are later edited, within the robotic language using in the robot controller, and play back repetitively to perform robot task. This conventional teaching method is time-consuming and somewhat dangerous. In the proposed method, the operator teaches the desired locations on the image acquired through CCD camera mounted on the robot hand. The robotic language program is automatically generated and downloaded to the robot controller. This teaching process is implemented through an off-line programming software. The OLP is developed for an robotic assembly system used in this study. In order to transform the location on image coordinates into robot coordinates, a calibration process is established. The proposed teaching method is implemented and evaluated on an assembly system for soldering electronic parts on a circuit board. A six-axis articulated robot executes assembly task according to the off-line teaching in the system.

  • PDF

A Study on Implementation of Image Processing System for the Defect Inspection of polyethylene (팔레트의 불량검사를 위한 영상 처리 시스템 구현)

  • Kim, Kyoung-Min;Kang, Jong-Su;Park, Joong-Jo;Song, Myeong-Hyun
    • Proceedings of the KIEE Conference
    • /
    • 2001.07d
    • /
    • pp.2738-2740
    • /
    • 2001
  • This paper describes a study on implementation of image processing systems for the defect inspection of polyethylene. In order to detect the edge, the Robert filter is used. And we use to the labeling algorithm for feature extraction. Labeling the conected regions of a image is a fundamental computation in image analysis and machine vision, with a large number of application. This algorithm is designed for the defect inspection of polyethylene.

  • PDF

Lightweight CNN based Meter Digit Recognition

  • Sharma, Akshay Kumar;Kim, Kyung Ki
    • Journal of Sensor Science and Technology
    • /
    • v.30 no.1
    • /
    • pp.15-19
    • /
    • 2021
  • Image processing is one of the major techniques that are used for computer vision. Nowadays, researchers are using machine learning and deep learning for the aforementioned task. In recent years, digit recognition tasks, i.e., automatic meter recognition approach using electric or water meters, have been studied several times. However, two major issues arise when we talk about previous studies: first, the use of the deep learning technique, which includes a large number of parameters that increase the computational cost and consume more power; and second, recent studies are limited to the detection of digits and not storing or providing detected digits to a database or mobile applications. This paper proposes a system that can detect the digital number of meter readings using a lightweight deep neural network (DNN) for low power consumption and send those digits to an Android mobile application in real-time to store them and make life easy. The proposed lightweight DNN is computationally inexpensive and exhibits accuracy similar to those of conventional DNNs.

Deep Reinforcement Learning in ROS-based autonomous robot navigation

  • Roland, Cubahiro;Choi, Donggyu;Jang, Jongwook
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.47-49
    • /
    • 2022
  • Robot navigation has seen a major improvement since the the rediscovery of the potential of Artificial Intelligence (AI) and the attention it has garnered in research circles. A notable achievement in the area was Deep Learning (DL) application in computer vision with outstanding daily life applications such as face-recognition, object detection, and more. However, robotics in general still depend on human inputs in certain areas such as localization, navigation, etc. In this paper, we propose a study case of robot navigation based on deep reinforcement technology. We look into the benefits of switching from traditional ROS-based navigation algorithms towards machine learning approaches and methods. We describe the state-of-the-art technology by introducing the concepts of Reinforcement Learning (RL), Deep Learning (DL) and DRL before before focusing on visual navigation based on DRL. The case study preludes further real life deployment in which mobile navigational agent learns to navigate unbeknownst areas.

  • PDF