• Title/Summary/Keyword: RGB Color Detector

Search Result 7, Processing Time 0.021 seconds

Development Small Size RGB Sensor for Providing Long Detecting Range (원거리 검출범위를 제공하는 소형 RGB 센서 개발)

  • Seo, Jae Yong;Lee, Si Hyun
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.52 no.12
    • /
    • pp.174-182
    • /
    • 2015
  • In this paper, we developed the small size RGB sensor that recognizes a long distance using a low-cost color sensor. Light receiving portion of the sensor was used as a camera lens for far distance recognition, and illuminating unit was increased the strength of the light by using a high-power white LED and a lens mounted on the reflector. RGB color recognition algorithm consists of the learning process and the realtime recognition process. We obtain a normalized RGB color reference data in the learning process using the specimens painted with target colors, and classifies the three colors using the Mahalanobis distance in recognition process. We apply the developed the RGB color recognition sensor to a prototype of the part classification system and evaluate the performance of its.

Properties of Photo Detector using SOI NMOSFET (SOI NMOSFET을 이용한 Photo Detector의 특성)

  • 김종준;정두연;이종호;오환술
    • Journal of the Korean Institute of Electrical and Electronic Material Engineers
    • /
    • v.15 no.7
    • /
    • pp.583-590
    • /
    • 2002
  • In this paper, a new Silicon on Insulator (SOI)-based photodetector was proposed, and its basic operation principle was explained. Fabrication steps of the detector are compatible with those of conventional SOI CMOS technology. With the proposed structure, RGB (Read, Green, Blue) which are three primary colors of light can be realized without using any organic color filters. It was shown that the characteristics of the SOI-based detector are better than those of bulk-based detector. To see the response characteristics to the green (G) among RGB, SOI and bulk NMOSFETS were fabricated using $1.5\mu m$ CMOS technology and characterized. We obtained optimum optical response characteristics at $V_{GS}=0.35 V$ in NMOSFET with threshold voltage of 0.72 V. Drain bias should be less than about 1.5 V to avoid any problem from floating body effect, since the body of the SOI NMOSFET was floated. The SOI and the bulk NMOSFETS shown maximum drain currents at the wavelengths of incident light around 550 nm and 750 nm, respectively. Therefore the SOI detector is more suitable for the G color detector.

Crop-row Detection by Color Line Sensor

  • Ha, S.ta;T.Kobaysahi;K.Sakai
    • Proceedings of the Korean Society for Agricultural Machinery Conference
    • /
    • 1993.10a
    • /
    • pp.353-362
    • /
    • 1993
  • The purpose of this study is to develop a crop-row detector which can be applied to an automatic row following control for cultivators or thinning machines. In this report, a possibility of new crop-row detecting method was discussed. This detecting method consists of two principal means. One is the hardware means to convert the two dimensional crop-row vision to the compacted one dimensional information. The conversion is achieved by a color line sensor and a rotating mirror. In order to extract crop-row , R and G signals of RGB color system are used. The locations of two different points on the target row are detected by this means. Another is the software means to estimate the offset value and the heading angle between the detector and the target row which can be assumed as a straight line. As a result of discussion, it was concluded that this detecting method would be accurate enough for practical use.

  • PDF

A Background Segmentation and Feature Point Extraction Method of Human Motion Recognition (동작인식을 위한 배경 분할 및 특징점 추출 방법)

  • You, Hwi-Jong;Kim, Tae-Young
    • Journal of Korea Game Society
    • /
    • v.11 no.2
    • /
    • pp.161-166
    • /
    • 2011
  • In this paper, we propose a novel background segmentation and feature point extraction method of a human motion for the augmented reality game. First, our method transforms input image from RGB color space to HSV color space, then segments a skin colored area using double threshold of H, S value. And it also segments a moving area using the time difference images and then removes the noise of the area using the Hessian affine region detector. The skin colored area with the moving area is segmented as a human motion. Next, the feature points for the human motion are extracted by calculating the center point for each block in the previously obtained image. The experiments on various input images show that our method is capable of correct background segmentation and feature points extraction 12 frames per second.

Pedestrian Detection Algorithm using a Gabor Filter Bank (Gabor Filter Bank를 이용한 보행자 검출 알고리즘)

  • Lee, Sewon;Jang, Jin-Won;Baek, Kwang-Ryul
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.20 no.9
    • /
    • pp.930-935
    • /
    • 2014
  • A Gabor filter is a linear filter used for edge detectionas frequency and orientation representations of Gabor filters are similar to those of the human visual system. In this thesis, we propose a pedestrian detection algorithm using a Gabor filter bank. In order to extract the features of the pedestrian, we use various image processing algorithms and data structure algorithms. First, color image segmentation is performed to consider the information of the RGB color space. Second, histogram equalization is performed to enhance the brightness of the input images. Third, convolution is performed between a Gabor filter bank and the enhanced images. Fourth, statistical values are calculated by using the integral image (summed area table) method. The calculated statistical values are used for the feature matrix of the pedestrian area. To evaluate the proposed algorithm, the INRIA pedestrian database and SVM (Support Vector Machine) are used, and we compare the proposed algorithm and the HOG (Histogram of Oriented Gradient) pedestrian detector, presentlyreferred to as the methodology of pedestrian detection algorithm. The experimental results show that the proposed algorithm is more accurate compared to the HOG pedestrian detector.

Estimating vegetation index for outdoor free-range pig production using YOLO

  • Sang-Hyon Oh;Hee-Mun Park;Jin-Hyun Park
    • Journal of Animal Science and Technology
    • /
    • v.65 no.3
    • /
    • pp.638-651
    • /
    • 2023
  • The objective of this study was to quantitatively estimate the level of grazing area damage in outdoor free-range pig production using a Unmanned Aerial Vehicles (UAV) with an RGB image sensor. Ten corn field images were captured by a UAV over approximately two weeks, during which gestating sows were allowed to graze freely on the corn field measuring 100 × 50 m2. The images were corrected to a bird's-eye view, and then divided into 32 segments and sequentially inputted into the YOLOv4 detector to detect the corn images according to their condition. The 43 raw training images selected randomly out of 320 segmented images were flipped to create 86 images, and then these images were further augmented by rotating them in 5-degree increments to create a total of 6,192 images. The increased 6,192 images are further augmented by applying three random color transformations to each image, resulting in 24,768 datasets. The occupancy rate of corn in the field was estimated efficiently using You Only Look Once (YOLO). As of the first day of observation (day 2), it was evident that almost all the corn had disappeared by the ninth day. When grazing 20 sows in a 50 × 100 m2 cornfield (250 m2/sow), it appears that the animals should be rotated to other grazing areas to protect the cover crop after at least five days. In agricultural technology, most of the research using machine and deep learning is related to the detection of fruits and pests, and research on other application fields is needed. In addition, large-scale image data collected by experts in the field are required as training data to apply deep learning. If the data required for deep learning is insufficient, a large number of data augmentation is required.

A Road Luminance Measurement Application based on Android (안드로이드 기반의 도로 밝기 측정 어플리케이션 구현)

  • Choi, Young-Hwan;Kim, Hongrae;Hong, Min
    • Journal of Internet Computing and Services
    • /
    • v.16 no.2
    • /
    • pp.49-55
    • /
    • 2015
  • According to the statistics of traffic accidents over recent 5 years, traffic accidents during the night times happened more than the day times. There are various causes to occur traffic accidents and the one of the major causes is inappropriate or missing street lights that make driver's sight confused and causes the traffic accidents. In this paper, with smartphones, we designed and implemented a lane luminance measurement application which stores the information of driver's location, driving, and lane luminance into database in real time to figure out the inappropriate street light facilities and the area that does not have any street lights. This application is implemented under Native C/C++ environment using android NDK and it improves the operation speed than code written in Java or other languages. To measure the luminance of road, the input image with RGB color space is converted to image with YCbCr color space and Y value returns the luminance of road. The application detects the road lane and calculates the road lane luminance into the database sever. Also this application receives the road video image using smart phone's camera and improves the computational cost by allocating the ROI(Region of interest) of input images. The ROI of image is converted to Grayscale image and then applied the canny edge detector to extract the outline of lanes. After that, we applied hough line transform method to achieve the candidated lane group. The both sides of lane is selected by lane detection algorithm that utilizes the gradient of candidated lanes. When the both lanes of road are detected, we set up a triangle area with a height 20 pixels down from intersection of lanes and the luminance of road is estimated from this triangle area. Y value is calculated from the extracted each R, G, B value of pixels in the triangle. The average Y value of pixels is ranged between from 0 to 100 value to inform a luminance of road and each pixel values are represented with color between black and green. We store car location using smartphone's GPS sensor into the database server after analyzing the road lane video image with luminance of road about 60 meters ahead by wireless communication every 10 minutes. We expect that those collected road luminance information can warn drivers about safe driving or effectively improve the renovation plans of road luminance management.