• Title/Summary/Keyword: Canny Edge Algorithm

Search Result 107, Processing Time 0.023 seconds

A Study on Finding the Rail Space in Elevators Using Matched Filter

  • Song, Myong-Lyol
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.6
    • /
    • pp.57-65
    • /
    • 2019
  • In this paper, we study on finding the rail space in elevators by analyzing each image captured with CCD camera. We propose a method that applies one-dimensional matched filter to the pixels of a selected search space in the vertical line at a horizontal position and decides the position with the thickness of the space being represented by a black thick line in captured images. The pattern similarity representing how strongly the associated image pixels resemble with the thick line is defined and calculated with respect to each position along the vertical line of pixels. The position and thickness of the line are decided from the point having the maximum in pattern similarity graph. In the experiments of the proposed method under different illuminational conditions, it is observed that all the pattern similarity graphs show similar shape around door area independent of the conditions and the method can effectively detect the rail space if the rails are illuminated with even weak light. The method can be used for real-time embedded systems because of its simple algorithm, in which it is implemented in simple structure of program with small amount of operations in comparison with the conventional approaches using Canny edge detection and Hough transform.

Morphology-Based Step Response Extraction and Regularized Iterative Point Spread Function Estimation & Image Restoration (수리형태학적 분석을 통한 계단응답 추출 및 반복적 정칙화 방법을 이용한 점확산함수 추정 및 영상 복원)

  • Park, Young-Uk;Jeon, Jae-Hwan;Lee, Jin-Hee;Kang, Nam-Oh;Paik, Joon-Ki
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.46 no.6
    • /
    • pp.26-35
    • /
    • 2009
  • In this paper, we present morphology-based step region extraction and regularized iterative point-spread-function (PSF) estimation methods. The proposed PSF estimation method uses canny edge detector to extract the edge of the input image. We extract feasible vertical and horizontal edges using morphology analysis, such as the hit-or-miss transform. Given extracted edges we estimate the optimal step-response using flattening and normalization processes. The PSF is finally characterized by solving the equation which relates the optimal step response and the 2D isotropic PSF. We shows the restored image by the estimated PSF. The proposed algorithm can be applied a fully digital auto-focusing system without using mechanical focusing parts.

Salient Object Extraction from Video Sequences using Contrast Map and Motion Information (대비 지도와 움직임 정보를 이용한 동영상으로부터 중요 객체 추출)

  • Kwak, Soo-Yeong;Ko, Byoung-Chul;Byun, Hye-Ran
    • Journal of KIISE:Software and Applications
    • /
    • v.32 no.11
    • /
    • pp.1121-1135
    • /
    • 2005
  • This paper proposes a moving object extraction method using the contrast map and salient points. In order to make the contrast map, we generate three-feature maps such as luminance map, color map and directional map and extract salient points from an image. By using these features, we can decide the Attention Window(AW) location easily The purpose of the AW is to remove the useless regions in the image such as background as well as to reduce the amount of image processing. To create the exact location and flexible size of the AW, we use motion feature instead of pre-assumptions or heuristic parameters. After determining of the AW, we find the difference of edge to inner area from the AW. Then, we can extract horizontal candidate region and vortical candidate region. After finding both horizontal and vertical candidates, intersection regions through logical AND operation are further processed by morphological operations. The proposed algorithm has been applied to many video sequences which have static background like surveillance type of video sequences. The moving object was quite well segmented with accurate boundaries.

Sensor Fusion Docking System of Drone and Ground Vehicles Using Image Object Detection (영상 객체 검출을 이용한 드론과 지상로봇의 센서 융합 도킹 시스템)

  • Beck, Jong-Hwan;Park, Hee-Su;Oh, Se-Ryeong;Shin, Ji-Hun;Kim, Sang-Hoon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.6 no.4
    • /
    • pp.217-222
    • /
    • 2017
  • Recent studies for working robot in dangerous places have been carried out on large unmanned ground vehicles or 4-legged robots with the advantage of long working time, but it is difficult to apply in practical dangerous fields which require the real-time system with high locomotion and capability of delicate working. This research shows the collaborated docking system of drone and ground vehicles which combines image processing algorithm and laser sensors for effective detection of docking markers, and is finally capable of moving a long distance and doing very delicate works. We proposed the docking system of drone and ground vehicles with sensor fusion which also suggests two template matching methods appropriate for this application. The system showed 95% docking success rate in 50 docking attempts.

A Morphology Technique-Based Boundary Detection in a Two-Dimensional QR Code (2차원 QR코드에서 모폴로지 기반의 경계선 검출 방법)

  • Park, Kwang Wook;Lee, Jong Yun
    • Journal of Digital Convergence
    • /
    • v.13 no.2
    • /
    • pp.159-175
    • /
    • 2015
  • The two-dimensional QR code has advantages such as directional nature, enough data storage capacity, ability of error correction, and ability of data restoration. There are two major issues like speed and correctiveness of recognition in the two-dimensional QR code. Therefore, this paper proposes a morphology-based algorithm of detecting the interest region of a barcode. Our research contents can be summarized as follows. First, the interest region of a barcode image was detected by close operations in morphology. Second, after that, the boundary of the barcode are detected by intersecting four cross line outside in a code. Three, the projected image is then rectified into a two-dimensional barcode in a square shape by the reverse-perspective transform. In result, it shows that our detection and recognition rates for the barcode image is also 97.20% and 94.80%, respectively and that outperforms than previous methods in various illumination and distorted image environments.

3D Film Image Inspection Based on the Width of Optimized Height of Histogram (히스토그램의 최적 높이의 폭에 기반한 3차원 필름 영상 검사)

  • Jae-Eun Lee;Jong-Nam Kim
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.23 no.2
    • /
    • pp.107-114
    • /
    • 2022
  • In order to classify 3D film images as right or wrong, it is necessary to detect the pattern in a 3D film image. However, if the contrast of the pixels in the 3D film image is low, it is not easy to classify as the right and wrong 3D film images because the pattern in the image might not be clear. In this paper, we propose a method of classifying 3D film images as right or wrong by comparing the width at a specific frequency of each histogram after obtaining the histogram. Since, it is classified using the width of the histogram, the analysis process is not complicated. From the experiment, the histograms of right and wrong 3D film images were distinctly different, and the proposed algorithm reflects these features, and showed that all 3D film images were accurately classified at a specific frequency of the histogram. The performance of the proposed algorithm was verified to be the best through the comparison test with the other methods such as image subtraction, otsu thresholding, canny edge detection, morphological geodesic active contour, and support vector machines, and it was shown that excellent classification accuracy could be obtained without detecting the patterns in 3D film images.

A Road Luminance Measurement Application based on Android (안드로이드 기반의 도로 밝기 측정 어플리케이션 구현)

  • Choi, Young-Hwan;Kim, Hongrae;Hong, Min
    • Journal of Internet Computing and Services
    • /
    • v.16 no.2
    • /
    • pp.49-55
    • /
    • 2015
  • According to the statistics of traffic accidents over recent 5 years, traffic accidents during the night times happened more than the day times. There are various causes to occur traffic accidents and the one of the major causes is inappropriate or missing street lights that make driver's sight confused and causes the traffic accidents. In this paper, with smartphones, we designed and implemented a lane luminance measurement application which stores the information of driver's location, driving, and lane luminance into database in real time to figure out the inappropriate street light facilities and the area that does not have any street lights. This application is implemented under Native C/C++ environment using android NDK and it improves the operation speed than code written in Java or other languages. To measure the luminance of road, the input image with RGB color space is converted to image with YCbCr color space and Y value returns the luminance of road. The application detects the road lane and calculates the road lane luminance into the database sever. Also this application receives the road video image using smart phone's camera and improves the computational cost by allocating the ROI(Region of interest) of input images. The ROI of image is converted to Grayscale image and then applied the canny edge detector to extract the outline of lanes. After that, we applied hough line transform method to achieve the candidated lane group. The both sides of lane is selected by lane detection algorithm that utilizes the gradient of candidated lanes. When the both lanes of road are detected, we set up a triangle area with a height 20 pixels down from intersection of lanes and the luminance of road is estimated from this triangle area. Y value is calculated from the extracted each R, G, B value of pixels in the triangle. The average Y value of pixels is ranged between from 0 to 100 value to inform a luminance of road and each pixel values are represented with color between black and green. We store car location using smartphone's GPS sensor into the database server after analyzing the road lane video image with luminance of road about 60 meters ahead by wireless communication every 10 minutes. We expect that those collected road luminance information can warn drivers about safe driving or effectively improve the renovation plans of road luminance management.