• Title/Summary/Keyword: object shape recognition

Search Result 163, Processing Time 0.03 seconds

Building Points Classification from Raw LiDAR Data by Information Theory (정보이론에 의한 LiDAR 원시자료의 건물포인트 분류기법 연구)

  • Choi Yun-Woong;Jang Young-Woon;Cho Gi-Sung
    • Proceedings of the Korean Society of Surveying, Geodesy, Photogrammetry, and Cartography Conference
    • /
    • 2006.04a
    • /
    • pp.469-473
    • /
    • 2006
  • In general, a classification process between ground data and non-ground data, which include building objects, is required prior to producing a DEM for a certain surface reconstruction from LiDAR data in which the DEM can be produced from the ground data, and certain objects like buildings can be reconstructed using non-ground data. Thus, an exact classification between ground and non-ground data from LiDAR data is the most important factor in the ground reconstruction process using LiDAR data. In particular, building objects can be largely used as digital maps, orthophotos, and urban planning regarding the object in the ground and become an essential to providing three dimensional information for certain urban areas. In this study, an entropy theory, which has been used as a standard of disorder or uncertainty for data used in the information theory, is used to apply a more objective and generalized method in the recognition and segmentation of buildings from raw LiDAR data. In particular, a method that directly uses the raw LiDAR data, which is a type of point shape vector data, without any changes, to a type of normal lattices was proposed, and the existing algorithm that segments LiDAR data into ground and non-ground data as a binarization manner was improved. In addition, this study proposes a generalized building extraction method that excludes precedent information for buildings and topographies and subsidiary materials, which have different data sources.

  • PDF

A Development of Object Shape Recognition Module using Laser Sensor (레이저 센서를 이용한 물체의 형상인식 모듈 개발)

  • Kwak, Sung-Hwan;Lee, Seung-Kyu;Lee, Seung-Jae;Kim, Young-Sik;Choi, Joong-Koung;Park, Mu-Hun
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2008.10a
    • /
    • pp.215-221
    • /
    • 2008
  • 무인 운반설비의 자동화 시스템 개발의 한 부분으로써 여러 Vision 센서 중 레이저 센서를 이용하여 작업공간상에 있는 판재류와 코일류의 경계부분을 인식한다. 다음으로 인식한 물체의 경계를 이용하여 3차원 공간상의 위치좌표를 추출하여 무인크레인에 이동해야할 위치좌표를 전달한다. 본 연구에서는, 첫 번째 레이저 센서를 이용한 물체의 경계 추출, 두 번째 레이저 센서의 z축 기울기 각 추출, 세 번째 인식한 경계를 이용하여 물체의 2차원 위치좌표 추출, 네 번째 레이저 센서를 이용하여 판재와 코일의 판별, 다섯 번째 물체 판별의 결과에 따른 판재와 코일의 3차원 위치좌표 추출을 목적으로 한다. 본 연구의 결과는 무인 운반설비의 자동화 시스템 개발에 상당한 도움이 될 것으로 기대된다.

  • PDF

Automatic Extraction and Measurement of Visual Features of Mushroom (Lentinus edodes L.) (표고 외관 특징점의 자동 추출 및 측정)

  • Hwang, Heon;Lee, Yong-Guk
    • Journal of Bio-Environment Control
    • /
    • v.1 no.1
    • /
    • pp.37-51
    • /
    • 1992
  • Quantizing and extracting visual features of mushroom(Lentinus edodes L.) are crucial to the sorting and grading automation, the growth state measurement, and the dried performance indexing. A computer image processing system was utilized for the extraction and measurement of visual features of front and back sides of the mushroom. The image processing system is composed of the IBM PC compatible 386DK, ITEX PCVISION Plus frame grabber, B/W CCD camera, VGA color graphic monitor, and image output RGB monitor. In this paper, an automatic thresholding algorithm was developed to yield the segmented binary image representing skin states of the front and back sides. An eight directional Freeman's chain coding was modified to solve the edge disconnectivity by gradually expanding the mask size of 3$\times$3 to 9$\times$9. A real scaled geometric quantity of the object was directly extracted from the 8-directional chain element. The external shape of the mushroom was analyzed and converted to the quantitative feature patterns. Efficient algorithms for the extraction of the selected feature patterns and the recognition of the front and back side were developed. The developed algorithms were coded in a menu driven way using MS_C language Ver.6.0, PC VISION PLUS library fuctions, and VGA graphic functions.

  • PDF

Development of Stair Climbing Robot for Delivery Based on Deep Learning (딥러닝 기반 자율주행 계단 등반 물품운송 로봇 개발)

  • Mun, Gi-Il;Lee, Seung-Hyeon;Choo, Jeong-Pil;Oh, Yeon-U;Lee, Sang-Soon
    • Journal of the Semiconductor & Display Technology
    • /
    • v.21 no.4
    • /
    • pp.121-125
    • /
    • 2022
  • This paper deals with the development of a deep-learning-based robot that recognizes various types of stairs and performs a mission to go up to the target floor. The overall motion sequence of the robot is performed based on the ROS robot operating system, and it is possible to detect the shape of the stairs required to implement the motion sequence through rapid object recognition through YOLOv4 and Cuda acceleration calculations. Using the ROS operating system installed in Jetson Nano, a system was built to support communication between Arduino DUE and OpenCM 9.04 with heterogeneous hardware and to control the movement of the robot by aligning the received sensors and data. In addition, the web server for robot control was manufactured as ROS web server, and flow chart and basic ROS communication were designed to enable control through computer and smartphone through message passing.

A Comparative Study on Artificial in Intelligence Model Performance between Image and Video Recognition in the Fire Detection Area (화재 탐지 영역의 이미지와 동영상 인식 사이 인공지능 모델 성능 비교 연구)

  • Jeong Rok Lee;Dae Woong Lee;Sae Hyun Jeong;Sang Jeong
    • Journal of the Society of Disaster Information
    • /
    • v.19 no.4
    • /
    • pp.968-975
    • /
    • 2023
  • Purpose: We would like to confirm that the false positive rate of flames/smoke is high when detecting fires. Propose a method and dataset to recognize and classify fire situations to reduce the false detection rate. Method: Using the video as learning data, the characteristics of the fire situation were extracted and applied to the classification model. For evaluation, the model performance of Yolov8 and Slowfast were compared and analyzed using the fire dataset conducted by the National Information Society Agency (NIA). Result: YOLO's detection performance varies sensitively depending on the influence of the background, and it was unable to properly detect fires even when the fire scale was too large or too small. Since SlowFast learns the time axis of the video, we confirmed that detects fire excellently even in situations where the shape of an atypical object cannot be clearly inferred because the surrounding area is blurry or bright. Conclusion: It was confirmed that the fire detection rate was more appropriate when using a video-based artificial intelligence detection model rather than using image data.

A Selection of Threshold for the Generalized Hough Transform: A Probabilistic Approach (일반화된 허프변환의 임계값 선택을 위한 확률적 접근방식)

  • Chang, Ji Y.
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.51 no.1
    • /
    • pp.161-171
    • /
    • 2014
  • When the Hough transform is applied to identify an instance of a given model, the output is typically a histogram of votes cast by a set of image features into a parameter space. The next step is to threshold the histogram of counts to hypothesize a given match. The question is "What is a reasonable choice of the threshold?" In a standard implementation of the Hough transform, the threshold is selected heuristically, e.g., some fraction of the highest cell count. Setting the threshold too low can give rise to a false alarm of a given shape(Type I error). On the other hand, setting the threshold too high can result in mis-detection of a given shape(Type II error). In this paper, we derive two conditional probability functions of cell counts in the accumulator array of the generalized Hough transform(GHough), that can be used to select a scientific threshold at the peak detection stage of the Ghough.

A Study on the Deep Neural Network based Recognition Model for Space Debris Vision Tracking System (심층신경망 기반 우주파편 영상 추적시스템 인식모델에 대한 연구)

  • Lim, Seongmin;Kim, Jin-Hyung;Choi, Won-Sub;Kim, Hae-Dong
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.45 no.9
    • /
    • pp.794-806
    • /
    • 2017
  • It is essential to protect the national space assets and space environment safely as a space development country from the continuously increasing space debris. And Active Debris Removal(ADR) is the most active way to solve this problem. In this paper, we studied the Artificial Neural Network(ANN) for a stable recognition model of vision-based space debris tracking system. We obtained the simulated image of the space environment by the KARICAT which is the ground-based space debris clearing satellite testbed developed by the Korea Aerospace Research Institute, and created the vector which encodes structure and color-based features of each object after image segmentation by depth discontinuity. The Feature Vector consists of 3D surface area, principle vector of point cloud, 2D shape and color information. We designed artificial neural network model based on the separated Feature Vector. In order to improve the performance of the artificial neural network, the model is divided according to the categories of the input feature vectors, and the ensemble technique is applied to each model. As a result, we confirmed the performance improvement of recognition model by ensemble technique.

A Hand Gesture Recognition System using 3D Tracking Volume Restriction Technique (3차원 추적영역 제한 기법을 이용한 손 동작 인식 시스템)

  • Kim, Kyung-Ho;Jung, Da-Un;Lee, Seok-Han;Choi, Jong-Soo
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.6
    • /
    • pp.201-211
    • /
    • 2013
  • In this paper, we propose a hand tracking and gesture recognition system. Our system employs a depth capture device to obtain 3D geometric information of user's bare hand. In particular, we build a flexible tracking volume and restrict the hand tracking area, so that we can avoid diverse problems caused by conventional object detection/tracking systems. The proposed system computes running average of the hand position, and tracking volume is actively adjusted according to the statistical information that is computed on the basis of uncertainty of the user's hand motion in the 3D space. Once the position of user's hand is obtained, then the system attempts to detect stretched fingers to recognize finger gesture of the user's hand. In order to test the proposed framework, we built a NUI system using the proposed technique, and verified that our system presents very stable performance even in the case that multiple objects exist simultaneously in the crowded environment, as well as in the situation that the scene is occluded temporarily. We also verified that our system ensures running speed of 24-30 frames per second throughout the experiments.

Effective Detection of Target Region Using a Machine Learning Algorithm (기계 학습 알고리즘을 이용한 효과적인 대상 영역 분할)

  • Jang, Seok-Woo;Lee, Gyungju;Jung, Myunghee
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.19 no.5
    • /
    • pp.697-704
    • /
    • 2018
  • Since the face in image content corresponds to individual information that can distinguish a specific person from other people, it is important to accurately detect faces not hidden in an image. In this paper, we propose a method to accurately detect a face from input images using a deep learning algorithm, which is one of the machine learning methods. In the proposed method, image input via the red-green-blue (RGB) color model is first changed to the luminance-chroma: blue-chroma: red-chroma ($YC_bC_r$) color model; then, other regions are removed using the learned skin color model, and only the skin regions are segmented. A CNN model-based deep learning algorithm is then applied to robustly detect only the face region from the input image. Experimental results show that the proposed method more efficiently segments facial regions from input images. The proposed face area-detection method is expected to be useful in practical applications related to multimedia and shape recognition.

Image Analysis for Discrimination of Neoplastic Cellis in Spatial Frequency Domain (종양세포식별을 위한 공간주파수영역에서의 화상해석)

  • 나철훈;김창원;김현재
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.18 no.3
    • /
    • pp.385-396
    • /
    • 1993
  • In this paper, a improved method of digital image analysis required in basic medical science for diagnosis of cells was proposed. The object image was the thyroid gland cell image, and the purpose was automatic discrimination of three classes cells(normal cell, follicular neoplastic cells, and papillary neoplastic cells) by difference of chromatin patterns. To segment the cell nucleus from background, the region segmentation algorithm by edge tracing was proposed. And feature parameter was obtained from discrete Fourier transformation of image. After construct a feature sample group of each cells, experiment of discrimination was executed with any verification cells. As a consequency of using features proposed in this paper, get a better recognition rate(70-90%) than previously reported papers, and this method give shape to get objectivity and fixed quantity in diagnosis of cells, The methods described in this paper be used immediately for discrimination of neoplastic cells.

  • PDF