• Title/Summary/Keyword: 객체검출 모델

Search Result 240, Processing Time 0.032 seconds

Automated Analyses of Ground-Penetrating Radar Images to Determine Spatial Distribution of Buried Cultural Heritage (매장 문화재 공간 분포 결정을 위한 지하투과레이더 영상 분석 자동화 기법 탐색)

  • Kwon, Moonhee;Kim, Seung-Sep
    • Economic and Environmental Geology
    • /
    • v.55 no.5
    • /
    • pp.551-561
    • /
    • 2022
  • Geophysical exploration methods are very useful for generating high-resolution images of underground structures, and such methods can be applied to investigation of buried cultural properties and for determining their exact locations. In this study, image feature extraction and image segmentation methods were applied to automatically distinguish the structures of buried relics from the high-resolution ground-penetrating radar (GPR) images obtained at the center of Silla Kingdom, Gyeongju, South Korea. The major purpose for image feature extraction analyses is identifying the circular features from building remains and the linear features from ancient roads and fences. Feature extraction is implemented by applying the Canny edge detection and Hough transform algorithms. We applied the Hough transforms to the edge image resulted from the Canny algorithm in order to determine the locations the target features. However, the Hough transform requires different parameter settings for each survey sector. As for image segmentation, we applied the connected element labeling algorithm and object-based image analysis using Orfeo Toolbox (OTB) in QGIS. The connected components labeled image shows the signals associated with the target buried relics are effectively connected and labeled. However, we often find multiple labels are assigned to a single structure on the given GPR data. Object-based image analysis was conducted by using a Large-Scale Mean-Shift (LSMS) image segmentation. In this analysis, a vector layer containing pixel values for each segmented polygon was estimated first and then used to build a train-validation dataset by assigning the polygons to one class associated with the buried relics and another class for the background field. With the Random Forest Classifier, we find that the polygons on the LSMS image segmentation layer can be successfully classified into the polygons of the buried relics and those of the background. Thus, we propose that these automatic classification methods applied to the GPR images of buried cultural heritage in this study can be useful to obtain consistent analyses results for planning excavation processes.

Line-Segment Feature Analysis Algorithm for Handwritten-Digits Data Reduction (필기체 숫자 데이터 차원 감소를 위한 선분 특징 분석 알고리즘)

  • Kim, Chang-Min;Lee, Woo-Beom
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.10 no.4
    • /
    • pp.125-132
    • /
    • 2021
  • As the layers of artificial neural network deepens, and the dimension of data used as an input increases, there is a problem of high arithmetic operation requiring a lot of arithmetic operation at a high speed in the learning and recognition of the neural network (NN). Thus, this study proposes a data dimensionality reduction method to reduce the dimension of the input data in the NN. The proposed Line-segment Feature Analysis (LFA) algorithm applies a gradient-based edge detection algorithm using median filters to analyze the line-segment features of the objects existing in an image. Concerning the extracted edge image, the eigenvalues corresponding to eight kinds of line-segment are calculated, using 3×3 or 5×5-sized detection filters consisting of the coefficient values, including [0, 1, 2, 4, 8, 16, 32, 64, and 128]. Two one-dimensional 256-sized data are produced, accumulating the same response values from the eigenvalue calculated with each detection filter, and the two data elements are added up. Two LFA256 data are merged to produce 512-sized LAF512 data. For the performance evaluation of the proposed LFA algorithm to reduce the data dimension for the recognition of handwritten numbers, as a result of a comparative experiment, using the PCA technique and AlexNet model, LFA256 and LFA512 showed a recognition performance respectively of 98.7% and 99%.

Assessment of the FC-DenseNet for Crop Cultivation Area Extraction by Using RapidEye Satellite Imagery (RapidEye 위성영상을 이용한 작물재배지역 추정을 위한 FC-DenseNet의 활용성 평가)

  • Seong, Seon-kyeong;Na, Sang-il;Choi, Jae-wan
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_1
    • /
    • pp.823-833
    • /
    • 2020
  • In order to stably produce crops, there is an increasing demand for effective crop monitoring techniques in domestic agricultural areas. In this manuscript, a cultivation area extraction method by using deep learning model is developed, and then, applied to satellite imagery. Training dataset for crop cultivation areas were generated using RapidEye satellite images that include blue, green, red, red-edge, and NIR bands useful for vegetation and environmental analysis, and using this, we tried to estimate the crop cultivation area of onion and garlic by deep learning model. In order to training the model, atmospheric-corrected RapidEye satellite images were used, and then, a deep learning model using FC-DenseNet, which is one of the representative deep learning models for semantic segmentation, was created. The final crop cultivation area was determined as object-based data through combination with cadastral maps. As a result of the experiment, it was confirmed that the FC-DenseNet model learned using atmospheric-corrected training data can effectively detect crop cultivation areas.

A Driver's Condition Warning System using Eye Aspect Ratio (눈 영상비를 이용한 운전자 상태 경고 시스템)

  • Shin, Moon-Chang;Lee, Won-Young
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.15 no.2
    • /
    • pp.349-356
    • /
    • 2020
  • This paper introduces the implementation of a driver's condition warning system using eye aspect ratio to prevent a car accident. The proposed driver's condition warning system using eye aspect ratio consists of a camera, that is required to detect eyes, the Raspberrypie that processes information on eyes from the camera, buzzer and vibrator, that are required to warn the driver. In order to detect and recognize driver's eyes, the histogram of oriented gradients and face landmark estimation based on deep-learning are used. Initially the system calculates the eye aspect ratio of the driver from 6 coordinates around the eye and then gets each eye aspect ratio values when the eyes are opened and closed. These two different eye aspect ratio values are used to calculate the threshold value that is necessary to determine the eye state. Because the threshold value is adaptively determined according to the driver's eye aspect ratio, the system can use the optimal threshold value to determine the driver's condition. In addition, the system synthesizes an input image from the gray-scaled and LAB model images to operate in low lighting conditions.

Study on Face recognition algorithm using the eye detection (눈 검출을 이용한 얼굴인식 알고리즘에 관한 연구)

  • Park, Byung-Joon;Kim, Ki-young;Kim, Sun-jib
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.8 no.6
    • /
    • pp.491-496
    • /
    • 2015
  • Cloud computing has emerged with promise to decrease the cost of server additional cost and expanding the data storage and ease for computer resource sharing and apply the new technologies. However, Cloud computing also raises many new security concerns due to the new structure of the cloud service models. Therefore, the secure user authentication is required when the user is using cloud computing. This paper, we propose the enhanced AdaBoost algorithm for access cloud security zone. The AdaBoost algorithm despite the disadvantage of not detect a face inclined at least 20, is widely used because of speed and responsibility. In the experimental results confirm that a face inclined at least 20 degrees tilted face was recognized. Using the FEI Face Database that can be used in research to obtain a result of 98% success rate of the algorithm perform. The 2% failed rate is due to eye detection error which is the people wearing glasses in the picture.

Development of an intelligent edge computing device equipped with on-device AI vision model (온디바이스 AI 비전 모델이 탑재된 지능형 엣지 컴퓨팅 기기 개발)

  • Kang, Namhi
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.22 no.5
    • /
    • pp.17-22
    • /
    • 2022
  • In this paper, we design a lightweight embedded device that can support intelligent edge computing, and show that the device quickly detects an object in an image input from a camera device in real time. The proposed system can be applied to environments without pre-installed infrastructure, such as an intelligent video control system for industrial sites or military areas, or video security systems mounted on autonomous vehicles such as drones. The On-Device AI(Artificial intelligence) technology is increasingly required for the widespread application of intelligent vision recognition systems. Computing offloading from an image data acquisition device to a nearby edge device enables fast service with less network and system resources than AI services performed in the cloud. In addition, it is expected to be safely applied to various industries as it can reduce the attack surface vulnerable to various hacking attacks and minimize the disclosure of sensitive data.

Temporal Analysis of Agricultural Reservoir Water Surface Area using Remote Sensing and CNN (위성영상 및 CNN을 활용한 소규모 농업용 저수지의 수표면적 시계열 분석)

  • Yang, Mi-Hye;Nam, Won-Ho;Lee, Hee-Jin;Kim, Taegon
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2021.06a
    • /
    • pp.118-118
    • /
    • 2021
  • 최근 지구 온난화 현상으로 인한 기후변화로 이상기후 현상이 발생하고 있으며 이로 인해 장기적으로 폭염의 빈도 및 강도 상승에 따른 가뭄 피해 우려가 증가하고 있다. 농업 가뭄은 강수량 부족, 토양 수분 부족, 저수량 부족 등 농업분야에 영향을 주는 인자들과 관련되어 있어 농작물 생육 및 수확량 감소를 야기한다. 우리나라는 논농사가 주를 이루고 있어 국내 농업 가뭄은 주수원공인 농업용 저수지의 가용저수용량으로 판단 가능하다. 따라서 안정적인 농업용수 공급을 위해 수리시설물의 모니터링, 공급량 등의 분석이 이루어져야 하며, 농업 가뭄에 대비하기 위해 농업용 저수지의 가용저수용량 파악이 필요하다. 수자원 분야에서 지점자료의 시·공간적 한계점을 보완하기 위해 인공위성 자료를 활용한 연구가 활발히 이루어지고 있으며, 본 연구에서는 위성영상 자료 및 딥러닝 기반 알고리즘을 적용하여 농업용 저수지 수표면 탐지 및 시계열 분석을 목적으로 한다. 위성영상 자료는 5일 주기 및 10 m 공간해상도를 가진 Sentinel-2 위성영상 자료를 활용하고자 하였으며, 딥러닝에 적용하기 위하여 100장 이상의 영상 이미지를 구축하였다. 딥러닝 기반 알고리즘으로는 Convolutional Neural Network (CNN)을 활용하였으며, CNN은 주로 이미지 분류나 객체 검출 문제를 해결하기 위해 제안된 모델로 최근 픽셀 단위로 분류가 가능한 알고리즘이 개발되어 높은 정확도의 수표면 탐지가 가능할 것으로 판단된다. 따라서 본 연구에서는 CNN 기반 수표면 탐지 알고리즘을 개발하여 Sentinel-2 영상 기준 경기도 안성시를 대상으로 소규모 농업용 저수지의 수표면적에 대한 시계열 데이터를 분석하고자 한다.

  • PDF

Region of Interest Extraction and Bilinear Interpolation Application for Preprocessing of Lipreading Systems (입 모양 인식 시스템 전처리를 위한 관심 영역 추출과 이중 선형 보간법 적용)

  • Jae Hyeok Han;Yong Ki Kim;Mi Hye Kim
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.4
    • /
    • pp.189-198
    • /
    • 2024
  • Lipreading is one of the important parts of speech recognition, and several studies have been conducted to improve the performance of lipreading in lipreading systems for speech recognition. Recent studies have used method to modify the model architecture of lipreading system to improve recognition performance. Unlike previous research that improve recognition performance by modifying model architecture, we aim to improve recognition performance without any change in model architecture. In order to improve the recognition performance without modifying the model architecture, we refer to the cues used in human lipreading and set other regions such as chin and cheeks as regions of interest along with the lip region, which is the existing region of interest of lipreading systems, and compare the recognition rate of each region of interest to propose the highest performing region of interest In addition, assuming that the difference in normalization results caused by the difference in interpolation method during the process of normalizing the size of the region of interest affects the recognition performance, we interpolate the same region of interest using nearest neighbor interpolation, bilinear interpolation, and bicubic interpolation, and compare the recognition rate of each interpolation method to propose the best performing interpolation method. Each region of interest was detected by training an object detection neural network, and dynamic time warping templates were generated by normalizing each region of interest, extracting and combining features, and mapping the dimensionality reduction of the combined features into a low-dimensional space. The recognition rate was evaluated by comparing the distance between the generated dynamic time warping templates and the data mapped to the low-dimensional space. In the comparison of regions of interest, the result of the region of interest containing only the lip region showed an average recognition rate of 97.36%, which is 3.44% higher than the average recognition rate of 93.92% in the previous study, and in the comparison of interpolation methods, the bilinear interpolation method performed 97.36%, which is 14.65% higher than the nearest neighbor interpolation method and 5.55% higher than the bicubic interpolation method. The code used in this study can be found a https://github.com/haraisi2/Lipreading-Systems.

U-healthcare Based System for Sleeping Control and Remote Monitoring (u-헬스케어기반의 수면제어 및 원격모니터링 시스템)

  • Kim, Dong-Ho;Jeong, Chang-Won;Joo, Su-Chong
    • Journal of Internet Computing and Services
    • /
    • v.8 no.1
    • /
    • pp.33-45
    • /
    • 2007
  • Using switches and sensors informing the current on or off state, this paper suggests a sleeping control and remote monitoring system that not only can recognize the sleeping situations but also can control for keeping an appropriate sleeping situation remotely, And we show an example that this system is applied to the healthcare sleeping mat, Our system comprises the following 3 parts: a part for detecting the sleeping situations, a part for extracting sensing data and sending/receiving the relating situated data, and a part controlling and monitoring the all of sleeping situations. In details, in order to develop our system, we used the touch and pressure-sensitive sensors with On/Off functions for a purpose of the first part, The second part consists of the self-developed embedded board with the socket based communication as well as extracting real-time sensing data. And the third part is implemented by service modules for providing controlling and monitoring functions previously described. Finally, these service modules are implemented by the TMO scheme, one of real-time object-oriented programming models and the communications among them is supported using the TMOSM of distributed real-time middleware.

  • PDF

Corneal Ulcer Region Detection With Semantic Segmentation Using Deep Learning

  • Im, Jinhyuk;Kim, Daewon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.9
    • /
    • pp.1-12
    • /
    • 2022
  • Traditional methods of measuring corneal ulcers were difficult to present objective basis for diagnosis because of the subjective judgment of the medical staff through photographs taken with special equipment. In this paper, we propose a method to detect the ulcer area on a pixel basis in corneal ulcer images using a semantic segmentation model. In order to solve this problem, we performed the experiment to detect the ulcer area based on the DeepLab model which has the highest performance in semantic segmentation model. For the experiment, the training and test data were selected and the backbone network of DeepLab model which set as Xception and ResNet, respectively were evaluated and compared the performances. We used Dice similarity coefficient and IoU value as an indicator to evaluate the performances. Experimental results show that when 'crop & resized' images are added to the dataset, it segment the ulcer area with an average accuracy about 93% of Dice similarity coefficient on the DeepLab model with ResNet101 as the backbone network. This study shows that the semantic segmentation model used for object detection also has an ability to make significant results when classifying objects with irregular shapes such as corneal ulcers. Ultimately, we will perform the extension of datasets and experiment with adaptive learning methods through future studies so that they can be implemented in real medical diagnosis environment.