• Title/Summary/Keyword: 합성신경망

Search Result 649, Processing Time 0.027 seconds

Deep-Learning-Based Mine Detection Using Simulated Data (시뮬레이션 데이터 기반으로 학습된 딥러닝 모델을 활용한 지뢰식별연구)

  • Buhwan Jeon;Chunju Lee
    • Journal of The Korean Institute of Defense Technology
    • /
    • v.5 no.4
    • /
    • pp.16-21
    • /
    • 2023
  • Although the global number of landmines is on a declining trend, the damages caused by previously buried landmines persist. In light of this, the present study contemplates solutions to issues and constraints that may arise due to the improvement of mine detection equipment and the reduction in the number of future soldiers. Current mine detectors lack data storage capabilities, posing limitations on data collection for research purposes. Additionally, practical data collection in real-world environments demands substantial time and manpower. Therefore, in this study, gprMax simulation was utilized to generate data. The lightweight CNN-based model, MobileNet, was trained and validated with real data, achieving a high identification rate of 97.35%. Consequently, the potential integration of technologies such as deep learning and simulation into geographical detection equipment is highlighted, offering a pathway to address potential future challenges. The study aims to somewhat alleviate these issues and anticipates contributing to the development of our military capabilities in becoming a future scientific and technological force.

  • PDF

Rainfall image DB construction for rainfall intensity estimation from CCTV videos: focusing on experimental data in a climatic environment chamber (CCTV 영상 기반 강우강도 산정을 위한 실환경 실험 자료 중심 적정 강우 이미지 DB 구축 방법론 개발)

  • Byun, Jongyun;Jun, Changhyun;Kim, Hyeon-Joon;Lee, Jae Joon;Park, Hunil;Lee, Jinwook
    • Journal of Korea Water Resources Association
    • /
    • v.56 no.6
    • /
    • pp.403-417
    • /
    • 2023
  • In this research, a methodology was developed for constructing an appropriate rainfall image database for estimating rainfall intensity based on CCTV video. The database was constructed in the Large-Scale Climate Environment Chamber of the Korea Conformity Laboratories, which can control variables with high irregularity and variability in real environments. 1,728 scenarios were designed under five different experimental conditions. 36 scenarios and a total of 97,200 frames were selected. Rain streaks were extracted using the k-nearest neighbor algorithm by calculating the difference between each image and the background. To prevent overfitting, data with pixel values greater than set threshold, compared to the average pixel value for each image, were selected. The area with maximum pixel variability was determined by shifting with every 10 pixels and set as a representative area (180×180) for the original image. After re-transforming to 120×120 size as an input data for convolutional neural networks model, image augmentation was progressed under unified shooting conditions. 92% of the data showed within the 10% absolute range of PBIAS. It is clear that the final results in this study have the potential to enhance the accuracy and efficacy of existing real-world CCTV systems with transfer learning.

A Study on the Restoration of Korean Traditional Palace Image by Adjusting the Receptive Field of Pix2Pix (Pix2Pix의 수용 영역 조절을 통한 전통 고궁 이미지 복원 연구)

  • Hwang, Won-Yong;Kim, Hyo-Kwan
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.15 no.5
    • /
    • pp.360-366
    • /
    • 2022
  • This paper presents a AI model structure for restoring Korean traditional palace photographs, which remain only black-and-white photographs, to color photographs using Pix2Pix, one of the adversarial generative neural network techniques. Pix2Pix consists of a combination of a synthetic image generator model and a discriminator model that determines whether a synthetic image is real or fake. This paper deals with an artificial intelligence model by adjusting a receptive field of the discriminator, and analyzes the results by considering the characteristics of the ancient palace photograph. The receptive field of Pix2Pix, which is used to restore black-and-white photographs, was commonly used in a fixed size, but a fixed size of receptive field is not suitable for a photograph which consisting with various change in an image. This paper observed the result of changing the size of the existing fixed a receptive field to identify the proper size of the discriminator that could reflect the characteristics of ancient palaces. In this experiment, the receptive field of the discriminator was adjusted based on the prepared ancient palace photos. This paper measure a loss of the model according to the change in a receptive field of the discriminator and check the results of restored photos using a well trained AI model from experiments.

Semantic Segmentation of Hazardous Facilities in Rural Area Using U-Net from KOMPSAT Ortho Mosaic Imagery (KOMPSAT 정사모자이크 영상으로부터 U-Net 모델을 활용한 농촌위해시설 분류)

  • Sung-Hyun Gong;Hyung-Sup Jung;Moung-Jin Lee;Kwang-Jae Lee;Kwan-Young Oh;Jae-Young Chang
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.6_3
    • /
    • pp.1693-1705
    • /
    • 2023
  • Rural areas, which account for about 90% of the country's land area, are increasing in importance and value as a space that performs various public functions. However, facilities that adversely affect residents' lives, such as livestock facilities, factories, and solar panels, are being built indiscriminately near residential areas, damaging the rural environment and landscape and lowering the quality of residents' lives. In order to prevent disorderly development in rural areas and manage rural space in a planned manner, detection and monitoring of hazardous facilities in rural areas is necessary. Data can be acquired through satellite imagery, which can be acquired periodically and provide information on the entire region. Effective detection is possible by utilizing image-based deep learning techniques using convolutional neural networks. Therefore, U-Net model, which shows high performance in semantic segmentation, was used to classify potentially hazardous facilities in rural areas. In this study, KOMPSAT ortho-mosaic optical imagery provided by the Korea Aerospace Research Institute in 2020 with a spatial resolution of 0.7 meters was used, and AI training data for livestock facilities, factories, and solar panels were produced by hand for training and inference. After training with U-Net, pixel accuracy of 0.9739 and mean Intersection over Union (mIoU) of 0.7025 were achieved. The results of this study can be used for monitoring hazardous facilities in rural areas and are expected to be used as basis for rural planning.

Comparative Research of Image Classification and Image Segmentation Methods for Mapping Rural Roads Using a High-resolution Satellite Image (고해상도 위성영상을 이용한 농촌 도로 매핑을 위한 영상 분류 및 영상 분할 방법 비교에 관한 연구)

  • CHOUNG, Yun-Jae;GU, Bon-Yup
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.24 no.3
    • /
    • pp.73-82
    • /
    • 2021
  • Rural roads are the significant infrastructure for developing and managing the rural areas, hence the utilization of the remote sensing datasets for managing the rural roads is necessary for expanding the rural transportation infrastructure and improving the life quality of the rural residents. In this research, the two different methods such as image classification and image segmentation were compared for mapping the rural road based on the given high-resolution satellite image acquired in the rural areas. In the image classification method, the deep learning with the multiple neural networks was employed to the given high-resolution satellite image for generating the object classification map, then the rural roads were mapped by extracting the road objects from the generated object classification map. In the image segmentation method, the multiresolution segmentation was employed to the same satellite image for generating the segment image, then the rural roads were mapped by merging the road objects located on the rural roads on the satellite image. We used the 100 checkpoints for assessing the accuracy of the two rural roads mapped by the different methods and drew the following conclusions. The image segmentation method had the better performance than the image classification method for mapping the rural roads using the give satellite image, because some of the rural roads mapped by the image classification method were not identified due to the miclassification errors occurred in the object classification map, while all of the rural roads mapped by the image segmentation method were identified. However some of the rural roads mapped by the image segmentation method also had the miclassfication errors due to some rural road segments including the non-rural road objects. In future research the object-oriented classification or the convolutional neural networks widely used for detecting the precise objects from the image sources would be used for improving the accuracy of the rural roads using the high-resolution satellite image.

Semantic Segmentation of Drone Imagery Using Deep Learning for Seagrass Habitat Monitoring (잘피 서식지 모니터링을 위한 딥러닝 기반의 드론 영상 의미론적 분할)

  • Jeon, Eui-Ik;Kim, Seong-Hak;Kim, Byoung-Sub;Park, Kyung-Hyun;Choi, Ock-In
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.2_1
    • /
    • pp.199-215
    • /
    • 2020
  • A seagrass that is marine vascular plants plays an important role in the marine ecosystem, so periodic monitoring ofseagrass habitatsis being performed. Recently, the use of dronesthat can easily acquire very high-resolution imagery is increasing to efficiently monitor seagrass habitats. And deep learning based on a convolutional neural network has shown excellent performance in semantic segmentation. So, studies applied to deep learning models have been actively conducted in remote sensing. However, the segmentation accuracy was different due to the hyperparameter, various deep learning models and imagery. And the normalization of the image and the tile and batch size are also not standardized. So,seagrass habitats were segmented from drone-borne imagery using a deep learning that shows excellent performance in this study. And it compared and analyzed the results focused on normalization and tile size. For comparison of the results according to the normalization, tile and batch size, a grayscale image and grayscale imagery converted to Z-score and Min-Max normalization methods were used. And the tile size isincreased at a specific interval while the batch size is allowed the memory size to be used as much as possible. As a result, IoU was 0.26 ~ 0.4 higher than that of Z-score normalized imagery than other imagery. Also, it wasfound that the difference to 0.09 depending on the tile and batch size. The results were different according to the normalization, tile and batch. Therefore, this experiment found that these factors should have a suitable decision process.

Evaluation of Transfer Learning in Gastroscopy Image Classification using Convolutional Neual Network (합성곱 신경망을 활용한 위내시경 이미지 분류에서 전이학습의 효용성 평가)

  • Park, Sung Jin;Kim, Young Jae;Park, Dong Kyun;Chung, Jun Won;Kim, Kwang Gi
    • Journal of Biomedical Engineering Research
    • /
    • v.39 no.5
    • /
    • pp.213-219
    • /
    • 2018
  • Stomach cancer is the most diagnosed cancer in Korea. When gastric cancer is detected early, the 5-year survival rate is as high as 90%. Gastroscopy is a very useful method for early diagnosis. But the false negative rate of gastric cancer in the gastroscopy was 4.6~25.8% due to the subjective judgment of the physician. Recently, the image classification performance of the image recognition field has been advanced by the convolutional neural network. Convolutional neural networks perform well when diverse and sufficient amounts of data are supported. However, medical data is not easy to access and it is difficult to gather enough high-quality data that includes expert annotations. So This paper evaluates the efficacy of transfer learning in gastroscopy classification and diagnosis. We obtained 787 endoscopic images of gastric endoscopy at Gil Medical Center, Gachon University. The number of normal images was 200, and the number of abnormal images was 587. The image size was reconstructed and normalized. In the case of the ResNet50 structure, the classification accuracy before and after applying the transfer learning was improved from 0.9 to 0.947, and the AUC was also improved from 0.94 to 0.98. In the case of the InceptionV3 structure, the classification accuracy before and after applying the transfer learning was improved from 0.862 to 0.924, and the AUC was also improved from 0.89 to 0.97. In the case of the VGG16 structure, the classification accuracy before and after applying the transfer learning was improved from 0.87 to 0.938, and the AUC was also improved from 0.89 to 0.98. The difference in the performance of the CNN model before and after transfer learning was statistically significant when confirmed by T-test (p < 0.05). As a result, transfer learning is judged to be an effective method of medical data that is difficult to collect good quality data.

Deep Learning based Color Restoration of Corrupted Black and White Facial Photos (딥러닝 기반 손상된 흑백 얼굴 사진 컬러 복원)

  • Woo, Shin Jae;Kim, Jong-Hyun;Lee, Jung;Song, Chang-Germ;Kim, Sun-Jeong
    • Journal of the Korea Computer Graphics Society
    • /
    • v.24 no.2
    • /
    • pp.1-9
    • /
    • 2018
  • In this paper, we propose a method to restore corrupted black and white facial images to color. Previous studies have shown that when coloring damaged black and white photographs, such as old ID photographs, the area around the damaged area is often incorrectly colored. To solve this problem, this paper proposes a method of restoring the damaged area of input photo first and then performing colorization based on the result. The proposed method consists of two steps: BEGAN (Boundary Equivalent Generative Adversarial Networks) model based restoration and CNN (Convolutional Neural Network) based coloring. Our method uses the BEGAN model, which enables a clearer and higher resolution image restoration than the existing methods using the DCGAN (Deep Convolutional Generative Adversarial Networks) model for image restoration, and performs colorization based on the restored black and white image. Finally, we confirmed that the experimental results of various types of facial images and masks can show realistic color restoration results in many cases compared with the previous studies.

Evaluation of Classification and Accuracy in Chest X-ray Images using Deep Learning with Convolution Neural Network (컨볼루션 뉴럴 네트워크 기반의 딥러닝을 이용한 흉부 X-ray 영상의 분류 및 정확도 평가)

  • Song, Ho-Jun;Lee, Eun-Byeol;Jo, Heung-Joon;Park, Se-Young;Kim, So-Young;Kim, Hyeon-Jeong;Hong, Joo-Wan
    • Journal of the Korean Society of Radiology
    • /
    • v.14 no.1
    • /
    • pp.39-44
    • /
    • 2020
  • The purpose of this study was learning about chest X-ray image classification and accuracy research through Deep Learning using big data technology with Convolution Neural Network. Normal 1,583 and Pneumonia 4,289 were used in chest X-ray images. The data were classified as train (88.8%), validation (0.2%) and test (11%). Constructed as Convolution Layer, Max pooling layer size 2×2, Flatten layer, and Image Data Generator. The number of filters, filter size, drop out, epoch, batch size, and loss function values were set when the Convolution layer were 3 and 4 respectively. The test data verification results showed that the predicted accuracy was 94.67% when the number of filters was 64-128-128-128, filter size 3×3, drop out 0.25, epoch 5, batch size 15, and loss function RMSprop was 4. In this study, the classification of chest X-ray Normal and Pneumonia was predictable with high accuracy, and it is believed to be of great help not only to chest X-ray images but also to other medical images.

Applicability Evaluation for Discharge Model Using Curve Number and Convolution Neural Network (Curve Number 및 Convolution Neural Network를 이용한 유출모형의 적용성 평가)

  • Song, Chul Min;Lee, Kwang Hyun
    • Ecology and Resilient Infrastructure
    • /
    • v.7 no.2
    • /
    • pp.114-125
    • /
    • 2020
  • Despite the various artificial neural networks that have been developed, most of the discharge models in previous studies have been developed using deep neural networks. This study aimed to develop a discharge model using a convolution neural network (CNN), which was used to solve classification problems. Furthermore, the applicability of CNN was evaluated. The photographs (pictures or images) for input data to CNN could not clearly show the characteristics of the study area as well as precipitation. Hence, the model employed in this study had to use numerical images. To solve the problem, the CN of NRCS was used to generate images as input data for the model. The generated images showed a good possibility of applicability as input data. Moreover, a new application of CN, which had been used only for discharge prediction, was proposed in this study. As a result of CNN training, the model was trained and generalized stably. Comparison between the actual and predicted values had an R2 of 0.79, which was relatively high. The model showed good performance in terms of the Pearson correlation coefficient (0.84), the Nash-Sutcliffe efficiency (NSE) (0.63), and the root mean square error (24.54 ㎥/s).