• Title/Summary/Keyword: GAN(Generative Adversarial Network

Search Result 176, Processing Time 0.025 seconds

Crack Detection on Bridge Deck Using Generative Adversarial Networks and Deep Learning (적대적 생성 신경망과 딥러닝을 이용한 교량 상판의 균열 감지)

  • Ji, Bongjun
    • Journal of the Korean Recycled Construction Resources Institute
    • /
    • v.9 no.3
    • /
    • pp.303-310
    • /
    • 2021
  • Cracks in bridges are important factors that indicate the condition of bridges and should be monitored periodically. However, a visual inspection conducted by a human expert has problems in cost, time, and reliability. Therefore, in recent years, researches to apply a deep learning model are started to be conducted. Deep learning requires sufficient data on the situations to be predicted, but bridge crack data is relatively difficult to obtain. In particular, it is difficult to collect a large amount of crack data in a specific situation because the shape of bridge cracks may vary depending on the bridge's design, location, and construction method. This study developed a crack detection model that generates and trains insufficient crack data through a Generative Adversarial Network. GAN successfully generated data statistically similar to the given crack data, and accordingly, crack detection was possible with about 3% higher accuracy when using the generated image than when the generated image was not used. This approach is expected to effectively improve the performance of the detection model as it is applied when crack detection on bridges is required, though there is not enough data, also when there is relatively little or much data f or one class.

Generation of optical fringe patterns using deep learning (딥러닝을 이용한 광학적 프린지 패턴의 생성)

  • Kang, Ji-Won;Kim, Dong-Wook;Seo, Young-Ho
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.12
    • /
    • pp.1588-1594
    • /
    • 2020
  • In this paper, we discuss a data balancing method for learning a neural network that generates digital holograms using a deep neural network (DNN). Deep neural networks are based on deep learning (DL) technology and use a generative adversarial network (GAN) series. The fringe pattern, which is the basic unit of a hologram to be created through a deep neural network, has very different data types depending on the hologram plane and the position of the object. However, because the criteria for classifying the data are not clear, an imbalance in the training data may occur. The imbalance of learning data acts as a factor of instability in learning. Therefore, it presents a method for classifying and balancing data for which the classification criteria are not clear. And it shows that learning is stabilized through this.

Eyeglass Remover Network based on a Synthetic Image Dataset

  • Kang, Shinjin;Hahn, Teasung
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.4
    • /
    • pp.1486-1501
    • /
    • 2021
  • The removal of accessories from the face is one of the essential pre-processing stages in the field of face recognition. However, despite its importance, a robust solution has not yet been provided. This paper proposes a network and dataset construction methodology to remove only the glasses from facial images effectively. To obtain an image with the glasses removed from an image with glasses by the supervised learning method, a network that converts them and a set of paired data for training is required. To this end, we created a large number of synthetic images of glasses being worn using facial attribute transformation networks. We adopted the conditional GAN (cGAN) frameworks for training. The trained network converts the in-the-wild face image with glasses into an image without glasses and operates stably even in situations wherein the faces are of diverse races and ages and having different styles of glasses.

Planetary Long-Range Deep 2D Global Localization Using Generative Adversarial Network (생성적 적대 신경망을 이용한 행성의 장거리 2차원 깊이 광역 위치 추정 방법)

  • Ahmed, M.Naguib;Nguyen, Tuan Anh;Islam, Naeem Ul;Kim, Jaewoong;Lee, Sukhan
    • The Journal of Korea Robotics Society
    • /
    • v.13 no.1
    • /
    • pp.26-30
    • /
    • 2018
  • Planetary global localization is necessary for long-range rover missions in which communication with command center operator is throttled due to the long distance. There has been number of researches that address this problem by exploiting and matching rover surroundings with global digital elevation maps (DEM). Using conventional methods for matching, however, is challenging due to artifacts in both DEM rendered images, and/or rover 2D images caused by DEM low resolution, rover image illumination variations and small terrain features. In this work, we use train CNN discriminator to match rover 2D image with DEM rendered images using conditional Generative Adversarial Network architecture (cGAN). We then use this discriminator to search an uncertainty bound given by visual odometry (VO) error bound to estimate rover optimal location and orientation. We demonstrate our network capability to learn to translate rover image into DEM simulated image and match them using Devon Island dataset. The experimental results show that our proposed approach achieves ~74% mean average precision.

A study on speech disentanglement framework based on adversarial learning for speaker recognition (화자 인식을 위한 적대학습 기반 음성 분리 프레임워크에 대한 연구)

  • Kwon, Yoohwan;Chung, Soo-Whan;Kang, Hong-Goo
    • The Journal of the Acoustical Society of Korea
    • /
    • v.39 no.5
    • /
    • pp.447-453
    • /
    • 2020
  • In this paper, we propose a system to extract effective speaker representations from a speech signal using a deep learning method. Based on the fact that speech signal contains identity unrelated information such as text content, emotion, background noise, and so on, we perform a training such that the extracted features only represent speaker-related information but do not represent speaker-unrelated information. Specifically, we propose an auto-encoder based disentanglement method that outputs both speaker-related and speaker-unrelated embeddings using effective loss functions. To further improve the reconstruction performance in the decoding process, we also introduce a discriminator popularly used in Generative Adversarial Network (GAN) structure. Since improving the decoding capability is helpful for preserving speaker information and disentanglement, it results in the improvement of speaker verification performance. Experimental results demonstrate the effectiveness of our proposed method by improving Equal Error Rate (EER) on benchmark dataset, Voxceleb1.

Voice Conversion using Generative Adversarial Nets conditioned by Phonetic Posterior Grams (Phonetic Posterior Grams에 의해 조건화된 적대적 생성 신경망을 사용한 음성 변환 시스템)

  • Lim, Jin-su;Kang, Cheon-seong;Kim, Dong-Ha;Kim, Kyung-sup
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2018.10a
    • /
    • pp.369-372
    • /
    • 2018
  • This paper suggests non-parallel-voice-conversion network conversing voice between unmapped voice pair as source voice and target voice. Conventional voice conversion researches used learning methods that minimize spectrogram's distance error. Not only these researches have some problem that is lost spectrogram resolution by methods averaging pixels. But also have used parallel data that is hard to collect. This research uses PPGs that is input voice's phonetic data and a GAN learning method to generate more clear voices. To evaluate the suggested method, we conduct MOS test with GMM based Model. We found that the performance is improved compared to the conventional methods.

  • PDF

Land Use and Land Cover Mapping from Kompsat-5 X-band Co-polarized Data Using Conditional Generative Adversarial Network

  • Jang, Jae-Cheol;Park, Kyung-Ae
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.1
    • /
    • pp.111-126
    • /
    • 2022
  • Land use and land cover (LULC) mapping is an important factor in geospatial analysis. Although highly precise ground-based LULC monitoring is possible, it is time consuming and costly. Conversely, because the synthetic aperture radar (SAR) sensor is an all-weather sensor with high resolution, it could replace field-based LULC monitoring systems with low cost and less time requirement. Thus, LULC is one of the major areas in SAR applications. We developed a LULC model using only KOMPSAT-5 single co-polarized data and digital elevation model (DEM) data. Twelve HH-polarized images and 18 VV-polarized images were collected, and two HH-polarized images and four VV-polarized images were selected for the model testing. To train the LULC model, we applied the conditional generative adversarial network (cGAN) method. We used U-Net combined with the residual unit (ResUNet) model to generate the cGAN method. When analyzing the training history at 1732 epochs, the ResUNet model showed a maximum overall accuracy (OA) of 93.89 and a Kappa coefficient of 0.91. The model exhibited high performance in the test datasets with an OA greater than 90. The model accurately distinguished water body areas and showed lower accuracy in wetlands than in the other LULC types. The effect of the DEM on the accuracy of LULC was analyzed. When assessing the accuracy with respect to the incidence angle, owing to the radar shadow caused by the side-looking system of the SAR sensor, the OA tended to decrease as the incidence angle increased. This study is the first to use only KOMPSAT-5 single co-polarized data and deep learning methods to demonstrate the possibility of high-performance LULC monitoring. This study contributes to Earth surface monitoring and the development of deep learning approaches using the KOMPSAT-5 data.

A Design of Behavior Recognition method through GAN-based skeleton data generation (GAN 기반 관절 데이터 생성을 통한 행동 인식 방법 설계)

  • Kim, Jinah;Moon, Nammee
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2022.11a
    • /
    • pp.592-593
    • /
    • 2022
  • 다중 데이터 기반의 행동 인식 과정에서 데이터 수집 반경이 비교적 제한되는 영상 데이터의 결측에 대한 보완이 요구된다. 본 논문에서는 6축 센서 데이터를 이용하여 결측된 영상 데이터를 생성함으로써 행동 인식의 성능을 개선하는 방법을 제안한다. 가속도와 자이로 센서로부터 수집된 행동 데이터를 이용하여 GAN(Generative Adversarial Network)을 통해 영상에서의 관절(Skeleton) 움직임에 대한 데이터를 생성하고자 한다. 이를 위해 DeepLabCut 기반 모델 학습을 통해 관절 좌표를 추출하며, 전처리된 센서 시퀀스 데이터를 가지고 GRU 기반 GAN 모델을 통해 관절 좌표에 대한 영상 시퀀스 데이터를 생성한다. 생성된 영상 시퀀스 데이터는 영상 데이터의 결측이 발생했을 때 대신 행동 인식 모델의 입력값으로 활용될 수 있어 성능 향상을 기대할 수 있다.

Comparison of CNN and GAN-based Deep Learning Models for Ground Roll Suppression (그라운드-롤 제거를 위한 CNN과 GAN 기반 딥러닝 모델 비교 분석)

  • Sangin Cho;Sukjoon Pyun
    • Geophysics and Geophysical Exploration
    • /
    • v.26 no.2
    • /
    • pp.37-51
    • /
    • 2023
  • The ground roll is the most common coherent noise in land seismic data and has an amplitude much larger than the reflection event we usually want to obtain. Therefore, ground roll suppression is a crucial step in seismic data processing. Several techniques, such as f-k filtering and curvelet transform, have been developed to suppress the ground roll. However, the existing methods still require improvements in suppression performance and efficiency. Various studies on the suppression of ground roll in seismic data have recently been conducted using deep learning methods developed for image processing. In this paper, we introduce three models (DnCNN (De-noiseCNN), pix2pix, and CycleGAN), based on convolutional neural network (CNN) or conditional generative adversarial network (cGAN), for ground roll suppression and explain them in detail through numerical examples. Common shot gathers from the same field were divided into training and test datasets to compare the algorithms. We trained the models using the training data and evaluated their performances using the test data. When training these models with field data, ground roll removed data are required; therefore, the ground roll is suppressed by f-k filtering and used as the ground-truth data. To evaluate the performance of the deep learning models and compare the training results, we utilized quantitative indicators such as the correlation coefficient and structural similarity index measure (SSIM) based on the similarity to the ground-truth data. The DnCNN model exhibited the best performance, and we confirmed that other models could also be applied to suppress the ground roll.

Generation of contrast enhanced computed tomography image using deep learning network

  • Woo, Sang-Keun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.3
    • /
    • pp.41-47
    • /
    • 2019
  • In this paper, we propose a application of conditional generative adversarial network (cGAN) for generation of contrast enhanced computed tomography (CT) image. Two types of CT data which were the enhanced and non-enhanced were used and applied by the histogram equalization for adjusting image intensities. In order to validate the generation of contrast enhanced CT data, the structural similarity index measurement (SSIM) was performed. Prepared generated contrast CT data were analyzed the statistical analysis using paired sample t-test. In order to apply the optimized algorithm for the lymph node cancer, they were calculated by short to long axis ratio (S/L) method. In the case of the model trained with CT data and their histogram equalized SSIM were $0.905{\pm}0.048$ and $0.908{\pm}0.047$. The tumor S/L of generated contrast enhanced CT data were validated similar to the ground truth when they were compared to scanned contrast enhanced CT data. It is expected that advantages of Generated contrast enhanced CT data based on deep learning are a cost-effective and less radiation exposure as well as further anatomical information with non-enhanced CT data.