• Title/Summary/Keyword: GAN(Generative Adversarial Networks)

Search Result 116, Processing Time 0.025 seconds

A Study on Image Quality Improvement for 3D Pagoda Restoration (3D 탑복원을 위한 화질 개선에 관한 연구)

  • Kim, Beom Jun-Ji;Lee, Hyun-woo;Kim, Ki-hyeop;Kim, Eun-ji;Kim, Young-jin;Lee, Byong-Kwon
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2022.07a
    • /
    • pp.145-147
    • /
    • 2022
  • 본 논문에서는 훼손되어 식별할 수 없는 탑 이미지를 비롯해 낮은 해상도의 탑 이미지를 개선하기 위해 우리는 탑 이미지의 화질 개선을 인공지능을 이용하여 빠르게 개선을 해 보고자 한다. 최근에 Generative Adversarial Networks(GANS) 알고리즘에서 SrGAN 알고리즘이 나오면서 이미지 생성, 이미지 복원, 해상도 변화 분야가 지속해서 발전하고 있다. 이에 본 연구에서는 다양한 GAN 알고리즘을 화질 개선에 적용해 보았다. 탑 이미지에 GAN 알고리즘 중 SrGan을 적용하였으며 실험한 결과 Srgan 알고리즘은 학습이 진행되었으며, 낮은 해상도의 탑 이미지가 높은 해상도, 초고해상도 이미지가 생성되는 것을 확인했다.

  • PDF

Deep Learning Music Genre Classification System Model Improvement Using Generative Adversarial Networks (GAN) (생성적 적대 신경망(GAN)을 이용한 딥러닝 음악 장르 분류 시스템 모델 개선)

  • Bae, Jun
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.7
    • /
    • pp.842-848
    • /
    • 2020
  • Music markets have entered the era of streaming. In order to select and propose music that suits the taste of music consumers, there is an active demand and research on an automatic music genre classification system. We propose a method to improve the accuracy of genre unclassified songs, which was a lack of the previous system, by using a generative adversarial network (GAN) to further develop the automatic voting system for deep learning music genre using Softmax proposed in the previous paper. In the previous study, if the spectrogram of the song was ambiguous to grasp the genre of the song, it was forced to leave it as an unclassified song. In this paper, we proposed a system that increases the accuracy of genre classification of unclassified songs by converting the spectrogram of unclassified songs into an easy-to-read spectrogram using GAN. And the result of the experiment was able to derive an excellent result compared to the existing method.

Data Augmentation Effect of StyleGAN-Generated Images in Deep Neural Network Training for Medical Image Classification (의료영상 분류를 위한 심층신경망 훈련에서 StyleGAN 합성 영상의 데이터 증강 효과 분석)

  • Hansang Lee;Arha Woo;Helen Hong
    • Journal of the Korea Computer Graphics Society
    • /
    • v.30 no.4
    • /
    • pp.19-29
    • /
    • 2024
  • In this paper, we examine the effectiveness of StyleGAN-generated images for data augmentation in training deep neural networks for medical image classification. We apply StyleGAN data augmentation to train VGG-16 networks for pneumonia diagnosis from chest X-ray images and focal liver lesion classification from abdominal CT images. Through quantitative and qualitative analyses, our experiments reveal that StyleGAN data augmentation expands the outer class boundaries in the feature space. Thanks to this expansion characteristics, the StyleGAN data augmentation can enhance classification performance when properly combined with real training images.

A Study on the Synthetic ECG Generation for User Recognition (사용자 인식을 위한 가상 심전도 신호 생성 기술에 관한 연구)

  • Kim, Min Gu;Kim, Jin Su;Pan, Sung Bum
    • Smart Media Journal
    • /
    • v.8 no.4
    • /
    • pp.33-37
    • /
    • 2019
  • Because the ECG signals are time-series data acquired as time elapses, it is important to obtain comparative data the same in size as the enrolled data every time. This paper suggests a network model of GAN (Generative Adversarial Networks) based on an auxiliary classifier to generate synthetic ECG signals which may address the different data size issues. The Cosine similarity and Cross-correlation are used to examine the similarity of synthetic ECG signals. The analysis shows that the Average Cosine similarity was 0.991 and the Average Euclidean distance similarity based on cross-correlation was 0.25: such results indicate that data size difference issue can be resolved while the generated synthetic ECG signals, similar to real ECG signals, can create synthetic data even when the registered data are not the same as the comparative data in size.

Interactive Web using CycleGAN (CycleGAN을 이용한 인터랙티브 웹페이지)

  • Kim, Jiwon;Jung, Haejung;Kim, Dongho
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • fall
    • /
    • pp.280-282
    • /
    • 2021
  • 최근에 딥러닝 기술인 GAN (Generative Adversarial Networks) 연구는 Image-to-Image translation 분야에서 활발하게 이뤄지고 있다. 이러한 기술을 바탕으로 사용자에게 편의와 재미를 제공하는 서비스가 애플리케이션 및 웹사이트의 형태로 개발되고 있다. 이에 본 논문은 CycleGAN 모델을 사용하여 이미지를 변환하고, 이를 인터랙티브 웹페이지를 통해 사용자와 실시간으로 상호작용하며 결과 이미지를 제공할 수 있는 방법을 연구하였다. 모델을 구현하기 위해 Tensorflow 및 Keras를 사용하였고, Django와 HTML5, CSS, JavaScript를 사용하여 웹사이트를 제작하였다.

  • PDF

GAN-based camouflage pattern generation parameter optimization system for improving assimilation rate with environment (야생 환경과의 동화율 개선을 위한 GAN 알고리즘 기반 위장 패턴 생성 파라미터 최적화 시스템)

  • Park, JunHyeok;Park, Seungmin;Cho, Dae-Soo
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2022.07a
    • /
    • pp.511-512
    • /
    • 2022
  • 동물무늬는 서식지에 따라 야생에서 천적으로부터 살아남을 수 있는 중요한 역할을 한다. 동물무늬의 역할 중 하나인 자연과 야생 환경에서 천적의 눈을 피해 위장하는 기능이 있기 때문인데 본 논문에서는 기존 위장무늬의 개선을 위한 GAN 알고리즘 기반 위장 패턴 생성모델을 제안한다. 이 모델은 단순히 색상만을 사용하여 위장무늬의 윤곽선을 Blur 처리를 해서 사람의 관측을 흐리게 만드는 기존의 모델의 단순함을 보완하여 GAN 알고리즘의 활용기술인 Deep Dream을 활용하여 경사 상승법을 통해 특정 층의 필터 값을 조절하여 원하는 부분에 대한 구분되는 패턴을 생성할 수 있어 색뿐만 아니라 위장의 기능이 있는 동물무늬와 섞어 자연과 야생 환경에서 더욱 동화율이 높아진 위장 패턴을 생성하고자 한다.

  • PDF

Comparison of online video(OTT) content production technology based on artificial intelligence customized recommendation service (인공지능 맞춤 추천서비스 기반 온라인 동영상(OTT) 콘텐츠 제작 기술 비교)

  • CHUN, Sanghun;SHIN, Seoung-Jung
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.21 no.3
    • /
    • pp.99-105
    • /
    • 2021
  • In addition to the OTT video production service represented by Nexflix and YouTube, a personalized recommendation system for content with artificial intelligence has become common. YouTube's personalized recommendation service system consists of two neural networks, one neural network consisting of a recommendation candidate generation model and the other consisting of a ranking network. Netflix's video recommendation system consists of two data classification systems, divided into content-based filtering and collaborative filtering. As the online platform-led content production is activated by the Corona Pandemic, the field of virtual influencers using artificial intelligence is emerging. Virtual influencers are produced with GAN (Generative Adversarial Networks) artificial intelligence, and are unsupervised learning algorithms in which two opposing systems compete with each other. This study also researched the possibility of developing AI platform based on individual recommendation and virtual influencer (metabus) as a core content of OTT in the future.

Study on hole-filling technique of motion capture images using GANs (Generative Adversarial Networks) (GANs(Generative Adversarial Networks)를 활용한 모션캡처 이미지의 hole-filling 기법 연구)

  • Shin, Kwang-Seong;Shin, Seong-Yoon
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2019.05a
    • /
    • pp.160-161
    • /
    • 2019
  • As a method for modeling a three-dimensional object, there are a method using a 3D scanner, a method using a motion capture system, and a method using a Kinect system. Through this method, a portion that is not captured due to occlusion occurs in the process of creating a three-dimensional object. In order to implement a perfect three-dimensional object, it is necessary to arbitrarily fill the obscured part. There is a technique to fill the unexposed part by various image processing methods. In this study, we propose a method using GANs, which is the latest trend of unsupervised machine learning, as a method for more natural hole-filling.

  • PDF

Image-to-Image Translation Based on U-Net with R2 and Attention (R2와 어텐션을 적용한 유넷 기반의 영상 간 변환에 관한 연구)

  • Lim, So-hyun;Chun, Jun-chul
    • Journal of Internet Computing and Services
    • /
    • v.21 no.4
    • /
    • pp.9-16
    • /
    • 2020
  • In the Image processing and computer vision, the problem of reconstructing from one image to another or generating a new image has been steadily drawing attention as hardware advances. However, the problem of computer-generated images also continues to emerge when viewed with human eyes because it is not natural. Due to the recent active research in deep learning, image generating and improvement problem using it are also actively being studied, and among them, the network called Generative Adversarial Network(GAN) is doing well in the image generating. Various models of GAN have been presented since the proposed GAN, allowing for the generation of more natural images compared to the results of research in the image generating. Among them, pix2pix is a conditional GAN model, which is a general-purpose network that shows good performance in various datasets. pix2pix is based on U-Net, but there are many networks that show better performance among U-Net based networks. Therefore, in this study, images are generated by applying various networks to U-Net of pix2pix, and the results are compared and evaluated. The images generated through each network confirm that the pix2pix model with Attention, R2, and Attention-R2 networks shows better performance than the existing pix2pix model using U-Net, and check the limitations of the most powerful network. It is suggested as a future study.

Comparison of GAN Deep Learning Methods for Underwater Optical Image Enhancement

  • Kim, Hong-Gi;Seo, Jung-Min;Kim, Soo Mee
    • Journal of Ocean Engineering and Technology
    • /
    • v.36 no.1
    • /
    • pp.32-40
    • /
    • 2022
  • Underwater optical images face various limitations that degrade the image quality compared with optical images taken in our atmosphere. Attenuation according to the wavelength of light and reflection by very small floating objects cause low contrast, blurry clarity, and color degradation in underwater images. We constructed an image data of the Korean sea and enhanced it by learning the characteristics of underwater images using the deep learning techniques of CycleGAN (cycle-consistent adversarial network), UGAN (underwater GAN), FUnIE-GAN (fast underwater image enhancement GAN). In addition, the underwater optical image was enhanced using the image processing technique of Image Fusion. For a quantitative performance comparison, UIQM (underwater image quality measure), which evaluates the performance of the enhancement in terms of colorfulness, sharpness, and contrast, and UCIQE (underwater color image quality evaluation), which evaluates the performance in terms of chroma, luminance, and saturation were calculated. For 100 underwater images taken in Korean seas, the average UIQMs of CycleGAN, UGAN, and FUnIE-GAN were 3.91, 3.42, and 2.66, respectively, and the average UCIQEs were measured to be 29.9, 26.77, and 22.88, respectively. The average UIQM and UCIQE of Image Fusion were 3.63 and 23.59, respectively. CycleGAN and UGAN qualitatively and quantitatively improved the image quality in various underwater environments, and FUnIE-GAN had performance differences depending on the underwater environment. Image Fusion showed good performance in terms of color correction and sharpness enhancement. It is expected that this method can be used for monitoring underwater works and the autonomous operation of unmanned vehicles by improving the visibility of underwater situations more accurately.