• Title/Summary/Keyword: DCGAN

Search Result 32, Processing Time 0.031 seconds

Photo-realistic Face Image Generation by DCGAN with error relearning (심층 적대적 생성 신경망의 오류 재학습을 이용한 얼굴 영상 생성 모델)

  • Ha, Yong-Wook;Hong, Dong-jin;Cha, Eui-Young
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2018.10a
    • /
    • pp.617-619
    • /
    • 2018
  • In this paper, We suggest a face image generating GAN model which is improved by an additive discriminator. This discriminator is trained to be specialized in preventing frequent mistake of generator. To verify the model suggested, we used $^*Inception$ score. We used 155,680 images of $^*celebA$ which is frontal face. We earned average 1.742p at Inception score and it is much better score compare to previous model.

  • PDF

A Study on Virtual Tooth Image Generation Using Deep Learning - Based on the number of learning (심층 학습을 활용한 가상 치아 이미지 생성 연구 -학습 횟수를 중심으로)

  • Bae, EunJeong;Jeong, Junho;Son, Yunsik;Lim, JoonYeon
    • Journal of Technologic Dentistry
    • /
    • v.42 no.1
    • /
    • pp.1-8
    • /
    • 2020
  • Purpose: Among the virtual teeth generated by Deep Convolutional Generative Adversarial Networks (DCGAN), the optimal data was analyzed for the number of learning. Methods: We extracted 50 mandibular first molar occlusal surfaces and trained 4,000 epoch with DCGAN. The learning screen was saved every 50 times and evaluated on a Likert 5-point scale according to five classification criteria. Results were analyzed by one-way ANOVA and tukey HSD post hoc analysis (α = 0.05). Results: It was the highest with 83.90±6.32 in the number of group3 (2,050-3,000) learning and statistically significant in the group1 (50-1,000) and the group2 (1,050-2,000). Conclusion: Since there is a difference in the optimal virtual tooth generation according to the number of learning, it is necessary to analyze the learning frequency section in various ways.

Assessment and Analysis of Fidelity and Diversity for GAN-based Medical Image Generative Model (GAN 기반 의료영상 생성 모델에 대한 품질 및 다양성 평가 및 분석)

  • Jang, Yoojin;Yoo, Jaejun;Hong, Helen
    • Journal of the Korea Computer Graphics Society
    • /
    • v.28 no.2
    • /
    • pp.11-19
    • /
    • 2022
  • Recently, various researches on medical image generation have been suggested, and it becomes crucial to accurately evaluate the quality and diversity of the generated medical images. For this purpose, the expert's visual turing test, feature distribution visualization, and quantitative evaluation through IS and FID are evaluated. However, there are few methods for quantitatively evaluating medical images in terms of fidelity and diversity. In this paper, images are generated by learning a chest CT dataset of non-small cell lung cancer patients through DCGAN and PGGAN generative models, and the performance of the two generative models are evaluated in terms of fidelity and diversity. The performance is quantitatively evaluated through IS and FID, which are one-dimensional score-based evaluation methods, and Precision and Recall, Improved Precision and Recall, which are two-dimensional score-based evaluation methods, and the characteristics and limitations of each evaluation method are also analyzed in medical imaging.

A Study on the Generation of Webtoons through Fine-Tuning of Diffusion Models (확산모델의 미세조정을 통한 웹툰 생성연구)

  • Kyungho Yu;Hyungju Kim;Jeongin Kim;Chanjun Chun;Pankoo Kim
    • Smart Media Journal
    • /
    • v.12 no.7
    • /
    • pp.76-83
    • /
    • 2023
  • This study proposes a method to assist webtoon artists in the process of webtoon creation by utilizing a pretrained Text-to-Image model to generate webtoon images from text. The proposed approach involves fine-tuning a pretrained Stable Diffusion model using a webtoon dataset transformed into the desired webtoon style. The fine-tuning process, using LoRA technique, completes in a quick training time of approximately 4.5 hours with 30,000 steps. The generated images exhibit the representation of shapes and backgrounds based on the input text, resulting in the creation of webtoon-like images. Furthermore, the quantitative evaluation using the Inception score shows that the proposed method outperforms DCGAN-based Text-to-Image models. If webtoon artists adopt the proposed Text-to-Image model for webtoon creation, it is expected to significantly reduce the time required for the creative process.

Deep Learning based Color Restoration of Corrupted Black and White Facial Photos (딥러닝 기반 손상된 흑백 얼굴 사진 컬러 복원)

  • Woo, Shin Jae;Kim, Jong-Hyun;Lee, Jung;Song, Chang-Germ;Kim, Sun-Jeong
    • Journal of the Korea Computer Graphics Society
    • /
    • v.24 no.2
    • /
    • pp.1-9
    • /
    • 2018
  • In this paper, we propose a method to restore corrupted black and white facial images to color. Previous studies have shown that when coloring damaged black and white photographs, such as old ID photographs, the area around the damaged area is often incorrectly colored. To solve this problem, this paper proposes a method of restoring the damaged area of input photo first and then performing colorization based on the result. The proposed method consists of two steps: BEGAN (Boundary Equivalent Generative Adversarial Networks) model based restoration and CNN (Convolutional Neural Network) based coloring. Our method uses the BEGAN model, which enables a clearer and higher resolution image restoration than the existing methods using the DCGAN (Deep Convolutional Generative Adversarial Networks) model for image restoration, and performs colorization based on the restored black and white image. Finally, we confirmed that the experimental results of various types of facial images and masks can show realistic color restoration results in many cases compared with the previous studies.

Synthetic Image Dataset Generation for Defense using Generative Adversarial Networks (국방용 합성이미지 데이터셋 생성을 위한 대립훈련신경망 기술 적용 연구)

  • Yang, Hunmin
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.22 no.1
    • /
    • pp.49-59
    • /
    • 2019
  • Generative adversarial networks(GANs) have received great attention in the machine learning field for their capacity to model high-dimensional and complex data distribution implicitly and generate new data samples from the model distribution. This paper investigates the model training methodology, architecture, and various applications of generative adversarial networks. Experimental evaluation is also conducted for generating synthetic image dataset for defense using two types of GANs. The first one is for military image generation utilizing the deep convolutional generative adversarial networks(DCGAN). The other is for visible-to-infrared image translation utilizing the cycle-consistent generative adversarial networks(CycleGAN). Each model can yield a great diversity of high-fidelity synthetic images compared to training ones. This result opens up the possibility of using inexpensive synthetic images for training neural networks while avoiding the enormous expense of collecting large amounts of hand-annotated real dataset.

A Study on Classification System using Generative Adversarial Networks (GAN을 활용한 분류 시스템에 관한 연구)

  • Bae, Sangjung;Lim, Byeongyeon;Jung, Jihak;Na, Chulhun;Jung, Hoekyung
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2019.05a
    • /
    • pp.338-340
    • /
    • 2019
  • Recently, the speed and size of data accumulation are increasing due to the development of networks. There are many difficulties in classifying these data. One of the difficulties is the difficulty of labeling. Labeling is usually done by people, but it is very difficult for everyone to understand the data in the same way and it is very difficult to label them on the same basis. In order to solve this problem, we implemented GAN to generate new image based on input image and to learn input data indirectly by using it for learning. This suggests that the accuracy of classification can be increased by increasing the number of learning data.

  • PDF

HiGANCNN: A Hybrid Generative Adversarial Network and Convolutional Neural Network for Glaucoma Detection

  • Alsulami, Fairouz;Alseleahbi, Hind;Alsaedi, Rawan;Almaghdawi, Rasha;Alafif, Tarik;Ikram, Mohammad;Zong, Weiwei;Alzahrani, Yahya;Bawazeer, Ahmed
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.9
    • /
    • pp.23-30
    • /
    • 2022
  • Glaucoma is a chronic neuropathy that affects the optic nerve which can lead to blindness. The detection and prediction of glaucoma become possible using deep neural networks. However, the detection performance relies on the availability of a large number of data. Therefore, we propose different frameworks, including a hybrid of a generative adversarial network and a convolutional neural network to automate and increase the performance of glaucoma detection. The proposed frameworks are evaluated using five public glaucoma datasets. The framework which uses a Deconvolutional Generative Adversarial Network (DCGAN) and a DenseNet pre-trained model achieves 99.6%, 99.08%, 99.4%, 98.69%, and 92.95% of classification accuracy on RIMONE, Drishti-GS, ACRIMA, ORIGA-light, and HRF datasets respectively. Based on the experimental results and evaluation, the proposed framework closely competes with the state-of-the-art methods using the five public glaucoma datasets without requiring any manually preprocessing step.

Detection of Black Screen in Video Wall Controller Using CNN (컨볼루션 신경망에 기반한 비디오 월 컨트롤러의 블랙 스크린 감지)

  • Kim, Sung-jin
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.524-526
    • /
    • 2021
  • As the video wall controller market is growing rapidly, issues that have not been addressed so far are raised. One of them is a phenomenon in which a black screen is displayed on a multi-screen. Black screen is displayed due to an error in the video being displayed in the video wall controller. Human intervention is inevitable to recognize and solve the black screen. However, it is impossible for the operator to monitor the multi-screen 24 hours a day. In this paper, we propose a model that detects the black screen being displayed on the video wall controller. We propose a CNN based architecture to detect a black screen.

  • PDF

Face Morphing Using Generative Adversarial Networks (Generative Adversarial Networks를 이용한 Face Morphing 기법 연구)

  • Han, Yoon;Kim, Hyoung Joong
    • Journal of Digital Contents Society
    • /
    • v.19 no.3
    • /
    • pp.435-443
    • /
    • 2018
  • Recently, with the explosive development of computing power, various methods such as RNN and CNN have been proposed under the name of Deep Learning, which solve many problems of Computer Vision have. The Generative Adversarial Network, released in 2014, showed that the problem of computer vision can be sufficiently solved in unsupervised learning, and the generation domain can also be studied using learned generators. GAN is being developed in various forms in combination with various models. Machine learning has difficulty in collecting data. If it is too large, it is difficult to refine the effective data set by removing the noise. If it is too small, the small difference becomes too big noise, and learning is not easy. In this paper, we apply a deep CNN model for extracting facial region in image frame to GAN model as a preprocessing filter, and propose a method to produce composite images of various facial expressions by stably learning with limited collection data of two persons.