• Title/Summary/Keyword: generative learning

Search Result 285, Processing Time 0.029 seconds

Extraction of the OLED Device Parameter based on Randomly Generated Monte Carlo Simulation with Deep Learning (무작위 생성 심층신경망 기반 유기발광다이오드 흑점 성장가속 전산모사를 통한 소자 변수 추출)

  • You, Seung Yeol;Park, Il-Hoo;Kim, Gyu-Tae
    • Journal of the Semiconductor & Display Technology
    • /
    • v.20 no.3
    • /
    • pp.131-135
    • /
    • 2021
  • Numbers of studies related to optimization of design of organic light emitting diodes(OLED) through machine learning are increasing. We propose the generative method of the image to assess the performance of the device combining with machine learning technique. Principle parameter regarding dark spot growth mechanism of the OLED can be the key factor to determine the long-time performance. Captured images from actual device and randomly generated images at specific time and initial pinhole state are fed into the deep neural network system. The simulation reinforced by the machine learning technique can predict the device parameters accurately and faster. Similarly, the inverse design using multiple layer perceptron(MLP) system can infer the initial degradation factors at manufacturing with given device parameter to feedback the design of manufacturing process.

Deep Learning based Human Recognition using Integration of GAN and Spatial Domain Techniques

  • Sharath, S;Rangaraju, HG
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.8
    • /
    • pp.127-136
    • /
    • 2021
  • Real-time human recognition is a challenging task, as the images are captured in an unconstrained environment with different poses, makeups, and styles. This limitation is addressed by generating several facial images with poses, makeup, and styles with a single reference image of a person using Generative Adversarial Networks (GAN). In this paper, we propose deep learning-based human recognition using integration of GAN and Spatial Domain Techniques. A novel concept of human recognition based on face depiction approach by generating several dissimilar face images from single reference face image using Domain Transfer Generative Adversarial Networks (DT-GAN) combined with feature extraction techniques such as Local Binary Pattern (LBP) and Histogram is deliberated. The Euclidean Distance (ED) is used in the matching section for comparison of features to test the performance of the method. A database of millions of people with a single reference face image per person, instead of multiple reference face images, is created and saved on the centralized server, which helps to reduce memory load on the centralized server. It is noticed that the recognition accuracy is 100% for smaller size datasets and a little less accuracy for larger size datasets and also, results are compared with present methods to show the superiority of proposed method.

Study on 2D Sprite *3.Generation Using the Impersonator Network

  • Yongjun Choi;Beomjoo Seo;Shinjin Kang;Jongin Choi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.7
    • /
    • pp.1794-1806
    • /
    • 2023
  • This study presents a method for capturing photographs of users as input and converting them into 2D character animation sprites using a generative adversarial network-based artificial intelligence network. Traditionally, 2D character animations have been created by manually creating an entire sequence of sprite images, which incurs high development costs. To address this issue, this study proposes a technique that combines motion videos and sample 2D images. In the 2D sprite generation process that uses the proposed technique, a sequence of images is extracted from real-life images captured by the user, and these are combined with character images from within the game. Our research aims to leverage cutting-edge deep learning-based image manipulation techniques, such as the GAN-based motion transfer network (impersonator) and background noise removal (U2 -Net), to generate a sequence of animation sprites from a single image. The proposed technique enables the creation of diverse animations and motions just one image. By utilizing these advancements, we focus on enhancing productivity in the game and animation industry through improved efficiency and streamlined production processes. By employing state-of-the-art techniques, our research enables the generation of 2D sprite images with various motions, offering significant potential for boosting productivity and creativity in the industry.

Application of transfer learning to develop radar-based rainfall prediction model with GAN(Generative Adversarial Network) for multiple dam domains (다중 댐 유역에 대한 강우예측모델 개발을 위한 전이학습 기법의 적용)

  • Choi, Suyeon;Kim, Yeonjoo
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2022.05a
    • /
    • pp.61-61
    • /
    • 2022
  • 최근 머신러닝 기술의 발달에 따라 이를 활용한 레이더 자료기반 강우예측기법이 활발히 개발되고 있다. 기존 머신러닝을 이용한 강우예측모델 개발 관련 연구는 주로 한 지역에 대해 수행되며, 데이터 기반으로 훈련되는 머신러닝 기법의 특성상 개발된 모델이 훈련된 지역에 대해서만 좋은 성능을 보인다는 한계점이 존재한다. 이러한 한계점을 해결하기 위해 사전 훈련된 모델을 이용하여 새로운 데이터에 대해 모델을 훈련하는 전이학습 기법 (transfer learning)을 적용하여 여러 유역에 대한 강우예측모델을 개발하고자 하였다. 본 연구에서는 사전 훈련된 강우예측 모델로 생성적 적대 신경망 기반 기법(Generative Adversarial Network, GAN)을 이용한 미래 강우예측모델을 사용하였다. 해당 모델은 기상청에서 제공된 2014년~2017년 여름의 레이더 이미지 자료를 이용하여 초단기, 단기 강우예측을 수행하도록 학습시켰으며, 2018년 레이더 이미지 자료를 이용한 단기강우예측 모의에서 좋은 성능을 보였다. 본 연구에서는 훈련된 모델을 이용해 새로운 댐 유역(안동댐, 충주댐)에 대한 강우예측모델을 개발하기 위해 여러 전이학습 기법을 적용하고, 그 결과를 비교하였다. 결과를 통해 새로운 데이터로 처음부터 훈련시킨 모델보다 전이학습 기법을 사용하였을 때 좋은 성능을 보이는 것을 확인하였으며, 이를 통해 여러 댐 유역에 대한 모델 개발 시 전이학습 기법이 효율적으로 적용될 수 있음을 확인하였다.

  • PDF

Effects of Visual Working Memory Capacity and the Type and Contents of Graphic Annotation on Multimedia English Learning (시공간 작업기억 용량과 그림 자료의 유형과 내용이 초등학생의 영어 단어 학습에 미치는 영향)

  • Do, Kyung-Soo;Cha, Yu-Young
    • Korean Journal of Cognitive Science
    • /
    • v.19 no.4
    • /
    • pp.369-396
    • /
    • 2008
  • The purpose of this article is to investigate the effect of visual working memory, the types and contents of graphic annotations on English learning. The participants of the experiments were 5th and 6th graders. The result showed that animation was effective only in the word test for children with large visual working memory, whereas text-only-annotation yielded better performance in the comprehension test in Experiment 1. The effect of relevance of animations was tested in Experiment 2. Context-relevant-animations yielded better comprehension than the animations denoting the typical meaning, whereas the opposite pattern was reported in the word test. The result of the two experiments was interpreted in terms of cognitive load theory and the generative theory of multimedia learning.

  • PDF

Perceptual Photo Enhancement with Generative Adversarial Networks (GAN 신경망을 통한 자각적 사진 향상)

  • Que, Yue;Lee, Hyo Jong
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2019.05a
    • /
    • pp.522-524
    • /
    • 2019
  • In spite of a rapid development in the quality of built-in mobile cameras, their some physical restrictions hinder them to achieve the satisfactory results of digital single lens reflex (DSLR) cameras. In this work we propose an end-to-end deep learning method to translate ordinary images by mobile cameras into DSLR-quality photos. The method is based on the framework of generative adversarial networks (GANs) with several improvements. First, we combined the U-Net with DenseNet and connected dense block (DB) in terms of U-Net. The Dense U-Net acts as the generator in our GAN model. Then, we improved the perceptual loss by using the VGG features and pixel-wise content, which could provide stronger supervision for contrast enhancement and texture recovery.

Face Recognition Research Based on Multi-Layers Residual Unit CNN Model

  • Zhang, Ruyang;Lee, Eung-Joo
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.11
    • /
    • pp.1582-1590
    • /
    • 2022
  • Due to the situation of the widespread of the coronavirus, which causes the problem of lack of face image data occluded by masks at recent time, in order to solve the related problems, this paper proposes a method to generate face images with masks using a combination of generative adversarial networks and spatial transformation networks based on CNN model. The system we proposed in this paper is based on the GAN, combined with multi-scale convolution kernels to extract features at different details of the human face images, and used Wasserstein divergence as the measure of the distance between real samples and synthetic samples in order to optimize Generator performance. Experiments show that the proposed method can effectively put masks on face images with high efficiency and fast reaction time and the synthesized human face images are pretty natural and real.

Object Tracking Based on Weighted Local Sub-space Reconstruction Error

  • Zeng, Xianyou;Xu, Long;Hu, Shaohai;Zhao, Ruizhen;Feng, Wanli
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.2
    • /
    • pp.871-891
    • /
    • 2019
  • Visual tracking is a challenging task that needs learning an effective model to handle the changes of target appearance caused by factors such as pose variation, illumination change, occlusion and motion blur. In this paper, a novel tracking algorithm based on weighted local sub-space reconstruction error is presented. First, accounting for the appearance changes in the tracking process, a generative weight calculation method based on structural reconstruction error is proposed. Furthermore, a template update scheme of occlusion-aware is introduced, in which we reconstruct a new template instead of simply exploiting the best observation for template update. The effectiveness and feasibility of the proposed algorithm are verified by comparing it with some state-of-the-art algorithms quantitatively and qualitatively.

Frontal Face Generation Algorithm from Multi-view Images Based on Generative Adversarial Network

  • Heo, Young- Jin;Kim, Byung-Gyu;Roy, Partha Pratim
    • Journal of Multimedia Information System
    • /
    • v.8 no.2
    • /
    • pp.85-92
    • /
    • 2021
  • In a face, there is much information of person's identity. Because of this property, various tasks such as expression recognition, identity recognition and deepfake have been actively conducted. Most of them use the exact frontal view of the given face. However, various directions of the face can be observed rather than the exact frontal image in real situation. The profile (side view) lacks information when comparing with the frontal view image. Therefore, if we can generate the frontal face from other directions, we can obtain more information on the given face. In this paper, we propose a combined style model based the conditional generative adversarial network (cGAN) for generating the frontal face from multi-view images that consist of characteristics that not only includes the style around the face (hair and beard) but also detailed areas (eye, nose, and mouth).

SkelGAN: A Font Image Skeletonization Method

  • Ko, Debbie Honghee;Hassan, Ammar Ul;Majeed, Saima;Choi, Jaeyoung
    • Journal of Information Processing Systems
    • /
    • v.17 no.1
    • /
    • pp.1-13
    • /
    • 2021
  • In this research, we study the problem of font image skeletonization using an end-to-end deep adversarial network, in contrast with the state-of-the-art methods that use mathematical algorithms. Several studies have been concerned with skeletonization, but a few have utilized deep learning. Further, no study has considered generative models based on deep neural networks for font character skeletonization, which are more delicate than natural objects. In this work, we take a step closer to producing realistic synthesized skeletons of font characters. We consider using an end-to-end deep adversarial network, SkelGAN, for font-image skeletonization, in contrast with the state-of-the-art methods that use mathematical algorithms. The proposed skeleton generator is proved superior to all well-known mathematical skeletonization methods in terms of character structure, including delicate strokes, serifs, and even special styles. Experimental results also demonstrate the dominance of our method against the state-of-the-art supervised image-to-image translation method in font character skeletonization task.