• 제목/요약/키워드: generative learning

검색결과 293건 처리시간 0.026초

시공간 작업기억 용량과 그림 자료의 유형과 내용이 초등학생의 영어 단어 학습에 미치는 영향 (Effects of Visual Working Memory Capacity and the Type and Contents of Graphic Annotation on Multimedia English Learning)

  • 도경수;차유영
    • 인지과학
    • /
    • 제19권4호
    • /
    • pp.369-396
    • /
    • 2008
  • 멀티미디어 학습에서 시공간 작업기억의 용량에 따라 그림 유형과 그림 내용의 효과가 달라지는지를 두 개의 실험에서 알아보았다. 실험 1에서는 영어단어 설명을 단어로만 주는 조건, 동영상과 같이 주는 조건, 그림과 같이 주는 조건을 비교하였다. 단어검사에서 동영상은 시공간 작업기억 용량이 큰 집단에서만 효과가 있었고, 이해검사에서는 단어설명만 주어진 조건에서 가장 수행이 좋았다. 실험 2에서는 동영상의 내용이 미치는 영향을 알아보았다. 단어검사에서는 전형적인 내용을 담은 동영상이 효과가 있었으나, 이해검사에서는 맥락에 적합한 내용의 동영상으로 학습한 집단이 이해 점수가 높았고, 이 경향은 시공간 작업기억 용량이 작은 집단에서 더 두드러지게 나타났다. 두 실험의 결과는 멀티미디어 학습의 효과는 인지부하의 정도에 달려 있다는 인지부하이론과 멀티미디어 학습을 여러 수준의 활동의 결과로 설명한 생성이론의 의해 잘 설명되는 것으로 해석되었다.

  • PDF

GAN 신경망을 통한 자각적 사진 향상 (Perceptual Photo Enhancement with Generative Adversarial Networks)

  • 궐월;이효종
    • 한국정보처리학회:학술대회논문집
    • /
    • 한국정보처리학회 2019년도 춘계학술발표대회
    • /
    • pp.522-524
    • /
    • 2019
  • In spite of a rapid development in the quality of built-in mobile cameras, their some physical restrictions hinder them to achieve the satisfactory results of digital single lens reflex (DSLR) cameras. In this work we propose an end-to-end deep learning method to translate ordinary images by mobile cameras into DSLR-quality photos. The method is based on the framework of generative adversarial networks (GANs) with several improvements. First, we combined the U-Net with DenseNet and connected dense block (DB) in terms of U-Net. The Dense U-Net acts as the generator in our GAN model. Then, we improved the perceptual loss by using the VGG features and pixel-wise content, which could provide stronger supervision for contrast enhancement and texture recovery.

Face Recognition Research Based on Multi-Layers Residual Unit CNN Model

  • Zhang, Ruyang;Lee, Eung-Joo
    • 한국멀티미디어학회논문지
    • /
    • 제25권11호
    • /
    • pp.1582-1590
    • /
    • 2022
  • Due to the situation of the widespread of the coronavirus, which causes the problem of lack of face image data occluded by masks at recent time, in order to solve the related problems, this paper proposes a method to generate face images with masks using a combination of generative adversarial networks and spatial transformation networks based on CNN model. The system we proposed in this paper is based on the GAN, combined with multi-scale convolution kernels to extract features at different details of the human face images, and used Wasserstein divergence as the measure of the distance between real samples and synthetic samples in order to optimize Generator performance. Experiments show that the proposed method can effectively put masks on face images with high efficiency and fast reaction time and the synthesized human face images are pretty natural and real.

Object Tracking Based on Weighted Local Sub-space Reconstruction Error

  • Zeng, Xianyou;Xu, Long;Hu, Shaohai;Zhao, Ruizhen;Feng, Wanli
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제13권2호
    • /
    • pp.871-891
    • /
    • 2019
  • Visual tracking is a challenging task that needs learning an effective model to handle the changes of target appearance caused by factors such as pose variation, illumination change, occlusion and motion blur. In this paper, a novel tracking algorithm based on weighted local sub-space reconstruction error is presented. First, accounting for the appearance changes in the tracking process, a generative weight calculation method based on structural reconstruction error is proposed. Furthermore, a template update scheme of occlusion-aware is introduced, in which we reconstruct a new template instead of simply exploiting the best observation for template update. The effectiveness and feasibility of the proposed algorithm are verified by comparing it with some state-of-the-art algorithms quantitatively and qualitatively.

Frontal Face Generation Algorithm from Multi-view Images Based on Generative Adversarial Network

  • Heo, Young- Jin;Kim, Byung-Gyu;Roy, Partha Pratim
    • Journal of Multimedia Information System
    • /
    • 제8권2호
    • /
    • pp.85-92
    • /
    • 2021
  • In a face, there is much information of person's identity. Because of this property, various tasks such as expression recognition, identity recognition and deepfake have been actively conducted. Most of them use the exact frontal view of the given face. However, various directions of the face can be observed rather than the exact frontal image in real situation. The profile (side view) lacks information when comparing with the frontal view image. Therefore, if we can generate the frontal face from other directions, we can obtain more information on the given face. In this paper, we propose a combined style model based the conditional generative adversarial network (cGAN) for generating the frontal face from multi-view images that consist of characteristics that not only includes the style around the face (hair and beard) but also detailed areas (eye, nose, and mouth).

SkelGAN: A Font Image Skeletonization Method

  • Ko, Debbie Honghee;Hassan, Ammar Ul;Majeed, Saima;Choi, Jaeyoung
    • Journal of Information Processing Systems
    • /
    • 제17권1호
    • /
    • pp.1-13
    • /
    • 2021
  • In this research, we study the problem of font image skeletonization using an end-to-end deep adversarial network, in contrast with the state-of-the-art methods that use mathematical algorithms. Several studies have been concerned with skeletonization, but a few have utilized deep learning. Further, no study has considered generative models based on deep neural networks for font character skeletonization, which are more delicate than natural objects. In this work, we take a step closer to producing realistic synthesized skeletons of font characters. We consider using an end-to-end deep adversarial network, SkelGAN, for font-image skeletonization, in contrast with the state-of-the-art methods that use mathematical algorithms. The proposed skeleton generator is proved superior to all well-known mathematical skeletonization methods in terms of character structure, including delicate strokes, serifs, and even special styles. Experimental results also demonstrate the dominance of our method against the state-of-the-art supervised image-to-image translation method in font character skeletonization task.

Research on AI Painting Generation Technology Based on the [Stable Diffusion]

  • Chenghao Wang;Jeanhun Chung
    • International journal of advanced smart convergence
    • /
    • 제12권2호
    • /
    • pp.90-95
    • /
    • 2023
  • With the rapid development of deep learning and artificial intelligence, generative models have achieved remarkable success in the field of image generation. By combining the stable diffusion method with Web UI technology, a novel solution is provided for the application of AI painting generation. The application prospects of this technology are very broad and can be applied to multiple fields, such as digital art, concept design, game development, and more. Furthermore, the platform based on Web UI facilitates user operations, making the technology more easily applicable to practical scenarios. This paper introduces the basic principles of Stable Diffusion Web UI technology. This technique utilizes the stability of diffusion processes to improve the output quality of generative models. By gradually introducing noise during the generation process, the model can generate smoother and more coherent images. Additionally, the analysis of different model types and applications within Stable Diffusion Web UI provides creators with a more comprehensive understanding, offering valuable insights for fields such as artistic creation and design.

생성적 적대 신경망을 활용한 다른 그림 찾기 생성 시스템 (Spot The Difference Generation System Using Generative Adversarial Networks)

  • 송성헌;문미경;최봉준
    • 한국컴퓨터정보학회:학술대회논문집
    • /
    • 한국컴퓨터정보학회 2021년도 제64차 하계학술대회논문집 29권2호
    • /
    • pp.673-674
    • /
    • 2021
  • 본 논문은 집중력 향상 놀이인 다른 그림 찾기를 자신이 좋아하는 주제를 배경으로 쉽게 생성할 수 있는 시스템을 제안한다. 아동기에 주로 진단이 되고 성인기까지 이어질 수 있는 주의력 결핍 과다활동 증후군(ADHD)을 조기에 예방하기 위해 본 논문에서는 선택한 그림의 일부분을 가지고 생성적 적대 신경망을 활용하여 새로운 물체를 생성해 낸 뒤 자연스럽게 원본 그림에 융화될 수 있도록 하는 것이 목표이다. 하나의 다른 그림 찾기 콘텐츠를 만드는 것은 포토샵과 같이 전문성을 가진 툴을 전문가가 오랜 시간 작업해야 하는 내용이다. 전문적인 기술이 필요한 작업 과정을 본 연구를 통해 일반인도 쉽게 작업할 수 있도록 하는 것을 최종 목표로 한다.

  • PDF

GAN을 이용한 게임 캐릭터 이미지 생성 (Game Character Image Generation Using GAN)

  • 김정기;정명준;차경애
    • 대한임베디드공학회논문지
    • /
    • 제18권5호
    • /
    • pp.241-248
    • /
    • 2023
  • GAN (Generative Adversarial Networks) creates highly sophisticated counterfeit products by learning real images or text and inferring commonalities. Therefore, it can be useful in fields that require the creation of large-scale images or graphics. In this paper, we implement GAN-based game character creation AI that can dramatically reduce illustration design work costs by providing expansion and automation of game character image creation. This is very efficient in game development as it allows mass production of various character images at low cost.

REVIEW OF DIFFUSION MODELS: THEORY AND APPLICATIONS

  • HYUNGJIN CHUNG;HYELIN NAM;JONG CHUL YE
    • Journal of the Korean Society for Industrial and Applied Mathematics
    • /
    • 제28권1호
    • /
    • pp.1-21
    • /
    • 2024
  • This review comprehensively explores the evolution, theoretical underpinnings, variations, and applications of diffusion models. Originating as a generative framework, diffusion models have rapidly ascended to the forefront of machine learning research, owing to their exceptional capability, stability, and versatility. We dissect the core principles driving diffusion processes, elucidating their mathematical foundations and the mechanisms by which they iteratively refine noise into structured data. We highlight pivotal advancements and the integration of auxiliary techniques that have significantly enhanced their efficiency and stability. Variants such as bridges that broaden the applicability of diffusion models to wider domains are introduced. We put special emphasis on the ability of diffusion models as a crucial foundation model, with modalities ranging from image, 3D assets, and video. The role of diffusion models as a general foundation model leads to its versatility in many of the downstream tasks such as solving inverse problems and image editing. Through this review, we aim to provide a thorough and accessible compendium for both newcomers and seasoned researchers in the field.