• 제목/요약/키워드: AI-based image generation

검색결과 37건 처리시간 0.024초

AI 기반 이미지 생성 기술의 농업 적용 가능성 (Agricultural Applicability of AI based Image Generation)

  • 윤승리;이예영;정은규;안태인
    • 생물환경조절학회지
    • /
    • 제33권2호
    • /
    • pp.120-128
    • /
    • 2024
  • 2022년 ChatGPT 출시 이후, 생성형 AI 산업은 엄청난 규모로 성장하였으며, 인지 작업에 혁신을 가져올 것으로 기대되고 있다. 특히 AI 기반 이미지 생성 기술은 현재 디지털 세계의 핵심적인 변화를 주도하고 있다. 본 연구는 대표적인 AI 이미지 생성 도구인 미드저니, 스테이블 디퓨전, 그리고 파이어플라이의 기술적 원리를 분석하고, 이미지 생성 결과를 비교함으로써 그 유용성을 평가하였다. 실험 결과, 이 AI 도구들은 대표 시설원예 작물인 토마토, 딸기, 파프리카, 오이의 과실 이미지를 실제와 유사하게 재현하였다. 특히 파이어플라이는 실제 온실 재배 작물 이미지를 매우 사실적으로 묘사하는 능력을 보여주었다. 그러나 모든 도구들은 작물이 자라는 온실의 환경적 맥락을 완전히 반영하는 데에 있어서 다소 한계를 보였다. 프롬프트 개선 및 레퍼런스 이미지를 활용하여 딸기과실 이미지와 시설 딸기재배 시스템을 보다 정교하게 생성하는 과정도 포함되었으며, 이러한 접근은 AI 이미지 생성 기술의 세밀한 조정이 가능함을 보여준다. 오이 과실 이미지 생성능력을 비교한 결과, AI 생성 도구들은 실제 이미지와 매우 유사한 이미지를 생성해 냄으로써 이미지 생성 점수(CLIP score)에 있어서 통계적 차이를 보이지 않았다. 본 연구는 AI 기반 이미지 생성 이미지 기술이 농업 분야에 활용될 수 있는 방안을 모색하며, 생성형 AI의 농업에 대한 적용을 긍정적으로 전망한다.

민화와 풍속화를 이용한 AI 기반의 콘텐츠 원천 데이터 생성 모델의 연구 (A Study of an AI-Based Content Source Data Generation Model using Folk Paintings and Genre Paintings)

  • 양석환;이영숙
    • 한국멀티미디어학회논문지
    • /
    • 제24권5호
    • /
    • pp.736-743
    • /
    • 2021
  • Due to COVID-19, the non-face-to-face content market is growing rapidly. However, most of the non-face-to-face content such as webtoons and web novels are produced based on the traditional culture of other countries, not Korean traditional culture. The biggest cause of this situation is the lack of reference materials for creating based on Korean traditional culture. Therefore, the need for materials on traditional Korean culture that can be used for content creation is emerging. In this paper, we propose a generation model of source data based on traditional folk paintings through the fusion of traditional Korean folk paintings and AI technology. The proposed model secures basic data based on folk tales, analyzes the style and characteristics of folk tales, and converts historical backgrounds and various stories related to folk tales into data. In addition, using the built data, various new stories are created based on AI technology. The proposed model is highly utilized in that it provides a foundation for new creation based on Korean traditional folk painting and AI technology.

Best Practice on Automatic Toon Image Creation from JSON File of Message Sequence Diagram via Natural Language based Requirement Specifications

  • Hyuntae Kim;Ji Hoon Kong;Hyun Seung Son;R. Young Chul Kim
    • International journal of advanced smart convergence
    • /
    • 제13권1호
    • /
    • pp.99-107
    • /
    • 2024
  • In AI image generation tools, most general users must use an effective prompt to craft queries or statements to elicit the desired response (image, result) from the AI model. But we are software engineers who focus on software processes. At the process's early stage, we use informal and formal requirement specifications. At this time, we adapt the natural language approach into requirement engineering and toon engineering. Most Generative AI tools do not produce the same image in the same query. The reason is that the same data asset is not used for the same query. To solve this problem, we intend to use informal requirement engineering and linguistics to create a toon. Therefore, we propose a sequence diagram and image generation mechanism by analyzing and applying key objects and attributes as an informal natural language requirement analysis. Identify morpheme and semantic roles by analyzing natural language through linguistic methods. Based on the analysis results, a sequence diagram and an image are generated through the diagram. We expect consistent image generation using the same image element asset through the proposed mechanism.

Research on AI Painting Generation Technology Based on the [Stable Diffusion]

  • Chenghao Wang;Jeanhun Chung
    • International journal of advanced smart convergence
    • /
    • 제12권2호
    • /
    • pp.90-95
    • /
    • 2023
  • With the rapid development of deep learning and artificial intelligence, generative models have achieved remarkable success in the field of image generation. By combining the stable diffusion method with Web UI technology, a novel solution is provided for the application of AI painting generation. The application prospects of this technology are very broad and can be applied to multiple fields, such as digital art, concept design, game development, and more. Furthermore, the platform based on Web UI facilitates user operations, making the technology more easily applicable to practical scenarios. This paper introduces the basic principles of Stable Diffusion Web UI technology. This technique utilizes the stability of diffusion processes to improve the output quality of generative models. By gradually introducing noise during the generation process, the model can generate smoother and more coherent images. Additionally, the analysis of different model types and applications within Stable Diffusion Web UI provides creators with a more comprehensive understanding, offering valuable insights for fields such as artistic creation and design.

A Case Study of Creative Art Based on AI Generation Technology

  • Qianqian Jiang;Jeanhun Chung
    • International journal of advanced smart convergence
    • /
    • 제12권2호
    • /
    • pp.84-89
    • /
    • 2023
  • In recent years, with the breakthrough of Artificial Intelligence (AI) technology in deep learning algorithms such as Generative Adversarial Networks (GANs) and Variational Autoencoders (VAE), AI generation technology has rapidly expanded in various sub-sectors in the art field. 2022 as the explosive year of AI-generated art, especially in the creation of AI-generated art creative design, many excellent works have been born, which has improved the work efficiency of art design. This study analyzed the application design characteristics of AI generation technology in two sub fields of artistic creative design of AI painting and AI animation production , and compares the differences between traditional painting and AI painting in the field of painting. Through the research of this paper, the advantages and problems in the process of AI creative design are summarized. Although AI art designs are affected by technical limitations, there are still flaws in artworks and practical problems such as copyright and income, but it provides a strong technical guarantee in the expansion of subdivisions of artistic innovation and technology integration, and has extremely high research value.

한국형 멀티모달 몽타주 앱을 위한 생성형 AI 연구 (Research on Generative AI for Korean Multi-Modal Montage App)

  • 임정현;차경애;고재필;홍원기
    • 서비스연구
    • /
    • 제14권1호
    • /
    • pp.13-26
    • /
    • 2024
  • 멀티모달 (multi-modal) 생성이란 텍스트, 이미지, 오디오 등 다양한 정보를 기반으로 결과를 도출하는 작업을 말한다. AI 기술의 비약적인 발전으로 인해 여러 가지 유형의 데이터를 종합적으로 처리해 결과를 도출하는 멀티모달 기반 시스템 또한 다양해지는 추세이다. 본 논문은 음성과 텍스트 인식을 활용하여 인물을 묘사하면, 몽타주 이미지를 생성하는 AI 시스템의 개발 내용을 소개한다. 기존의 몽타주 생성 기술은 서양인들의 외형을 기준으로 이루어진 반면, 본 논문에서 개발한 몽타주 생성 시스템은 한국인의 안면 특징을 바탕으로 모델을 학습한다. 따라서, 한국어에 특화된 음성과 텍스트의 멀티모달을 기반으로 보다 정확하고 효과적인 한국형 몽타주 이미지를 만들어낼 수 있다. 개발된 몽타주 생성 앱은 몽타주 초안으로 충분히 활용 가능하기 때문에 기존의 몽타주 제작 인력의 수작업을 획기적으로 줄여줄 수 있다. 이를 위해 한국지능정보사회진흥원의 AI-Hub에서 제공하는 페르소나 기반 가상 인물 몽타주 데이터를 활용하였다. AI-Hub는 AI 기술 및 서비스 개발에 필요한 인공지능 학습용 데이터를 구축하여 원스톱 제공을 목적으로 한 AI 통합 플랫폼이다. 이미지 생성 시스템은 고해상도 이미지를 생성하는데 사용하는 딥러닝 모델인 VQGAN과 한국어 기반 영상생성 모델인 KoDALLE 모델을 사용하여 구현하였다. 학습된 AI 모델은 음성과 텍스트를 이용해 묘사한 내용과 매우 유사한 얼굴의 몽타주 이미지가 생성됨을 확인할 수 있다. 개발된 몽타주 생성 앱의 실용성 검증을 위해 10명의 테스터가 사용한 결과 70% 이상이 만족한다는 응답을 보였다. 몽타주 생성 앱은 범죄자 검거 등 얼굴의 특징을 묘사하여 이미지화하는 여러 분야에서 다양하게 사용될 수 있을 것이다.

Performance Comparison According to Image Generation Method in NIDS (Network Intrusion Detection System) using CNN

  • Sang Hyun, Kim
    • International journal of advanced smart convergence
    • /
    • 제12권2호
    • /
    • pp.67-75
    • /
    • 2023
  • Recently, many studies have been conducted on ways to utilize AI technology in NIDS (Network Intrusion Detection System). In particular, CNN-based NIDS generally shows excellent performance. CNN is basically a method of using correlation between pixels existing in an image. Therefore, the method of generating an image is very important in CNN. In this paper, the performance comparison of CNN-based NIDS according to the image generation method was performed. The image generation methods used in the experiment are a direct conversion method and a one-hot encoding based method. As a result of the experiment, the performance of NIDS was different depending on the image generation method. In particular, it was confirmed that the method combining the direct conversion method and the one-hot encoding based method proposed in this paper showed the best performance.

A Comparative Analysis Between <Leonardo.Ai> and <Meshy> as AI Texture Generation Tools

  • Pingjian Jie;Xinyi Shan;Jeanhun Chung
    • International Journal of Advanced Culture Technology
    • /
    • 제11권4호
    • /
    • pp.333-339
    • /
    • 2023
  • In three-dimensional(3D) modeling, texturing plays a crucial role as a visual element, imparting detail and realism to models. In contrast to traditional texturing methods, the current trend involves utilizing AI tools such as Leonardo.Ai and Meshy to create textures for 3D models in a more efficient and precise manner. This paper focuses on 3D texturing, conducting a comprehensive comparative study of AI tools, specifically Leonardo.Ai and Meshy. By delving into the performance, functional differences, and respective application scopes of these two tools in the generation of 3D textures, we highlight potential applications and development trends within the realm of 3D texturing. The efficient use of AI tools in texture creation also has the potential to drive innovation and enhancement in the field of 3D modeling. In conclusion, this research aims to provide a comprehensive perspective for researchers, practitioners, and enthusiasts in related fields, fostering further innovation and development in this domain.

A Feasibility Study on RUNWAY GEN-2 for Generating Realistic Style Images

  • Yifan Cui;Xinyi Shan;Jeanhun Chung
    • International Journal of Internet, Broadcasting and Communication
    • /
    • 제16권1호
    • /
    • pp.99-105
    • /
    • 2024
  • Runway released an updated version, Gen-2, in March 2023, which introduced new features that are different from Gen-1: it can convert text and images into videos, or convert text and images together into video images based on text instructions. This update will be officially open to the public in June 2023, so more people can enjoy and use their creativity. With this new feature, users can easily transform text and images into impressive video creations. However, as with all new technologies, comes the instability of AI, which also affects the results generated by Runway. This article verifies the feasibility of using Runway to generate the desired video from several aspects through personal practice. In practice, I discovered Runway generation problems and propose improvement methods to find ways to improve the accuracy of Runway generation. And found that although the instability of AI is a factor that needs attention, through careful adjustment and testing, users can still make full use of this feature and create stunning video works. This update marks the beginning of a more innovative and diverse future for the digital creative field.