• Title/Summary/Keyword: Generative Model

Search Result 351, Processing Time 0.027 seconds

3D Object Generation and Renderer System based on VAE ResNet-GAN

  • Min-Su Yu;Tae-Won Jung;GyoungHyun Kim;Soonchul Kwon;Kye-Dong Jung
    • International journal of advanced smart convergence
    • /
    • v.12 no.4
    • /
    • pp.142-146
    • /
    • 2023
  • We present a method for generating 3D structures and rendering objects by combining VAE (Variational Autoencoder) and GAN (Generative Adversarial Network). This approach focuses on generating and rendering 3D models with improved quality using residual learning as the learning method for the encoder. We deep stack the encoder layers to accurately reflect the features of the image and apply residual blocks to solve the problems of deep layers to improve the encoder performance. This solves the problems of gradient vanishing and exploding, which are problems when constructing a deep neural network, and creates a 3D model of improved quality. To accurately extract image features, we construct deep layers of the encoder model and apply the residual function to learning to model with more detailed information. The generated model has more detailed voxels for more accurate representation, is rendered by adding materials and lighting, and is finally converted into a mesh model. 3D models have excellent visual quality and accuracy, making them useful in various fields such as virtual reality, game development, and metaverse.

A Research on Aesthetic Aspects of Checkpoint Models in [Stable Diffusion]

  • Ke Ma;Jeanhun Chung
    • International journal of advanced smart convergence
    • /
    • v.13 no.2
    • /
    • pp.130-135
    • /
    • 2024
  • The Stable diffsuion AI tool is popular among designers because of its flexible and powerful image generation capabilities. However, due to the diversity of its AI models, it needs to spend a lot of time testing different AI models in the face of different design plans, so choosing a suitable general AI model has become a big problem at present. In this paper, by comparing the AI images generated by two different Stable diffsuion models, the advantages and disadvantages of each model are analyzed from the aspects of the matching degree of the AI image and the prompt, the color composition and light composition of the image, and the general AI model that the generated AI image has an aesthetic sense is analyzed, and the designer does not need to take cumbersome steps. A satisfactory AI image can be obtained. The results show that Playground V2.5 model can be used as a general AI model, which has both aesthetic and design sense in various style design requirements. As a result, content designers can focus more on creative content development, and expect more groundbreaking technologies to merge generative AI with content design.

Face Recognition on complex backgrounds using Neural Network (복잡한 배경에서 신경망을 이용한 얼굴인식)

  • Han, Jun-Hee;Nam, Kee-Hwan;Park, Ho-Sik;Lee, Young-Sik;Jung, Yeon-Gil;Ra, Sang-Dong;Bae, Cheol-Soo
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • v.9 no.1
    • /
    • pp.1149-1152
    • /
    • 2005
  • Detecting faces in images with complex backgrounds is a difficult task. Our approach, which obtains state of the art results, is based on a generative neural network model: the Constrained Generative Model (CGM). To detect side view faces and to decrease the number of false alarms, a conditional mixture of networks is used. To decrease the computational time cost, a fast search algorithm is proposed. The level of performance reached, in terms of detection accuracy and processing time, allows to apply this detector to a real word application: the indexation of face images on the Web.

  • PDF

Evaluation of Sentimental Texts Automatically Generated by a Generative Adversarial Network (생성적 적대 네트워크로 자동 생성한 감성 텍스트의 성능 평가)

  • Park, Cheon-Young;Choi, Yong-Seok;Lee, Kong Joo
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.8 no.6
    • /
    • pp.257-264
    • /
    • 2019
  • Recently, deep neural network based approaches have shown a good performance for various fields of natural language processing. A huge amount of training data is essential for building a deep neural network model. However, collecting a large size of training data is a costly and time-consuming job. A data augmentation is one of the solutions to this problem. The data augmentation of text data is more difficult than that of image data because texts consist of tokens with discrete values. Generative adversarial networks (GANs) are widely used for image generation. In this work, we generate sentimental texts by using one of the GANs, CS-GAN model that has a discriminator as well as a classifier. We evaluate the usefulness of generated sentimental texts according to various measurements. CS-GAN model not only can generate texts with more diversity but also can improve the performance of its classifier.

Entity Embeddings for Enhancing Feasible and Diverse Population Synthesis in a Deep Generative Models (심층 생성모델 기반 합성인구 생성 성능 향상을 위한 개체 임베딩 분석연구)

  • Donghyun Kwon;Taeho Oh;Seungmo Yoo;Heechan Kang
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.22 no.6
    • /
    • pp.17-31
    • /
    • 2023
  • An activity-based model requires detailed population information to model individual travel behavior in a disaggregated manner. The recent innovative approach developed deep generative models with novel regularization terms that improves fidelity and diversity for population synthesis. Since the method relies on measuring the distance between distribution boundaries of the sample data and the generated sample, it is crucial to obtain well-defined continuous representation from the discretized dataset. Therefore, we propose an improved entity embedding models to enhance the performance of the regularization terms, which indirectly supports the synthesis in terms of feasible and diverse populations. Our results show a 28.87% improvement in the F1 score compared to the baseline method.

Software Education Class Model using Generative AI - Focusing on ChatGPT (생성형 AI를 활용한 소프트웨어교육 수업모델 연구 - ChatGPT를 중심으로)

  • Myung-suk Lee
    • Journal of Practical Engineering Education
    • /
    • v.16 no.3_spc
    • /
    • pp.275-282
    • /
    • 2024
  • This study studied a teaching model for software education using generative AI. The purpose of the study is to use ChatGPT as an instructor's assistant in programming classes for non-major students by using ChatGPT in software education. In addition, we designed ChatGPT to enable individual learning for learners and provide immediate feedback when students need it. The research method was conducted using ChatGPT as an assistant for non-computer majors taking a liberal arts Python class. In addition, we confirmed whether ChatGPT has the potential as an assistant in programming education for non-major students. Students actively used ChatGPT for writing assignments, correcting errors, writing coding, and acquiring knowledge, and confirmed various advantages, such as being able to focus on understanding the program rather than spending a lot of time resolving errors. We were able to see the potential for ChatGPT to increase students' learning efficiency, and we were able to see that more research is needed on its use in education. In the future, research will be conducted on the development, supplementation, and evaluation methods of educational models using ChatGPT.

Non-pneumatic Tire Design System based on Generative Adversarial Networks (적대적 생성 신경망 기반 비공기압 타이어 디자인 시스템)

  • JuYong Seong;Hyunjun Lee;Sungchul Lee
    • Journal of Platform Technology
    • /
    • v.11 no.6
    • /
    • pp.34-46
    • /
    • 2023
  • The design of non-pneumatic tires, which are created by filling the space between the wheel and the tread with elastomeric compounds or polygonal spokes, has become an important research topic in the automotive and aerospace industries. In this study, a system was designed for the design of non-pneumatic tires through the implementation of a generative adversarial network. We specifically examined factors that could impact the design, including the type of non-pneumatic tire, its intended usage environment, manufacturing techniques, distinctions from pneumatic tires, and how spoke design affects load distribution. Using OpenCV, various shapes and spoke configurations were generated as images, and a GAN model was trained on the projected GANs to generate shapes and spokes for non-pneumatic tire designs. The designed non-pneumatic tires were labeled as available or not, and a Vision Transformer image classification AI model was trained on these labels for classification purposes. Evaluation of the classification model show convergence to a near-zero loss and a 99% accuracy rate confirming the generation of non-pneumatic tire designs.

  • PDF

A StyleGAN Image Detection Model Based on Convolutional Neural Network (합성곱신경망 기반의 StyleGAN 이미지 탐지모델)

  • Kim, Jiyeon;Hong, Seung-Ah;Kim, Hamin
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.12
    • /
    • pp.1447-1456
    • /
    • 2019
  • As artificial intelligence technology is actively used in image processing, it is possible to generate high-quality fake images based on deep learning. Fake images generated using GAN(Generative Adversarial Network), one of unsupervised learning algorithms, have reached levels that are hard to discriminate from the naked eye. Detecting these fake images is required as they can be abused for crimes such as illegal content production, identity fraud and defamation. In this paper, we develop a deep-learning model based on CNN(Convolutional Neural Network) for the detection of StyleGAN fake images. StyleGAN is one of GAN algorithms and has an excellent performance in generating face images. We experiment with 48 number of experimental scenarios developed by combining parameters of the proposed model. We train and test each scenario with 300,000 number of real and fake face images in order to present a model parameter that improves performance in the detection of fake faces.

An Image-to-Image Translation GAN Model for Dental Prothesis Design (치아 보철물 디자인을 위한 이미지 대 이미지 변환 GAN 모델)

  • Tae-Min Kim;Jae-Gon Kim
    • Journal of Information Technology Services
    • /
    • v.22 no.5
    • /
    • pp.87-98
    • /
    • 2023
  • Traditionally, tooth restoration has been carried out by replicating teeth using plaster-based materials. However, recent technological advances have simplified the production process through the introduction of computer-aided design(CAD) systems. Nevertheless, dental restoration varies among individuals, and the skill level of dental technicians significantly influences the accuracy of the manufacturing process. To address this challenge, this paper proposes an approach to designing personalized tooth restorations using Generative Adversarial Network(GAN), a widely adopted technique in computer vision. The primary objective of this model is to create customized dental prosthesis for each patient by utilizing 3D data of the specific teeth to be treated and their corresponding opposite tooth. To achieve this, the 3D dental data is converted into a depth map format and used as input data for the GAN model. The proposed model leverages the network architecture of Pixel2Style2Pixel, which has demonstrated superior performance compared to existing models for image conversion and dental prosthesis generation. Furthermore, this approach holds promising potential for future advancements in dental and implant production.

Analysis and Forecast of Venture Capital Investment on Generative AI Startups: Focusing on the U.S. and South Korea (생성 AI 스타트업에 대한 벤처투자 분석과 예측: 미국과 한국을 중심으로)

  • Lee, Seungah;Jung, Taehyun
    • Asia-Pacific Journal of Business Venturing and Entrepreneurship
    • /
    • v.18 no.4
    • /
    • pp.21-35
    • /
    • 2023
  • Expectations surrounding generative AI technology and its profound ramifications are sweeping across various industrial domains. Given the anticipated pivotal role of the startup ecosystem in the utilization and advancement of generative AI technology, it is imperative to cultivate a deeper comprehension of the present state and distinctive attributes characterizing venture capital (VC) investments within this domain. The current investigation delves into South Korea's landscape of VC investment deals and prognosticates the projected VC investments by juxtaposing these against the United States, the frontrunner in the generative AI industry and its associated ecosystem. For analytical purposes, a compilation of 286 investment deals originating from 117 U.S. generative AI startups spanning the period from 2008 to 2023, as well as 144 investment deals from 42 South Korean generative AI startups covering the years 2011 to 2023, was amassed to construct new datasets. The outcomes of this endeavor reveal an upward trajectory in the count of VC investment deals within both the U.S. and South Korea during recent years. Predominantly, these deals have been concentrated within the early-stage investment realm. Noteworthy disparities between the two nations have also come to light. Specifically, in the U.S., in contrast to South Korea, the quantum of recent VC deals has escalated, marking an augmentation ranging from 285% to 488% in the corresponding developmental stage. While the interval between disparate investment stages demonstrated a slight elongation in South Korea relative to the U.S., this discrepancy did not achieve statistical significance. Furthermore, the proportion of VC investments channeled into generative AI enterprises, relative to the aggregate number of deals, exhibited a higher quotient in South Korea compared to the U.S. Upon a comprehensive sectoral breakdown of generative AI, it was discerned that within the U.S., 59.2% of total deals were concentrated in the text and model sectors, whereas in South Korea, 61.9% of deals centered around the video, image, and chat sectors. Through forecasting, the anticipated VC investments in South Korea from 2023 to 2029 were derived via four distinct models, culminating in an estimated average requirement of 3.4 trillion Korean won (ranging from at least 2.408 trillion won to a maximum of 5.919 trillion won). This research bears pragmatic significance as it methodically dissects VC investments within the generative AI domain across both the U.S. and South Korea, culminating in the presentation of an estimated VC investment projection for the latter. Furthermore, its academic significance lies in laying the groundwork for prospective scholarly inquiries by dissecting the current landscape of generative AI VC investments, a sphere that has hitherto remained void of rigorous academic investigation supported by empirical data. Additionally, the study introduces two innovative methodologies for the prediction of VC investment sums. Upon broader integration, application, and refinement of these methodologies within diverse academic explorations, they stand poised to enhance the prognosticative capacity pertaining to VC investment costs.

  • PDF