• Title/Summary/Keyword: AI Video

Search Result 178, Processing Time 0.037 seconds

Transforming Text into Video: A Proposed Methodology for Video Production Using the VQGAN-CLIP Image Generative AI Model

  • SukChang Lee
    • International Journal of Advanced Culture Technology
    • /
    • v.11 no.3
    • /
    • pp.225-230
    • /
    • 2023
  • With the development of AI technology, there is a growing discussion about Text-to-Image Generative AI. We presented a Generative AI video production method and delineated a methodology for the production of personalized AI-generated videos with the objective of broadening the landscape of the video domain. And we meticulously examined the procedural steps involved in AI-driven video production and directly implemented a video creation approach utilizing the VQGAN-CLIP model. The outcomes produced by the VQGAN-CLIP model exhibited a relatively moderate resolution and frame rate, and predominantly manifested as abstract images. Such characteristics indicated potential applicability in OTT-based video content or the realm of visual arts. It is anticipated that AI-driven video production techniques will see heightened utilization in forthcoming endeavors.

A Study on the Work Process of Creating AI SORA Videos (AI SORA 동영상 생성 제작의 작업 과정에 관한 고찰)

  • Cho, Hyun Kyung
    • The Journal of the Convergence on Culture Technology
    • /
    • v.10 no.5
    • /
    • pp.827-832
    • /
    • 2024
  • The AI program Sora is a video production model that can be used innovatively and is the starting point of a major paradigm shift in video planning and production in the future. In this paper, through consideration of the characteristics, application, and process of the AI video production program, the characteristics of the AI design video production method were understood, and the production algorithm was considered. The detailed consideration and characteristics of the work creation process for the video graphic AI video generation program that will be intensified every year were examined. Next, the method of generating a customized video with a text prompt and the process of innovative production results different from the previous production method were considered. In addition, the design direction through the generation of AI images was studied through the review of the strengths and weaknesses of the image details of the recently announced AI music video results. By considering the security of the AI generation video Sora and looking at the internal process of the actual AI process, it will be possible to present indicators for the future direction of AI video model production and education along with the direction of the design designer and education system. In the text and conclusion, we analyzed the strengths and weaknesses and future status of OpenAI Sora image, concluded how to apply the Sora model's capabilities, limitations, quality, and human creativity, and presented problems and alternatives through examples of the Sora model's capabilities and limitations to increase human creativity.

Artificial Intelligence-Based Video Content Generation (인공지능 기반 영상 콘텐츠 생성 기술 동향)

  • Son, J.W.;Han, M.H.;Kim, S.J.
    • Electronics and Telecommunications Trends
    • /
    • v.34 no.3
    • /
    • pp.34-42
    • /
    • 2019
  • This study introduces artificial intelligence (AI) techniques for video generation. For an effective illustration, techniques for video generation are classified as either semi-automatic or automatic. First, we discuss some recent achievements in semi-automatic video generation, and explain which types of AI techniques can be applied to produce films and improve film quality. Additionally, we provide an example of video content that has been generated by using AI techniques. Then, two automatic video-generation techniques are introduced with technical details. As there is currently no feasible automatic video-generation technique that can generate commercial videos, in this study, we explain their technical details, and suggest the future direction for researchers. Finally, we discuss several considerations for more practical automatic video-generation techniques.

A Comparative Study on the Features and Applications of AI Tools -Focus on PIKA Labs and RUNWAY

  • Biying Guo;Xinyi Shan;Jeanhun Chung
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.16 no.1
    • /
    • pp.86-91
    • /
    • 2024
  • In the field of artistic creation, the iterative development of AI-generated video software has pushed the boundaries of multimedia content creation and provided powerful creative tools for non-professionals. This paper extensively examines two leading AI-generated video software, PIKA Labs and RUNWAY, discussing their functions, performance differences, and application scopes in the video generation domain. Through detailed operational examples, a comparative analysis of their functionalities, as well as the advantages and limitations of each in generating video content, is presented. By comparison, it can be found that PIKA Labs and RUNWAY have excellent performance in stability and creativity. Therefore, the purpose of this study is to comprehensively elucidate the operating mechanisms of these two AI software, in order to intuitively demonstrate the advantages of each software. Simultaneously, this study provides valuable references for professionals and creators in the video production field, assisting them in selecting the most suitable tools for different scenarios, thereby advancing the application and development of AI-generated video software in multimedia content creation.

Enhancing Video Storyboarding with Artificial Intelligence: An Integrated Approach Using ChatGPT and Midjourney within AiSAC

  • Sukchang Lee
    • International Journal of Advanced Culture Technology
    • /
    • v.11 no.3
    • /
    • pp.253-259
    • /
    • 2023
  • The increasing incorporation of AI in video storyboard creation has been observed recently. Traditionally, the production of storyboards requires significant time, cost, and specialized expertise. However, the integration of AI can amplify the efficiency of storyboard creation and enhance storytelling. In Korea, AiSAC stands at the forefront of AI-driven storyboard platforms, boasting the capability to generate realistic images built on open datasets foundations. Yet, a notable limitation is the difficulty in intricately conveying a director's vision within the storyboard. To address this challenge, we proposed the application of image generation features from ChatGPT and Midjourney to AiSAC. Through this research, we aimed to enhance the efficiency of storyboard production and refined the intricacy of expression, thereby facilitating advancements in the video production process.

Resource Efficient AI Service Framework Associated with a Real-Time Object Detector

  • Jun-Hyuk Choi;Jeonghun Lee;Kwang-il Hwang
    • Journal of Information Processing Systems
    • /
    • v.19 no.4
    • /
    • pp.439-449
    • /
    • 2023
  • This paper deals with a resource efficient artificial intelligence (AI) service architecture for multi-channel video streams. As an AI service, we consider the object detection model, which is the most representative for video applications. Since most object detection models are basically designed for a single channel video stream, the utilization of the additional resource for multi-channel video stream processing is inevitable. Therefore, we propose a resource efficient AI service framework, which can be associated with various AI service models. Our framework is designed based on the modular architecture, which consists of adaptive frame control (AFC) Manager, multiplexer (MUX), adaptive channel selector (ACS), and YOLO interface units. In order to run only a single YOLO process without regard to the number of channels, we propose a novel approach efficiently dealing with multi-channel input streams. Through the experiment, it is shown that the framework is capable of performing object detection service with minimum resource utilization even in the circumstance of multi-channel streams. In addition, each service can be guaranteed within a deadline.

A Systematic Mapping Study on Artificial Intelligence Tools Used in Video Editing

  • Bieda, Igor;Panchenko, Taras
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.3
    • /
    • pp.312-318
    • /
    • 2022
  • From the past two eras, artificial intelligence has gained the attention of researchers of all research areas. Video editing is a task in the list that starts leveraging the blessing of Artificial Intelligence (AI). Since AI promises to make technology better use of human life although video editing technology is not new yet it is adopting new technologies like AI to become more powerful and sophisticated for video editors as well as users. Like other technologies, video editing will also be facilitated by the majestic power of AI in near future. There has been a lot of research that uses AI in video editing, yet there is no comprehensive literature review that systematically finds all of this work on one page so that new researchers can find research gaps in that area. In this research we conducted a statically approach called, systematic mapping study, to find answers to pre-proposed research questions. The aim and objective of this research are to find research gaps in our topic under discussion.

Proposal for AI Video Interview Using Image Data Analysis

  • Park, Jong-Youel;Ko, Chang-Bae
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.14 no.2
    • /
    • pp.212-218
    • /
    • 2022
  • In this paper, the necessity of AI video interview arises when conducting an interview for acquisition of excellent talent in a non-face-to-face situation due to similar situations such as Covid-19. As a matter to be supplemented in general AI interviews, it is difficult to evaluate the reliability and qualitative factors. In addition, the AI interview is conducted not in a two-way Q&A, rather in a one-sided Q&A process. This paper intends to fuse the advantages of existing AI interviews and video interviews. When conducting an interview using AI image analysis technology, it supplements subjective information that evaluates interview management and provides quantitative analysis data and HR expert data. In this paper, image-based multi-modal AI image analysis technology, bioanalysis-based HR analysis technology, and web RTC-based P2P image communication technology are applied. The goal of applying this technology is to propose a method in which biological analysis results (gaze, posture, voice, gesture, landmark) and HR information (opinions or features based on user propensity) can be processed on a single screen to select the right person for the hire.

Detection of video editing points using facial keypoints (얼굴 특징점을 활용한 영상 편집점 탐지)

  • Joshep Na;Jinho Kim;Jonghyuk Park
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.4
    • /
    • pp.15-30
    • /
    • 2023
  • Recently, various services using artificial intelligence(AI) are emerging in the media field as well However, most of the video editing, which involves finding an editing point and attaching the video, is carried out in a passive manner, requiring a lot of time and human resources. Therefore, this study proposes a methodology that can detect the edit points of video according to whether person in video are spoken by using Video Swin Transformer. First, facial keypoints are detected through face alignment. To this end, the proposed structure first detects facial keypoints through face alignment. Through this process, the temporal and spatial changes of the face are reflected from the input video data. And, through the Video Swin Transformer-based model proposed in this study, the behavior of the person in the video is classified. Specifically, after combining the feature map generated through Video Swin Transformer from video data and the facial keypoints detected through Face Alignment, utterance is classified through convolution layers. In conclusion, the performance of the image editing point detection model using facial keypoints proposed in this paper improved from 87.46% to 89.17% compared to the model without facial keypoints.

GreedyUCB1 based Monte-Carlo Tree Search for General Video Game Playing Artificial Intelligence (일반 비디오 게임 플레이 인공지능을 위한 GreedyUCB1기반 몬테카를로 트리 탐색)

  • Park, Hyunsoo;Kim, HyunTae;Kim, KyungJoong
    • KIISE Transactions on Computing Practices
    • /
    • v.21 no.8
    • /
    • pp.572-577
    • /
    • 2015
  • Generally, the existing Artificial Intelligence (AI) systems were designed for specific purposes and their capabilities handle only specific problems. Alternatively, Artificial General Intelligence can solve new problems as well as those that are already known. Recently, General Video Game Playing the game AI version of General Artificial Intelligence, has garnered a large amount of interest among Game Artificial Intelligence communities. Although video games are the sole concern, the design of a single AI that is capable of playing various video games is not an easy process. In this paper, we propose a GreedyUCB1 algorithm and rollout method that were formulated using the knowledge from a game analysis for the Monte-Carlo Tree Search game AI. An AI that used our method was ranked fourth at the GVG-AI (General Video Game-Artificial Intelligence) competition of the IEEE international conference of CIG (Computational Intelligence in Games) 2014.