• Title/Summary/Keyword: video generation

Search Result 580, Processing Time 0.021 seconds

Transmission of Multimedia Traffic over Mobile Ad-hoc Networks (모바일 ad-hoc 네트워크에서 멀티미디어 트래픽 전송)

  • Kim, Young-Dong
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.9 no.1
    • /
    • pp.95-101
    • /
    • 2005
  • In this paper, some performance characteristics of multimedia traffic for mobile ad-hoc networks is studied with simulations. Synthetic streaming video is considered as the multimedia traffic for MPEG-coded video in the simulation. The synthetic video stream is generated with a video stream generation algorithm. The algorithm generate VBR traffics for MPEG video streams with special predefined GOP(group of pictures) patterns that is consisted of a sequence of I(intra-coded), P(predicted-coded) and B(bidirectional-coded) frames. The synthetic VBR streams is transmitted through UDP protocol with on-demand mobile ad-hoc network routing protocols like as AODV and DSR. And performances for video streams through mobile ad-hoc networks is evaluated, the throughputs is compared between data and video traffics.

Analysis and Implementation of Distance Video Lecturing System (초고속통신망환경에서의 원격화상강의시스템 구현 및 분석)

  • 안종민;장종욱
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2002.05a
    • /
    • pp.90-95
    • /
    • 2002
  • This thesis is aiming at analysis and implementing Distance Video Lecturing System to guarantee the high quality in service through using less Network resource. Most of the usual Distance Video Lecturing System has been designed not in the system of Multiuser Conference but Point-to-Point. But from now on Distance Video Lecturing System is to be demanded on not only the expansion to Multiuser Distance Video Lecturing System, but high quality in service for users. To solve these problems, the Multicast Transmission Technology as the dimension of Network as well as highly efficient, advanced Compression Technology should be applied to Video Conference System in company with the expansion of Network bandwidth. The system suggested in this thesis is to help to offer an application service as Distance Video Lecturing System that will make popular on the basis of the character that the next generation Internet will be high speed Network.

  • PDF

A Comparative Study on the Features and Applications of AI Tools -Focus on PIKA Labs and RUNWAY

  • Biying Guo;Xinyi Shan;Jeanhun Chung
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.16 no.1
    • /
    • pp.86-91
    • /
    • 2024
  • In the field of artistic creation, the iterative development of AI-generated video software has pushed the boundaries of multimedia content creation and provided powerful creative tools for non-professionals. This paper extensively examines two leading AI-generated video software, PIKA Labs and RUNWAY, discussing their functions, performance differences, and application scopes in the video generation domain. Through detailed operational examples, a comparative analysis of their functionalities, as well as the advantages and limitations of each in generating video content, is presented. By comparison, it can be found that PIKA Labs and RUNWAY have excellent performance in stability and creativity. Therefore, the purpose of this study is to comprehensively elucidate the operating mechanisms of these two AI software, in order to intuitively demonstrate the advantages of each software. Simultaneously, this study provides valuable references for professionals and creators in the video production field, assisting them in selecting the most suitable tools for different scenarios, thereby advancing the application and development of AI-generated video software in multimedia content creation.

An Atlas Generation Method with Tiny Blocks Removal for Efficient 3DoF+ Video Coding (효율적인 3DoF+ 비디오 부호화를 위한 작은 블록 제거를 통한 아틀라스 생성 기법)

  • Lim, Sung-Gyun;Kim, Hyun-Ho;Kim, Jae-Gon
    • Journal of Broadcast Engineering
    • /
    • v.25 no.5
    • /
    • pp.665-671
    • /
    • 2020
  • MPEG-I is actively working on standardization on the coding of immersive video which provides up to 6 degree of freedom (6DoF) in terms of viewpoint. 3DoF+ video, which provides motion parallax to omnidirectional view of 360 video, renders a view at any desired viewpoint using multiple view videos acquisitioned in a limited 3D space covered with upper body motion at a fixed position. The MPEG-I visual group is developing a test model called TMIV (Test Model for Immersive Video) in the process of development of the standard for 3DoF+ video coding. In the TMIV, the redundancy between a set of input view videos is removed, and several atlases are generated by packing patches including the remaining texture and depth regions into frames as compact as possible, and coded. This paper presents an atlas generation method that removes small-sized blocks in the atlas for more efficient 3DoF+ video coding. The proposed method shows a performance improvement of BD-rate bit savings of 0.7% and 1.4%, respectively, in natural and graphic sequences compared to TMIV.

Abnormal Behavior Detection Based on Adaptive Background Generation for Intelligent Video Analysis (지능형 비디오 분석을 위한 적응적 배경 생성 기반의 이상행위 검출)

  • Lee, Seoung-Won;Kim, Tae-Kyung;Yoo, Jang-Hee;Paik, Joon-Ki
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.48 no.1
    • /
    • pp.111-121
    • /
    • 2011
  • Intelligent video analysis systems require techniques which can predict accidents and provide alarms to the monitoring personnel. In this paper, we present an abnormal behavior analysis technique based on adaptive background generation. More specifically, abnormal behaviors include fence climbing, abandoned objects, fainting persons, and loitering persons. The proposed video analysis system consists of (i) background generation and (ii) abnormal behavior analysis modules. For robust background generation, the proposed system updates static regions by detecting motion changes at each frame. In addition, noise and shadow removal steps are also were added to improve the accuracy of the object detection. The abnormal behavior analysis module extracts object information, such as centroid, silhouette, size, and trajectory. As the result of the behavior analysis function objects' behavior is configured and analyzed based on the a priori specified scenarios, such as fence climbing, abandoning objects, fainting, and loitering. In the experimental results, the proposed system was able to detect the moving object and analyze the abnormal behavior in complex environments.

Video-to-Video Generated by Collage Technique (콜라주 기법으로 해석한 비디오 생성)

  • Cho, Hyeongrae;Park, Gooman
    • Journal of Broadcast Engineering
    • /
    • v.26 no.1
    • /
    • pp.39-60
    • /
    • 2021
  • In the field of deep learning, there are many algorithms mainly after GAN in research related to generation, but in terms of generation, there are similarities and differences with art. If the generation in the engineering aspect is mainly to judge the presence or absence of a quantitative indicator or the correct answer and the incorrect answer, the creation in the artistic aspect creates a creation that interprets the world and human life by cross-validating and doubting the correct answer and incorrect answer from various perspectives. In this paper, the video generation ability of deep learning was interpreted from the perspective of collage and compared with the results made by the artist. The characteristic of the experiment is to compare and analyze how much GAN reproduces the result of the creator made with the collage technique and the difference between the creative part, and investigate the satisfaction level by making performance evaluation items for the reproducibility of GAN. In order to experiment on how much the creator's statement and purpose of expression were reproduced, a deep learning algorithm corresponding to the statement keyword was found and its similarity was compared. As a result of the experiment, GAN did not meet much expectations to express the collage technique. Nevertheless, the image association showed higher satisfaction than human ability, which is a positive discovery that GAN can show comparable ability to humans in terms of abstract creation.

Real-time Identification of Traffic Light and Road Sign for the Next Generation Video-Based Navigation System (차세대 실감 내비게이션을 위한 실시간 신호등 및 표지판 객체 인식)

  • Kim, Yong-Kwon;Lee, Ki-Sung;Cho, Seong-Ik;Park, Jeong-Ho;Choi, Kyoung-Ho
    • Journal of Korea Spatial Information System Society
    • /
    • v.10 no.2
    • /
    • pp.13-24
    • /
    • 2008
  • A next generation video based car navigation is researched to supplement the drawbacks of existed 2D based navigation and to provide the various services for safety driving. The components of this navigation system could be a load object database, identification module for load lines, and crossroad identification module, etc. In this paper, we proposed the traffic lights and road sign recognition method which can be effectively exploited for crossroad recognition in video-based car navigation systems. The method uses object color information and other spatial features in the video image. The results show average 90% recognition rate from 30m to 60m distance for traffic lights and 97% at 40-90m distance for load sign. The algorithm also achieves 46msec/frame processing time which also indicates the appropriateness of the algorithm in real-time processing.

  • PDF

Virtual View Rendering for 2D/3D Freeview Video Generation (2차원/3차원 자유시점 비디오 재생을 위한 가상시점 합성시스템)

  • Min, Dong-Bo;Sohn, Kwang-Hoon
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.45 no.4
    • /
    • pp.22-31
    • /
    • 2008
  • In this paper, we propose a new approach for efficient multiview stereo matching and virtual view generation, which are key technologies for 3DTV. We propose semi N-view & N-depth framework to estimate disparity maps efficiently and correctly. This framework reduces the redundancy on disparity estimation by using the information of neighboring views. The proposed method provides a user 2D/3D freeview video, and the user can select 2D/3D modes of freeview video. Experimental results show that the proposed method yields the accurate disparity maps and the synthesized novel view is satisfactory enough to provide user seamless freeview videos.

Automatic Generation of Video Metadata for the Super-personalized Recommendation of Media

  • Yong, Sung Jung;Park, Hyo Gyeong;You, Yeon Hwi;Moon, Il-Young
    • Journal of information and communication convergence engineering
    • /
    • v.20 no.4
    • /
    • pp.288-294
    • /
    • 2022
  • The media content market has been growing, as various types of content are being mass-produced owing to the recent proliferation of the Internet and digital media. In addition, platforms that provide personalized services for content consumption are emerging and competing with each other to recommend personalized content. Existing platforms use a method in which a user directly inputs video metadata. Consequently, significant amounts of time and cost are consumed in processing large amounts of data. In this study, keyframes and audio spectra based on the YCbCr color model of a movie trailer were extracted for the automatic generation of metadata. The extracted audio spectra and image keyframes were used as learning data for genre recognition in deep learning. Deep learning was implemented to determine genres among the video metadata, and suggestions for utilization were proposed. A system that can automatically generate metadata established through the results of this study will be helpful for studying recommendation systems for media super-personalization.

360 VR-based Sokcho Introduction Video Production (360 VR기반 속초 소개 영상 제작)

  • Lee, Jun-yeong;Im, So-Yeon;Park, Cheol-woo;Lee, Young-woo
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.493-495
    • /
    • 2022
  • This video is based on the newly emerged next-generation Media 360 VR. With the development of technology, digital content has developed and the opening of the COVID-19 pandemic untact era, people have found content that they can enjoy without going directly. 360 VR is a next-generation media that allows users to enjoy content in a three-dimensional manner as if they went to the site without having to go to the site. Using this, we would like to study the effective creation of local promotional videos.

  • PDF