• 제목/요약/키워드: Video Generation

검색결과 580건 처리시간 0.025초

인공지능 기반 영상 콘텐츠 생성 기술 동향 (Artificial Intelligence-Based Video Content Generation)

  • 손정우;한민호;김선중
    • 전자통신동향분석
    • /
    • 제34권3호
    • /
    • pp.34-42
    • /
    • 2019
  • This study introduces artificial intelligence (AI) techniques for video generation. For an effective illustration, techniques for video generation are classified as either semi-automatic or automatic. First, we discuss some recent achievements in semi-automatic video generation, and explain which types of AI techniques can be applied to produce films and improve film quality. Additionally, we provide an example of video content that has been generated by using AI techniques. Then, two automatic video-generation techniques are introduced with technical details. As there is currently no feasible automatic video-generation technique that can generate commercial videos, in this study, we explain their technical details, and suggest the future direction for researchers. Finally, we discuss several considerations for more practical automatic video-generation techniques.

3D Video Processing for 3DTV

  • Sohn, Kwang-Hoon
    • 한국정보디스플레이학회:학술대회논문집
    • /
    • 한국정보디스플레이학회 2007년도 7th International Meeting on Information Display 제7권2호
    • /
    • pp.1231-1234
    • /
    • 2007
  • This paper presents the overview of 3D video processing technologies for 3DTV such as 3D content generation, 3D video codec and video processing techniques for 3D displays. Some experimental results for 3D contents generation are shown in 3D mixed reality and 2D/3D conversion.

  • PDF

Video Captioning with Visual and Semantic Features

  • Lee, Sujin;Kim, Incheol
    • Journal of Information Processing Systems
    • /
    • 제14권6호
    • /
    • pp.1318-1330
    • /
    • 2018
  • Video captioning refers to the process of extracting features from a video and generating video captions using the extracted features. This paper introduces a deep neural network model and its learning method for effective video captioning. In this study, visual features as well as semantic features, which effectively express the video, are also used. The visual features of the video are extracted using convolutional neural networks, such as C3D and ResNet, while the semantic features are extracted using a semantic feature extraction network proposed in this paper. Further, an attention-based caption generation network is proposed for effective generation of video captions using the extracted features. The performance and effectiveness of the proposed model is verified through various experiments using two large-scale video benchmarks such as the Microsoft Video Description (MSVD) and the Microsoft Research Video-To-Text (MSR-VTT).

객체 기반 MPEG-4 동영상의 입체 변환 (Stereoscopic Conversion of Object-based MPEG-4 Video)

  • 박상훈;김만배;손현식
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2003년도 하계종합학술대회 논문집 Ⅳ
    • /
    • pp.2407-2410
    • /
    • 2003
  • In this paper, we propose a new stereoscopic video conversion methodology that converts two-dimensional (2-D) MPEG-4 video to stereoscopic video. In MPEG-4, each Image is composed of background object and primary object. In the first step of the conversion methodology, the camera motion type is determined for stereo Image generation. In the second step, the object-based stereo image generation is carried out. The background object makes use of a current image and a delayed image for its stereo image generation. On the other hand, the primary object uses a current image and its horizontally-shifted version to avoid the possible vertical parallax that could happen. Furthermore, URFA(Uncovered Region Filling Algorithm) is applied in the uncovered region which might be created after the stereo image generation of a primary object. In our experiment, show MPEG-4 test video and its stereoscopic video based upon out proposed methodology and analyze Its results.

  • PDF

FEASIBILITY ON GENERATING STEREO MOSAIC IMAGE

  • Noh, Myoung-Jong;Lee, Sung-Hun;Cho, Woo-Sug
    • 대한원격탐사학회:학술대회논문집
    • /
    • 대한원격탐사학회 2005년도 Proceedings of ISRS 2005
    • /
    • pp.201-204
    • /
    • 2005
  • Recently, the generation of panoramic images and high quality mosaic images from video sequences has been attempted by a variety of investigations. Among a matter of investigation, in this paper, left and right stereo mosaic image generation utilizing airborne-video sequence images is focused upon. The stereo mosaic image is generated by creating left and right mosaic image which is generated by front and rear slit having different viewing angle in consecutive video frame images. The generation of stereo mosaic image proposed in this paper consists of several processes: camera parameter estimation for each video frame image, rectification, slicing, motion parallax elimination and image mosaicking. However it is necessary to check the feasibility on generating stereo mosaic image as explained processes. Therefore, in this paper, we performed the feasibility test on generating stereo mosaic image using video frame images. In doing so, anaglyphic image for stereo mosaic images is generated and tested for feasibility check.

  • PDF

Automatic Poster Generation System Using Protagonist Face Analysis

  • Yeonhwi You;Sungjung Yong;Hyogyeong Park;Seoyoung Lee;Il-Young Moon
    • Journal of information and communication convergence engineering
    • /
    • 제21권4호
    • /
    • pp.287-293
    • /
    • 2023
  • With the rapid development of domestic and international over-the-top markets, a large amount of video content is being created. As the volume of video content increases, consumers tend to increasingly check data concerning the videos before watching them. To address this demand, video summaries in the form of plot descriptions, thumbnails, posters, and other formats are provided to consumers. This study proposes an approach that automatically generates posters to effectively convey video content while reducing the cost of video summarization. In the automatic generation of posters, face recognition and clustering are used to gather and classify character data, and keyframes from the video are extracted to learn the overall atmosphere of the video. This study used the facial data of the characters and keyframes as training data and employed technologies such as DreamBooth, a text-to-image generation model, to automatically generate video posters. This process significantly reduces the time and cost of video-poster production.

자동차 멀티미디어 시스템에서의 사진과 음악을 이용한 음악스토리 비디오 자동생성 기술 (Automatic Music-Story Video Generation Using Music Files and Photos in Automobile Multimedia System)

  • 김형국
    • 한국ITS학회 논문지
    • /
    • 제9권5호
    • /
    • pp.80-86
    • /
    • 2010
  • 본 논문에서는 차량 내의 멀티미디어 시스템에 장착되는 엔터테인먼트 기능 중의 하나인 음악스토리 자동생성 기술을 소개한다. 음악스토리 비디오 자동생성 기술은 개인이 소지하고 있는 휴대폰을 차량 내의 멀티미디어 시스템과 연결하여, 휴대폰 안에 저장된 음악과 사진의 결합을 통해 음악비디오를 자동으로 생성하는 멀티미디어 요소기술로서, 사용자에게 분위기에 맞게 음악을 들으면서 생성된 음악스토리 비디오를 즐기는 기능을 제공한다. 음악스토리 비디오 자동생성 기술에 대한 성능은 음악분류, 사진분류, 핵심단어 검출 등의 정확도와 생성된 음악스토리 비디오를 시청한 사용자의 MOS 결과를 통해 측정되었다.

A Review on Motion Estimation and Compensation for Versatile Video Coding Technology (VVC)

  • Choi, Young-Ju;Kim, Byung-Gyu
    • 한국멀티미디어학회논문지
    • /
    • 제22권7호
    • /
    • pp.770-779
    • /
    • 2019
  • Video coding technologies are progressively becoming more efficient and complex. The Versatile Video Coding (VVC) is a new state-of-the art video compression standard that is going to be a standard, as the next generation of High Efficiency Video Coding (HEVC) standard. To explore the future video coding technologies beyond the HEVC, numerous efficient methods have been adopted by the Joint Video Exploration Team (JVET). Since then, the next generation video coding standard named as VVC and its software model called VVC Test Model (VTM) have emerged. In this paper, several important coding features for motion estimation and motion compensation in the VVC standard is introduced and analyzed in terms of the performance. Improved coding tools introduced for ME and MC in VVC, can achieve much better and good balance between coding efficiency and coding complexity compared with the HEVC.

차선 이탈 경고 시스템의 성능 검증을 위한 가상의 오염 차선 이미지 및 비디오 생성 방법 (Virtual Contamination Lane Image and Video Generation Method for the Performance Evaluation of the Lane Departure Warning System)

  • 곽재호;김회율
    • 한국자동차공학회논문집
    • /
    • 제24권6호
    • /
    • pp.627-634
    • /
    • 2016
  • In this paper, an augmented video generation method to evaluate the performance of lane departure warning system is proposed. In our system, the input is a video which have road scene with general clean lane, and the content of output video is the same but the lane is synthesized with contamination image. In order to synthesize the contamination lane image, two approaches were used. One is example-based image synthesis, and the other is background-based image synthesis. Example-based image synthesis is generated in the assumption of the situation that contamination is applied to the lane, and background-based image synthesis is for the situation that the lane is erased due to aging. In this paper, a new contamination pattern generation method using Gaussian function is also proposed in order to produce contamination with various shape and size. The contamination lane video can be generated by shifting synthesized image as lane movement amount obtained empirically. Our experiment showed that the similarity between the generated contamination lane image and real lane image is over 90 %. Futhermore, we can verify the reliability of the video generated from the proposed method through the analysis of the change of lane recognition rate. In other words, the recognition rate based on the video generated from the proposed method is very similar to that of the real contamination lane video.

실버세대를 위한 동영상 영어사전의 개발 및 평가 (Development and Evaluation of Video English Dictionary for Silver Generation)

  • 김제영;박지수;손진곤
    • 정보처리학회논문지:소프트웨어 및 데이터공학
    • /
    • 제9권11호
    • /
    • pp.345-350
    • /
    • 2020
  • 본 논문은 실버세대 영어학습자를 위한 모바일 학습 콘텐츠를 구현하고 이를 평가하여 이들을 위한 콘텐츠 설계시 고려해야 할 사항에 대해 분석하고자 하였다. 실버세대의 신체적, 학습적 특징과 요구사항 분석을 근거로 하여 영어학습 콘텐츠로 동영상 영어사전을 개발하였고 이를 평가하였다. 동영상 영어사전은 입력방식으로 OCR을, 출력방식으로 동영상을 활용하여 개발하였고 17명의 실버세대들을 대상으로 학업성취도, 학습만족도, 사용의 용이성을 평가하였다. 분석결과 문자 영어사전과 동영상 영어사전 모두 학습만족도가 높은 것으로 나타났으나 학업성취도와 사용의 용이성에서는 문자로 된 영어사전보다 동영상 영어사전이 더 높은 결과를 나타냈다.