• Title/Summary/Keyword: Video content tagging information

Search Result 3, Processing Time 0.015 seconds

Food Media Content Study for an AI Smart Speaker

  • Kim, Kyoung-Ah
    • Journal of Multimedia Information System
    • /
    • v.6 no.4
    • /
    • pp.197-202
    • /
    • 2019
  • Society advances through technology, and technology has changed many lifestyles. The need for food is varying, but the availability of food is constantly changing as trends in production change. Combining the food industry and technology, a robot that delivers food and also cooks it has been developed. The time has come for a combination of food content and technology to advance the restaurant industry. This study discusses the application of a recommended food content media providing system using a curation engine that recommends contents according to individual tastes and preferences for the convenience of those who use food contents, using artificial intelligence speakers. We discuss the technologies required to develop video contents optimized for AI speakers with screens and shapes, combined with inset top boxes.

Character-based Subtitle Generation by Learning of Multimodal Concept Hierarchy from Cartoon Videos (멀티모달 개념계층모델을 이용한 만화비디오 컨텐츠 학습을 통한 등장인물 기반 비디오 자막 생성)

  • Kim, Kyung-Min;Ha, Jung-Woo;Lee, Beom-Jin;Zhang, Byoung-Tak
    • Journal of KIISE
    • /
    • v.42 no.4
    • /
    • pp.451-458
    • /
    • 2015
  • Previous multimodal learning methods focus on problem-solving aspects, such as image and video search and tagging, rather than on knowledge acquisition via content modeling. In this paper, we propose the Multimodal Concept Hierarchy (MuCH), which is a content modeling method that uses a cartoon video dataset and a character-based subtitle generation method from the learned model. The MuCH model has a multimodal hypernetwork layer, in which the patterns of the words and image patches are represented, and a concept layer, in which each concept variable is represented by a probability distribution of the words and the image patches. The model can learn the characteristics of the characters as concepts from the video subtitles and scene images by using a Bayesian learning method and can also generate character-based subtitles from the learned model if text queries are provided. As an experiment, the MuCH model learned concepts from 'Pororo' cartoon videos with a total of 268 minutes in length and generated character-based subtitles. Finally, we compare the results with those of other multimodal learning models. The Experimental results indicate that given the same text query, our model generates more accurate and more character-specific subtitles than other models.

3-tag-based Web Image Retrieval Technique (3-태그 기반의 웹 이미지 검색 기법)

  • Lee, Si-Hwa;Hwang, Dae-Hoon
    • Journal of Korea Multimedia Society
    • /
    • v.15 no.9
    • /
    • pp.1165-1173
    • /
    • 2012
  • One of the most popular technologies in Web2.0 is tagging, and it widely applies to Web content as well as multimedia data such as image and video. Web users have expected that tags by themselves would be reused in information search and maximize the search efficiency, but wrong tag by irresponsible Web users really has brought forth a incorrect search results. In past papers, we have gathered various information resources and tags scattered in Web, mapped one tag onto other tags, and clustered these tags according to the corelation between them. A 3-tag based search algorithm which use the clustered tags of past papers, is proposed in this paper. For performance evaluation of the proposed algorithm, our algorithm is compared with image search result of Flickr, typical tag based site, and is evaluated in accuracy and recall factor.