• Title/Summary/Keyword: 장면 구성

Search Result 357, Processing Time 0.028 seconds

Scene Change Detection Techniques Using DC components and Moving Vector in DCT-domain of MPEG systems (MPEG system의 DCT변환영역에서 DC성분과 움직임 벡터를 이용한 영상 장면전환 검출기법)

  • 박재두;이광형
    • Journal of the Korea Society of Computer and Information
    • /
    • v.4 no.3
    • /
    • pp.28-34
    • /
    • 1999
  • In this paper. we propose the method of Scene Change Detection for video sequence using the DC components and the moving vectors in the Macro Blocks in the DCT blocks. The proposed method detects the Scene Change which would not be related with the specific sequences in the compressed MPEG domain. To do this. we define new metrics for Scene Change Detection using the features of picture component and detect the exact Scene Change point of B-pictures using the characteristics of B-picture's sharp response for the moving vectors. In brief, we will detect the cut point using I-picture and the gradual scene changes such as dissolve, fade, wipe, etc. As a results, our proposed method shows good test results for the various MPEG sequences.

Text Extraction using Character-Edge Map Feature From Scene Images (장면 이미지로부터 문자-에지 맵 특징을 이용한 텍스트 추출)

  • Park, Jong-Cheon;Hwang, Dong-Guk;Lee, Woo-Ram;Kwon, Kyo-Hyun;Jun, Byoung-Min
    • Proceedings of the KAIS Fall Conference
    • /
    • 2006.05a
    • /
    • pp.139-142
    • /
    • 2006
  • 본 연구는 장면 이미지로부터 텍스트에 존재하는 문자-에지 특징을 이용하여 텍스트를 추출하는 방법을 제안한다. 캐니(Canny)에지 연산자를 이용하여 장면 이미지로부터 에지를 추출하고, 추출된 에지로부터 16종류의 에지-맵 생성한다. 생성된 에지 맵을 재구성하여 문자 특징을 갖는 8종류의 문자-에지 맵을 만단다. 텍스트는 배경과 잘 분리되는 특징이 있으므로 텍스트에 존재하는 '문자-에지 맵'의 특징을 이용하여 텍스트를 추출한다. 텍스트 영역에 대한 검증은 문자-에지 맵의 분포와 텍스트에 존재하는 글자간의 공백 특징으로 한다. 제안한 방법은 다양한 종류의 장면 이미지를 실험대상으로 하였고, 텍스트는 적어도 2글자 이상으로 구성된다는 제한조건과 너무 크거나 작은 텍스트는 텍스트 추출에서 제외하였다. 실험결과 텍스트 영역 추출률은 약 83%를 얻었다.

  • PDF

A case study on using MMT for transmission and synchronization of UWV content service (UWV 콘텐츠의 전송과 동기화를 위한 MMT 활용사례 연구)

  • Rhyu, Sung-yeul;An, Hyo-Min;Kim, Kyu-Heon
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2017.06a
    • /
    • pp.127-130
    • /
    • 2017
  • 본 논문은 UWV 서비스의 다양한 시나리오들을 소개하고 UWV서비스의 제공측 구성과 소비측 재생에 따라 개시되는 멀티미디어 콘텐츠 파일 전송의 형태를 검토한다. UWV 서비스의 전송 개시에 따라 적어도 하나 이상의 서비스 구성요소(멀티미디어 콘텐츠 파일)가 전송될 때 수신측에서의 엄밀한 동기화가 수행되지 못하면 전체 장면의 일부를 표시하는 View간의 일치가 이루어지지 않으며 이로 인해 사용자의 경험 측면에서 임장감이 낮아지거나 또는 불일치의 정도가 심한 경우 전체 장면을 이해하는데 어려움이 발생한다. 이에 본 논문에서는 기존의 전송 방식인 TS와 최근 MPEG에서 완료된 MMT를 사용해 UWV 서비스를 구성하는 콘텐츠의 전송시, 소비측에서의 동기화된 View 콘텐츠 재생을 달성하기 위한 전송 방법을 알아본다. 또한 시스템 구성을 통해 실제 전송시 요구사항의 달성여부와 함께 추가로 고려해야 할 사항들을 알아본다.

  • PDF

Semantic-based Scene Retrieval Using Ontologies for Video Server (비디오 서버에서 온톨로지를 이용한 의미기반 장면 검색)

  • Jung, Min-Young;Park, Sung-Han
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.45 no.5
    • /
    • pp.32-37
    • /
    • 2008
  • To ensure access to rapidly growing video collection, video indexing is becoming more and more important. In this paper, video ontology system for retrieving a video data based on a scene unit is proposed. The proposed system creates a semantic scene as a basic unit of video retrieval, and limits a domain of retrieval through a subject of that scene. The content of semantic scene is defined using the relationship between object and event included in the key frame of shots. The semantic gap between the low level feature and the high level feature is solved through the scene ontology to ensure the semantic-based retrieval.

Deep Neural Network-Based Scene Graph Generation for 3D Simulated Indoor Environments (3차원 가상 실내 환경을 위한 심층 신경망 기반의 장면 그래프 생성)

  • Shin, Donghyeop;Kim, Incheol
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.8 no.5
    • /
    • pp.205-212
    • /
    • 2019
  • Scene graph is a kind of knowledge graph that represents both objects and their relationships found in a image. This paper proposes a 3D scene graph generation model for three-dimensional indoor environments. An 3D scene graph includes not only object types, their positions and attributes, but also three-dimensional spatial relationships between them, An 3D scene graph can be viewed as a prior knowledge base describing the given environment within that the agent will be deployed later. Therefore, 3D scene graphs can be used in many useful applications, such as visual question answering (VQA) and service robots. This proposed 3D scene graph generation model consists of four sub-networks: object detection network (ObjNet), attribute prediction network (AttNet), transfer network (TransNet), relationship prediction network (RelNet). Conducting several experiments with 3D simulated indoor environments provided by AI2-THOR, we confirmed that the proposed model shows high performance.

Development of Intuitive Author for 3D Game/Animation Contents (효율적인 3D 게임 및 애니메이션 콘텐츠 제작을 위한 직관적인 저작 기술 개발)

  • Ji, Su-Mi;Lee, Jeong-Joong;Kim, Sung-Gook;Woo, Kyeong-Deok;Baik, Sung-Wook
    • Journal of Korea Multimedia Society
    • /
    • v.13 no.5
    • /
    • pp.780-791
    • /
    • 2010
  • This paper presents a semi-autonomous authoring method which can intuitively create 3D game/ animation contents for unexperienced users. In particular, this method considers their convenience to provide them with easier authoring environments. The 3D game/animation contents created by this method consist of several scene components connected to each other in a logical flow structure. Thus, it is possible to easily understand the overview of 3D content description for its production and manipulate these scene components for efficient management. Four different genres of game/animation contents have been developed by using an authoring tool based on this method. According to results of a survey intended for game developers who created game contents with the authoring tool, this method is shown to save contents development cost/time as well as raise the efficiency and convenience of whole process of contents creation to the expected level.

Video-Dissolve Detection using Characteristics of Neighboring Scenes (이웃 장면들의 특성을 이용한 비디오 디졸브 검출)

  • 원종운;최재각;박철현;김범수;곽동민;오상근;박길흠
    • Journal of KIISE:Information Networking
    • /
    • v.30 no.4
    • /
    • pp.504-512
    • /
    • 2003
  • In this paper, we propose a new adaptive dissolve detection method based on the analysis of a dissolve modeling error which is the difference between an ideally modeled dissolve curve with no correlation and an actual dissolve curve including a correlation. The proposed dissolve detection method consists of two steps. First, candidate dissolve regions are extracted using the characteristics of a downward convex parabola, then each candidate region is verified based oil the dissolve modeling error. If the dissolve modeling error for a candidate region is less than a threshold defined by the target modeling error with a target correlation, the candidate region is determined as a resolve region with a lower correlation than the target correlation. The threshold is adaptively determined based on the variances between the candidate regions and the target correlation. By considering the correlation between neighbor scenes, the proposed method is able to be a semantic scene-change detector. The proposed method was tested on various types of data and its performance proved to be more accurate and reliable regardless of variation of variance of test sequences when compared with other commonly use methods.

The Design of Object-based 3D Audio Broadcasting System (객체기반 3차원 오디오 방송 시스템 설계)

  • 강경옥;장대영;서정일;정대권
    • The Journal of the Acoustical Society of Korea
    • /
    • v.22 no.7
    • /
    • pp.592-602
    • /
    • 2003
  • This paper aims to describe the basic structure of novel object-based 3D audio broadcasting system To overcome current uni-directional audio broadcasting services, the object-based 3D audio broadcasting system is designed for providing the ability to interact with important audio objects as well as realistic 3D effects based on the MPEG-4 standard. The system is composed of 6 sub-modules. The audio input module collects the background sound object, which is recored by 3D microphone, and audio objects, which are recorded by monaural microphone or extracted through source separation method. The sound scene authoring module edits the 3D information of audio objects such as acoustical characteristics, location, directivity and etc. It also defines the final sound scene with a 3D background sound, which is intended to be delievered to a receiving terminal by producer. The encoder module encodes scene descriptors and audio objects for effective transmission. The decoder module extracts scene descriptors and audio objects from decoding received bistreams. The sound scene composition module reconstructs the 3D sound scene with scene descriptors and audio objects. The 3D sound renderer module maximizes the 3D sound effects through adapting the final sound to the listner's acoustical environments. It also receives the user's controls on audio objects and sends them to the scene composition module for changing the sound scene.

A Scene-based Tree Browsing Technique for Video Retrieval and Visual Summary (비디오 검색과 시각적 요약을 위한 장면 기반 계층적 브라우징 기법)

  • Im, Dong-Hyeok;Lee, Seok-Ryong;Jeong, Jin-Wan
    • Journal of KIISE:Databases
    • /
    • v.28 no.2
    • /
    • pp.181-187
    • /
    • 2001
  • 디지털 비디오의 사용이 일반화되어감에 따라 비디오 브라우징에 대한 연구가 더욱 요구되 어 지고 있다. 과거에 연구된 바 있는 VCR식 순차탐색기법은 아날로그 비디오 브라우징에 서 사용되던 고전적 방식을 다지털 비디오에 적용한 것이며, 키프레임 정적표현법은 비디오 를 구성하는 프레임을 보이는 방식이다. 이 두 방식이 디지털 비디오의 특성을 충분히 사용 하지 못하기 때문에, 최근에는 비디오 샷산의 계층적 관계를 기반으로 하는 계층적 브라우 징과 장면 간의 관계를 기반으로 하는 장면 기반 브라우징이 주목을 받고 있다. 본 논문에 서는 기존의 비디오 브라우징에 대한 연구들을 자세히 알아보고, 계층적 브라우징과 장면 기반 브라우징에서처럼 비디오 내의 각 장면에 바로 접근할 수 있을 뿐만 아니라, 계층적 브라우징처럼 비디오의 전체구조를 알기 쉽게 보여준다. 또한 브라우징의 결과는 시각적 요 약으로 사용될 수 있다.

  • PDF

The Method of Multi-screen Service using Scene Composition Technology based on HTML5 (HTML5 기반 장면구성 기술을 통한 멀티스크린 서비스 제공 방법)

  • Jo, Minwoo;Kim, Kyuheon
    • Journal of Broadcast Engineering
    • /
    • v.18 no.6
    • /
    • pp.895-910
    • /
    • 2013
  • Multi-screen service is a service that consumes more than one media in a number of terminals simultaneously or discriminately. This multi-screen service has become useful due to distribute of smart TV and terminals. Also, in case of hybrid broadcasting environment that is convergence of broadcasting and communication environment, it is able to provide various user experience through contents consumed by multiple screens. In hybrid broadcasting environment, scene composition technology can be used as an element technology for multi-screen service. Using scene composition technology, multiple media can be consumed complexly through the specified presentation time and space. Thus, multi-screen service based on the scene composition technology can provide spatial and temporal control and consumption of multiple media by linkage between the terminals. However, existing scene composition technologies are not able to use easily in hybrid broadcasting because of applicable environmental constraints, the difficulty in applying the various terminal and complexity. For this problems, HTML5 can be considered. HTML5 is expected to be applied in various smart terminals commonly, and provides consumption of diverse media. So, in this paper, it proposes the scene composition and multi-screen service technology based on HTML5 that is expected be used in various smart terminals providing hybrid broadcasting environment. For this, it includes the introduction in terms of HTML5 and multi-screen service, the method of providing information related with scene composition and multi-screen service through the extention of elements and attributes in HTML5, media signaling between terminals and the method of synchronization. In addition, the proposed scene composition and multi-screen service technology based on HTML5 was verified through the implementation and experiment.