• Title/Summary/Keyword: Visual Scene

Search Result 369, Processing Time 0.068 seconds

Abrupt Scene Change Detection Algorithm Using Macroblock Type and DC Coefficient in Compressed Domain (압축 도메인 상에서 메크로 블록 타입과 DC 계수를 사용한 급격한 장면 변화 검출 알고리즘)

  • 이흥렬;이웅희;이웅호;정동석
    • Proceedings of the IEEK Conference
    • /
    • 2003.07d
    • /
    • pp.1527-1530
    • /
    • 2003
  • Video is an important and challenge media and requires sophisticated indexing schemes for efficient retrieval from visual databases. Scene change detection is the first step for automatic indexing of video data. Recently, several scene change detection algorithms in the pixel and compressed domains have been reported in the literature. However, using pixel methods are computationally complex and are not very robust in detecting scene change detection. In this paper, we propose robust abrupt scene change detection using macroblock type and DC coefficient. Experimental results show that the proposed algorithm is robust for detection of most abrupt scene changes in the compressed domain.

  • PDF

Scene Recognition Using Local and Global Features (지역적, 전역적 특징을 이용한 환경 인식)

  • Kang, San-Deul;Hwang, Joong-Won;Jung, Hee-Chul;Han, Dong-Yoon;Sim, Sung-Dae;Kim, Jun-Mo
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.15 no.3
    • /
    • pp.298-305
    • /
    • 2012
  • In this paper, we propose an integrated algorithm for scene recognition, which has been a challenging computer vision problem, with application to mobile robot localization. The proposed scene recognition method utilizes SIFT and visual words as local-level features and GIST as a global-level feature. As local-level and global-level features complement each other, it results in improved performance for scene recognition. This improved algorithm is of low computational complexity and robust to image distortions.

GMM-KL Framework for Indoor Scene Matching (실내 환경 이미지 매칭을 위한 GMM-KL프레임워크)

  • Kim, Jun-Young;Ko, Han-Seok
    • Proceedings of the KIEE Conference
    • /
    • 2005.10b
    • /
    • pp.61-63
    • /
    • 2005
  • Retreiving indoor scene reference image from database using visual information is important issue in Robot Navigation. Scene matching problem in navigation robot is not easy because input image that is taken in navigation process is affinly distorted. We represent probabilistic framework for the feature matching between features in input image and features in database reference images to guarantee robust scene matching efficiency. By reconstructing probabilistic scene matching framework we get a higher precision than the existing feaure-feature matching scheme. To construct probabilistic framework we represent each image as Gaussian Mixture Model using Expectation Maximization algorithm using SIFT(Scale Invariant Feature Transform).

  • PDF

Salient Region Extraction based on Global Contrast Enhancement and Saliency Cut for Image Information Recognition of the Visually Impaired

  • Yoon, Hongchan;Kim, Baek-Hyun;Mukhriddin, Mukhiddinov;Cho, Jinsoo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.5
    • /
    • pp.2287-2312
    • /
    • 2018
  • Extracting key visual information from images containing natural scene is a challenging task and an important step for the visually impaired to recognize information based on tactile graphics. In this study, a novel method is proposed for extracting salient regions based on global contrast enhancement and saliency cuts in order to improve the process of recognizing images for the visually impaired. To accomplish this, an image enhancement technique is applied to natural scene images, and a saliency map is acquired to measure the color contrast of homogeneous regions against other areas of the image. The saliency maps also help automatic salient region extraction, referred to as saliency cuts, and assist in obtaining a binary mask of high quality. Finally, outer boundaries and inner edges are detected in images with natural scene to identify edges that are visually significant. Experimental results indicate that the method we propose in this paper extracts salient objects effectively and achieves remarkable performance compared to conventional methods. Our method offers benefits in extracting salient objects and generating simple but important edges from images containing natural scene and for providing information to the visually impaired.

An Optimization Technique of Scene Description for Effective Transmission of Interactive T-DMB Contents (대화형 T-DMB 컨텐츠의 효율적인 전송을 위한 장면기술정보 최적화 기법)

  • Li Song-Lu;Cheong Won-Sik;Jae Yoo-Young;Cha Kyung-Ae
    • Journal of Broadcast Engineering
    • /
    • v.11 no.3 s.32
    • /
    • pp.363-378
    • /
    • 2006
  • The Digital Multimedia Broadcasting(DMB) system is developed to offer high quality audio-visual multimedia contents to the mobile environment. The system adopts MPEG-4 standard for the main video, audio and other media format. It also adopts the MPEG-4 scene description for interactive multimedia contents. The animated and interactive contents can be actualized by BIFS(Binary Format for Scene), the binary format for scene description that refers to the spatio-temporal specifications and behaviors of the individual objects. As more interactive contents are, the scene description is also needed more high bitrate. However, the bandwidth for allocating meta data such as scene description is restrictive in mobile environment. On one hand, the DMB terminal starts demultiplexing content and decodes individual media by its own decoder. After decoding each media, rendering module presents each media stream according to the scene description. Thus the BIFS stream corresponding to the scene description should be decoded and parsed in advance of presenting media data. With these reason, the transmission delay of BIFS stream causes the delay of whole audio-visual scene presentation although the audio or video streams are encoded in very low bitrate. This paper presents the effective optimization technique for adapting BIFS stream into expected MPEG-2 TS bitrate without any bandwidth waste and avoiding the transmission delay of the initial scene description for interactive DMB contents.

Sensory Properties of Visual Scenes Experienced from Different Eye-Heights Arising from Individual Differences in Body-Heights (신장의 개인차로 인한 서로 다른 눈높이에서 경험된 시각장면의 감각적 특성)

  • Kim, Daegyu;Hyun, Joo-Seok
    • Journal of the Korea Convergence Society
    • /
    • v.9 no.11
    • /
    • pp.217-225
    • /
    • 2018
  • Different eye-heights due to individuals' body heights may cause different sensory experiences against the same visual scene, eventually leading to their longer-term psycho-social and developmental individual differences. Accordingly, the present study compared sensory properties of photographs for the same scene taken from two different camera-heights (i.e., eye-heights). Two sets of photographs were taken in parallel from two cameras attached to a different height on the same pedestrian's body. Analysis of the photographs revealed that both the levels of visual saliency and complexity were greater for the photographs taken from the high eye-height than those from the low eye-height. The results indicate a possible difference in sensory properties of visual scenes perceived from two different heights, potentially exposing taller individuals to richer and more diverse sensory experiences than shorter individuals.

MPEG-4 BIFS Optimization for Interactive T-DMB Content (지상파 DMB 컨텐츠의 MPEG-4 BIFS 최적화 기법)

  • Cha, Kyung-Ae
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.12 no.1
    • /
    • pp.54-60
    • /
    • 2007
  • The Digital Multimedia Broadcasting(DMB) system is developed to offer high quality multimedia content to the mobile environment. The system adopts the MPEG-4 standard for the main video, audio and other media format. For providing interactive contents, it also adopts the MPEG-4 scene description that refers to the spatio-temporal specifications and behaviors of individual objects. With more interactive contents, the scene description also needs higher bitrate. However, the bandwidth for allocating meta data, such as scene description is restrictive in the mobile environment. On one hand, the DMB terminal renders each media stream according to the scene description. Thus the binary format for scene(BIFS) stream corresponding to the scene description should be decoded and parsed in advance when presenting media data. With this reasoning, the transmission delay of the BIFS stream would cause the delay in transmitting whole audio-visual scene presentations, although the audio or video streams are encoded in very low bitrate. This paper presents the effective optimization technique in adapting the BIFS stream into the expected bitrate without any waste in bandwidth and avoiding transmission delays inthe initial scene description for interactive DMB content.

  • PDF

An Effective Moving Cast Shadow Removal in Gray Level Video for Intelligent Visual Surveillance (지능 영상 감시를 위한 흑백 영상 데이터에서의 효과적인 이동 투영 음영 제거)

  • Nguyen, Thanh Binh;Chung, Sun-Tae;Cho, Seongwon
    • Journal of Korea Multimedia Society
    • /
    • v.17 no.4
    • /
    • pp.420-432
    • /
    • 2014
  • In detection of moving objects from video sequences, an essential process for intelligent visual surveillance, the cast shadows accompanying moving objects are different from background so that they may be easily extracted as foreground object blobs, which causes errors in localization, segmentation, tracking and classification of objects. Most of the previous research results about moving cast shadow detection and removal usually utilize color information about objects and scenes. In this paper, we proposes a novel cast shadow removal method of moving objects in gray level video data for visual surveillance application. The proposed method utilizes observations about edge patterns in the shadow region in the current frame and the corresponding region in the background scene, and applies Laplacian edge detector to the blob regions in the current frame and the corresponding regions in the background scene. Then, the product of the outcomes of application determines moving object blob pixels from the blob pixels in the foreground mask. The minimal rectangle regions containing all blob pixles classified as moving object pixels are extracted. The proposed method is simple but turns out practically very effective for Adative Gaussian Mixture Model-based object detection of intelligent visual surveillance applications, which is verified through experiments.

A Visual Effect Retrieval System Design for Communication in Film-production - Focused on the Effect Using Computer Graphics Technology - (영화 비주얼 이펙트 제작의 커뮤니케이션을 위한 자료검색 시스템 제안 - 컴퓨터 그래픽 기술을 이용한 이펙트를 중심으로 -)

  • Jo, Kook-Jung;Suk, Hae-Jung
    • The Journal of the Korea Contents Association
    • /
    • v.9 no.6
    • /
    • pp.92-103
    • /
    • 2009
  • With the help of computer graphics technologies, the visual effects techniques using these technologies replaced most of special effects techniques which had been used for early films. For these changes, directors and visual effects creators make an effect in a scene through their mutual agreement in contemporary films. However, they undergo a lot of trial-and-error while making a visual effects scene because they cannot perfectly communicate their ideas due to the director's narrative language, and also because of the visual effect creator's language of computer graphics technology. This research suggests the design of a visual effects data retrieval system for efficient communication between directors and visual effects creators. This application provides the means to search a database analyzing visual effects scenes extracted from 14 remarkable movies in visual effect history by narrative and visual effects technique. They can search visual effects scenes using this application. also, this data can foster communication with directors and creators so they can make an efficient production pipeline.

Application of Virtual Studio Technology and Digital Human Monocular Motion Capture Technology -Based on <Beast Town> as an Example-

  • YuanZi Sang;KiHong Kim;JuneSok Lee;JiChu Tang;GaoHe Zhang;ZhengRan Liu;QianRu Liu;ShiJie Sun;YuTing Wang;KaiXing Wang
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.16 no.1
    • /
    • pp.106-123
    • /
    • 2024
  • This article takes the talk show "Beast Town" as an example to introduce the overall technical solution, technical difficulties and countermeasures for the combination of cartoon virtual characters and virtual studio technology, providing reference and experience for the multi-scenario application of digital humans. Compared with the live broadcast that combines reality and reality, we have further upgraded our virtual production technology and digital human-driven technology, adopted industry-leading real-time virtual production technology and monocular camera driving technology, and launched a virtual cartoon character talk show - "Beast Town" to achieve real Perfectly combined with virtuality, it further enhances program immersion and audio-visual experience, and expands infinite boundaries for virtual manufacturing. In the talk show, motion capture shooting technology is used for final picture synthesis. The virtual scene needs to present dynamic effects, and at the same time realize the driving of the digital human and the movement with the push, pull and pan of the overall picture. This puts forward very high requirements for multi-party data synchronization, real-time driving of digital people, and synthetic picture rendering. We focus on issues such as virtual and real data docking and monocular camera motion capture effects. We combine camera outward tracking, multi-scene picture perspective, multi-machine rendering and other solutions to effectively solve picture linkage and rendering quality problems in a deeply immersive space environment. , presenting users with visual effects of linkage between digital people and live guests.