• Title/Summary/Keyword: 2D Video

Search Result 910, Processing Time 0.027 seconds

Design of Interactive Operations using Prefetching in VoD System (VoD 시스템에서 선반입 기법을 이용한 대화식 동작의 설계)

  • Kim, Soon-Cheol
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.15 no.2
    • /
    • pp.31-39
    • /
    • 2010
  • VoD(Video-on-Demand) servers have to provide timely processing guarantees for continuous media and reduce the storage and bandwidth requirements for continuous media. The compression techniques make the bit rates of compressed video data significantly variable from frame to frame. A VoD system should be able to provide the client with interactive operations such as fast forward and fast rewind in addition to normal playback of movie. However, interactive operations require additional resources such as storage space, disk bandwidth, memory and network bandwidth. In a stored video application such as VoD system, it is possible that a priori disk access patterns can be used to reserve the system resources in advance. In addition, clients of VoD server spend most of their time in playback mode and the period of time spent in interactive mode is relatively small. In this paper, I present the new buffer management scheme that provides efficient support for interactive operations in a VoD server using variable bit rate continuous media. Simulation results show that our strategy achieves 34% increase of the number of accepted clients over the LRU strategy.

Overlay Text Graphic Region Extraction for Video Quality Enhancement Application (비디오 품질 향상 응용을 위한 오버레이 텍스트 그래픽 영역 검출)

  • Lee, Sanghee;Park, Hansung;Ahn, Jungil;On, Youngsang;Jo, Kanghyun
    • Journal of Broadcast Engineering
    • /
    • v.18 no.4
    • /
    • pp.559-571
    • /
    • 2013
  • This paper has presented a few problems when the 2D video superimposed the overlay text was converted to the 3D stereoscopic video. To resolve the problems, it proposes the scenario which the original video is divided into two parts, one is the video only with overlay text graphic region and the other is the video with holes, and then processed respectively. And this paper focuses on research only to detect and extract the overlay text graphic region, which is a first step among the processes in the proposed scenario. To decide whether the overlay text is included or not within a frame, it is used the corner density map based on the Harris corner detector. Following that, the overlay text region is extracted using the hybrid method of color and motion information of the overlay text region. The experiment shows the results of the overlay text region detection and extraction process in a few genre video sequence.

A Study of Video-Based Abnormal Behavior Recognition Model Using Deep Learning

  • Lee, Jiyoo;Shin, Seung-Jung
    • International journal of advanced smart convergence
    • /
    • v.9 no.4
    • /
    • pp.115-119
    • /
    • 2020
  • Recently, CCTV installations are rapidly increasing in the public and private sectors to prevent various crimes. In accordance with the increasing number of CCTVs, video-based abnormal behavior detection in control systems is one of the key technologies for safety. This is because it is difficult for the surveillance personnel who control multiple CCTVs to manually monitor all abnormal behaviors in the video. In order to solve this problem, research to recognize abnormal behavior using deep learning is being actively conducted. In this paper, we propose a model for detecting abnormal behavior based on the deep learning model that is currently widely used. Based on the abnormal behavior video data provided by AI Hub, we performed a comparative experiment to detect anomalous behavior through violence learning and fainting in videos using 2D CNN-LSTM, 3D CNN, and I3D models. We hope that the experimental results of this abnormal behavior learning model will be helpful in developing intelligent CCTV.

A Study of Mobile Ad-hoc Network Protocols for Ultra Narrowband Video Streaming over Tactical Combat Radio Networks (초협대역 영상전송 전투무선망을 위한 Mobile Ad-hoc Network 프로토콜 연구)

  • Seo, Myunghwan;Kim, Kihun;Ko, Yun-Soo;Kim, Kyungwoo;Kim, Donghyun;Choi, Jeung Won
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.23 no.4
    • /
    • pp.371-380
    • /
    • 2020
  • Video is principal information that facilitates commander's immediate command decision. Due to fading characteristics of radio link, however, it is difficult to stably transmit video in a multi-hop wireless environment. In this paper, we propose a MANET structure composed of a link adaptive routing protocol and a TDMA MAC protocol to stably transmit video traffic in a ultra-narrowband video streaming network. The routing protocol can adapt to link state change and select a stable route. The TDMA protocol enables collision-free video transmission to a destination using multi-hop dynamic resource allocation. As a result of simulation, the proposed MANET structure shows better video transmission performance than proposed MANET structure without link quality adaption, AODV with CSMA/CA, and OLSR with CSMA/CA structures.

Movement Search in Video Stream Using Shape Sequence (동영상에서 모양 시퀀스를 이용한 동작 검색 방법)

  • Choi, Min-Seok
    • Journal of Korea Multimedia Society
    • /
    • v.12 no.4
    • /
    • pp.492-501
    • /
    • 2009
  • Information on movement of objects in videos can be used as an important part in categorizing and separating the contents of a scene. This paper is proposing a shape-based movement-matching algorithm to effectively find the movement of an object in video streams. Information on object movement is extracted from the object boundaries from the input video frames becoming expressed in continuous 2D shape information while individual 2D shape information is converted into a lD shape feature using the shape descriptor. Object movement in video can be found as simply as searching for a word in a text without a separate movement segmentation process using the sequence of the shape descriptor listed according to order. The performance comparison results with the MPEG-7 shape variation descriptor showed that the proposed method can effectively express the movement information of the object and can be applied to movement search and analysis applications.

  • PDF

Virtual View Rendering for 2D/3D Freeview Video Generation (2차원/3차원 자유시점 비디오 재생을 위한 가상시점 합성시스템)

  • Min, Dong-Bo;Sohn, Kwang-Hoon
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.45 no.4
    • /
    • pp.22-31
    • /
    • 2008
  • In this paper, we propose a new approach for efficient multiview stereo matching and virtual view generation, which are key technologies for 3DTV. We propose semi N-view & N-depth framework to estimate disparity maps efficiently and correctly. This framework reduces the redundancy on disparity estimation by using the information of neighboring views. The proposed method provides a user 2D/3D freeview video, and the user can select 2D/3D modes of freeview video. Experimental results show that the proposed method yields the accurate disparity maps and the synthesized novel view is satisfactory enough to provide user seamless freeview videos.

Joint Source/Channel Coding Based on Two-Dimensional Optimization for Scalable H.264/AVC Video

  • Li, Xiao-Feng;Zhou, Ning;Liu, Hong-Sheng
    • ETRI Journal
    • /
    • v.33 no.2
    • /
    • pp.155-162
    • /
    • 2011
  • The scalable extension of the H.264/AVC video coding standard (SVC) demonstrates superb adaptability in video communications. Joint source and channel coding (JSCC) has been shown to be very effective for such scalable video consisting of parts of different significance. In this paper, a new JSCC scheme for SVC transmission over packet loss channels is proposed which performs two-dimensional optimization on the quality layers of each frame in a rate-distortion (R-D) sense as well as on the temporal hierarchical structure of frames under dependency constraints. To compute the end-to-end R-D points of a frame, a novel reduced trellis algorithm is developed with a significant reduction of complexity from the existing Viterbi-based algorithm. The R-D points of frames are sorted under the hierarchical dependency constraints and optimal JSCC solution is obtained in terms of the best R-D performance. Experimental results show that our scheme outperforms the existing scheme of [13] with average quality gains of 0.26 dB and 0.22 dB for progressive and non-progressive modes respectively.

MMT based V3C data packetizing method (MMT 기반 V3C 데이터 패킷화 방안)

  • Moon, Hyeongjun;Kim, Yeonwoong;Park, Seonghwan;Nam, Kwijung;Kim, Kyuhyeon
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2022.06a
    • /
    • pp.836-838
    • /
    • 2022
  • 3D Point Cloud는 3D 콘텐츠를 더욱 실감 나게 표현하기 위한 데이터 포맷이다. Point Cloud 데이터는 3차원 공간상에 존재하는 데이터로 기존의 2D 영상에 비해 거대한 용량을 가지고 있다. 최근 대용량 Point Cloud의 3D 데이터를 압축하기 위해 V-PCC(Video-based Point Cloud Compression)와 같은 다양한 방법이 제시되고 있다. 따라서 Point Cloud 데이터의 원활한 전송 및 저장을 위해서는 V-PCC와 같은 압축 기술이 요구된다. V-PCC는 Point Cloud의 데이터들을 Patch로써 뜯어내고 2D에 Projection 시켜 3D의 영상을 2D 형식으로 변환하고 2D로 변환된 Point Cloud 영상을 기존의 2D 압축 코덱을 활용하여 압축하는 기술이다. 이 V-PCC로 변환된 2D 영상은 기존 2D 영상을 전송하는 방식을 활용하여 네트워크 기반 전송이 가능하다. 본 논문에서는 V-PCC 방식으로 압축한 V3C 데이터를 방송망으로 전송 및 소비하기 위해 MPEG Media Transport(MMT) Packet을 만드는 패킷화 방안을 제안한다. 또한 Server와 Client에서 주고받은 V3C(Visual Volumetric Video Coding) 데이터의 비트스트림을 비교하여 검증한다.

  • PDF

Stereo Video Coding with Spatio-Temporal Scalability for Heterogeneous Collaboration Environments (이질적인 협업환경을 위한 시공간적 계위를 이용한 스테레오 비디오 압축)

  • Oh Sehchan;Lee Youngho;Woo Woontack
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.9
    • /
    • pp.1150-1160
    • /
    • 2004
  • In this paper, we propose a new 3D video coding method for heterogeneous display systems and network infrastructure over enhanced Access Grid (e-AG) using spatio-temporal scalability defined in MPEG-2. The proposed encoder produces several bit-streams for providing temporally and spatially scalable 3D video service. The generated bit-streams can be nelivered with proper spatio-temporal resolution according to network bandwidths and processing speeds, visualization capabilities of client systems. The functionality of proposed spatio-temporal scalability can be exploited for construction of highly scalable 3D video service in heterogeneous distributed environments.

Improving immersive video compression efficiency by reinforcement learning (강화학습 기반 몰입형 영상 압축 성능 향상 기법)

  • Kim, Dongsin;Oh, Byung Tae
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • fall
    • /
    • pp.33-36
    • /
    • 2021
  • In this paper, we propose a new method for improving compression efficiency of immersive video using reinforcement learning. Immersive video means a video that a user can directly experience, such as 3DOF+ videos and Point Cloud videos. It has a vast amount of information due to their characteristics. Therefore, lots of compression methods for immersive video are being studied, and generally, a method, which projects an 3D image into 2D image, is used. However, in this process, a region where information does not exist is created, and it can decrease the compression efficiency. To solve this problem, we propose the reinforcement learning-based filling method with considering the characteristics of images. Experimental results show that the performance is better than the conventional padding method.

  • PDF