• Title/Summary/Keyword: Video sequence

Search Result 507, Processing Time 0.023 seconds

Automatic Cast-list Analysis System in Broadcasting Videos (방송 비디오 등장인물 자동 분석 시스템)

  • 김기남;김형준;김회율
    • Journal of Broadcast Engineering
    • /
    • v.9 no.2
    • /
    • pp.164-173
    • /
    • 2004
  • In this paper, we propose a system that can analyze appearance interval of casts by detecting and recognizing casts in broadcasting videos. The cast is one of the most important characteristics in broadcasting videos such as drama and sports. In this paper, we propose the ACAV(Automatic Cast-list Analysis in Videos) system that analyzes cast-list automatically in video. The ACAV system consists of FAGIS(FAce reGIStration) which registers detected faces into the face DB and FACOG(FAce reCOGnition) that analyses the cast-list in video sequence using the face DB. We evaluate performance of the ACAV system by comparing with FaceIt, one of the most well-known commercial systems for the cast-list analysis. The ACAV shows face detection and recognition rates of 84.3% and 75.7% that are about 30% and 27.5% higher than those of FaceIt, respectively. The ACAV system can be applied to mass broadcasting videos management system for broadcasters and video management system of PVR(Personal Video Recorder) and mobile phone for the public.

An Efficient Transmission Method of Panoramic Multimedia Contents in a Limited Bandwidth Environment (제한적 네트워크 환경 하에서 효율적인 파노라마식 멀티미디어 콘텐츠 분할 전송 방법)

  • Kim, Byung-Chul;Lee, Gun-Hee;Lee, In-Jae;Kim, Kyu-Heon
    • Journal of Broadcast Engineering
    • /
    • v.16 no.5
    • /
    • pp.811-823
    • /
    • 2011
  • This paper proposes an efficient transmission method for the panoramic multimedia contents. The panoramic video provides wide sight and various view-point to the user. The traditional methods of the panoramic multimedia content transmission has several limitations, as follow; A client suffers a long initial delay time to play a panoramic video when it is transmitted through a limited bandwidth network, because the panoramic video has larger data size than a general video. And if a client's display device has limited resolution, such as mobile phone, laptop PC monitor, etc. it can not display the entire panoramic video that has a wide view video sequence. So, in order to overcome the obstacles, this paper proposes an efficient transmission of panoramic multimedia contents. This method will increase the transmission efficiency throughout the technique of the scene description in MPEG-4 system. Also we demonstrated the efficiency of the proposed method by comparison with existing methods.

Affine Model for Generating Stereo Mosaic Image from Video Frames (비디오 프레임 영상의 자유 입체 모자이크 영상 제작을 위한 부등각 모델 연구)

  • Noh, Myoung-Jong;Cho, Woo-Sug;Park, Jun-Ku;Koh, Jin-Woo
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.17 no.3
    • /
    • pp.49-56
    • /
    • 2009
  • Recently, a generation of high quality mosaic images from video sequences has been attempted by a variety of investigations. Among the matter of investigation, in this paper, generation on stereo mosaic utilizing airborne-video sequence images is focused upon. The stereo mosaic is made by creating left and right mosaic which are fabricated by front and rear slices having different viewing angle in consecutive video frames. For making the stereo mosaic, motion parameters which are able to define geometric relationship between consecutive video frames are determined. For determining motion parameters, affine model which is able to explain relative motion parameters is applied by this paper. The mosaicing method using relative motion parameters is called by free mosaic. The free mosaic proposed in this paper consists of 4 step processes: image registration with reference to first frame using affine model, front and rear slicing, stitching line definition and image mosaicing. As the result of experiment, the left and right mosaic image, anaglyphic image for stereo mosaic images are showed and analyzed y-parallax for checking accuracy.

  • PDF

A Study on Channel Decoder MAP Estimation Based on H.264 Syntax Rule (H-264 동영상 압축의 문법적 제한요소를 이용한 MAP기반의 Channel Decoder 성능 향상에 대한 연구)

  • Jeon, Yong-Jin;Seo, Dong-Wan;Choe, Yun-Sik
    • Proceedings of the KIEE Conference
    • /
    • 2003.11b
    • /
    • pp.295-298
    • /
    • 2003
  • In this paper, a novel maximum a posterion (MAP) estimation for the channel decoding of H.264 codes in the presence of transmission error is presented. Arithmetic codes with a forbidden symbol and trellis search techniques are employed in order to estimate the best transmitted. And, there has been growing interest of communication, the research about transmission of exact data is increasing. Unlike the case of voice transmission, noise has a fatal effect on the image transmission. The reason is that video coding standards have used the variable length coding. So, only one bit error affects the all video data compressed before resynchronization. For reasons of that, channel needs the channel codec, which is robust to channel error. But, usual channel decoder corrects the error only by channel error probability. So, designing source codec and channel codec, Instead of separating them, it is tried to combine them jointly. And many researches used the information of source redundancy In received data. But, these methods do not match to the video coding standards, because video ceding standards use not only one symbol but also many symbols in same data sequence. In this thesis, We try to design combined source-channel codec that is compatible with video coding standards. This MAP decoder is proposed by adding semantic structure and semantic constraint of video coding standards to the method using redundancy of the MAP decoders proposed previously. Then, We get the better performance than usual channel coder's.

  • PDF

Key Frame Extraction and Region Segmentation-based Video Retrieval in Compressed Domain (압축영역에서의 대표프레임 추출 및 영역분할기반 비디오 검색 기법)

  • 강응관;김성주;송호근;최종수
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.9B
    • /
    • pp.1713-1720
    • /
    • 1999
  • This paper presents a new key frame extraction technique, for scene change detection, using the proposed AHIM (Accumulative Histogram Intersection Measure) from the DC image constructed by DCT DC coefficients in the compressed video sequence that is video compression standard such as MPEG. For fast content-based browsing and video retrieval in a video database, we also provide a novel coarse-to-fine video indexing scheme. In the extracted key frame, we perform the region segmentation as a preprocessing. First, the segmented image is projected with the horizontal direction, then we transform the result into a histogram, which is saved as a database index. In the second step, we calculate the moments and change them into a distance value. From the simulation results, the proposed method clearly shows the validity and superiority in respect of computation time and memory space, and that in conjunction with other techniques for indexing, such as color, can provide a powerful framework for image indexing and retrieval.

  • PDF

Dual-stream Co-enhanced Network for Unsupervised Video Object Segmentation

  • Hongliang Zhu;Hui Yin;Yanting Liu;Ning Chen
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.4
    • /
    • pp.938-958
    • /
    • 2024
  • Unsupervised Video Object Segmentation (UVOS) is a highly challenging problem in computer vision as the annotation of the target object in the testing video is unknown at all. The main difficulty is to effectively handle the complicated and changeable motion state of the target object and the confusion of similar background objects in video sequence. In this paper, we propose a novel deep Dual-stream Co-enhanced Network (DC-Net) for UVOS via bidirectional motion cues refinement and multi-level feature aggregation, which can fully take advantage of motion cues and effectively integrate different level features to produce high-quality segmentation mask. DC-Net is a dual-stream architecture where the two streams are co-enhanced by each other. One is a motion stream with a Motion-cues Refine Module (MRM), which learns from bidirectional optical flow images and produces fine-grained and complete distinctive motion saliency map, and the other is an appearance stream with a Multi-level Feature Aggregation Module (MFAM) and a Context Attention Module (CAM) which are designed to integrate the different level features effectively. Specifically, the motion saliency map obtained by the motion stream is fused with each stage of the decoder in the appearance stream to improve the segmentation, and in turn the segmentation loss in the appearance stream feeds back into the motion stream to enhance the motion refinement. Experimental results on three datasets (Davis2016, VideoSD, SegTrack-v2) demonstrate that DC-Net has achieved comparable results with some state-of-the-art methods.

New Illumination compensation algorithm improving a multi-view video coding performance by advancing its temporal and inter-view correlation (다시점 비디오의 시공간적 중복도를 높여 부호화 성능을 향상시키는 새로운 조명 불일치 보상 기법)

  • Lee, Dong-Seok;Yoo, Ji-Sang
    • Journal of Broadcast Engineering
    • /
    • v.15 no.6
    • /
    • pp.768-782
    • /
    • 2010
  • Because of the different shooting position between multi-view cameras and the imperfect camera calibration, Illumination mismatches of multi-view video can happen. This variation can bring about the performance decrease of multi-view video coding(MVC) algorithm. A histogram matching algorithm can be applied to recompensate these inconsistencies in a prefiltering step. Once all camera frames of a multi-view sequence are adjusted to a predefined reference through the histogram matching, the coding efficiency of MVC is improved. However the histogram distribution can be different not only between neighboring views but also between sequential views on account of movements of camera angle and some objects, especially human. Therefore the histogram matching algorithm which references all frames in chose view is not appropriate for compensating the illumination differences of these sequence. Thus we propose new algorithms both the image classification algorithm which is applied two criteria to improve the correlation between inter-view frames and the histogram matching which references and matches with a group of pictures(GOP) as a unit to advance the correlation between successive frames. Experimental results show that the compression ratio for the proposed algorithm is improved comparing with the conventional algorithms.

Camera Motion Estimation using Geometrically Symmetric Points in Subsequent Video Frames (인접 영상 프레임에서 기하학적 대칭점을 이용한 카메라 움직임 추정)

  • Jeon, Dae-Seong;Mun, Seong-Heon;Park, Jun-Ho;Yun, Yeong-U
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.39 no.2
    • /
    • pp.35-44
    • /
    • 2002
  • The translation and the rotation of camera occur global motion which affects all over the frame in video sequence. With the video sequences containing global motion, it is practically impossible to extract exact video objects and to calculate genuine object motions. Therefore, high compression ratio cannot be achieved due to the large motion vectors. This problem can be solved when the global motion compensated frames are used. The existing camera motion estimation methods for global motion compensation have a large amount of computations in common. In this paper, we propose a simple global motion estimation algorithm that consists of linear equations without any repetition. The algorithm uses information .of symmetric points in the frame of the video sequence. The discriminant conditions to distinguish regions belonging to distant view from foreground in the frame are presented. Only for the distant view satisfying the discriminant conditions, the linear equations for the panning, tilting, and zooming parameters are applied. From the experimental results using the MPEG test sequences, we can confirm that the proposed algorithm estimates correct global motion parameters. Moreover the real-time capability of the proposed technique can be applicable to many MPEG-4 and MPEG-7 related areas.

A Content-based Video Rate-control Algorithm Interfaced to Human-eye (인간과 결합한 내용기반 동영상 율제어)

  • 황재정;진경식;황치규
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.28 no.3C
    • /
    • pp.307-314
    • /
    • 2003
  • In the general multiple video object coder, more interested objects such as speaker or moving object is consistently coded with higher priority. Since the priority of each object may not be fixed in the whole sequence and be variable on frame basis, it must be adjusted in a frame. In this paper, we analyze the independent rate control algorithm and global algorithm that the QP value is controled by the static parameters, object importance or priority, target PSNR, weighted distortion. The priority among static parameters is analyzed and adjusted into dynamic parameters according to the visual interests or importance obtained by camera interface. Target PSNR and weighted distortion are proportionally derived by using magnitude, motion, and distortion. We apply those parameters for the weighted distortion control and the priority-based control resulting in the efficient bit-rate distribution. As results of this paper, we achieved that fewer bits are allocated for video objects which has less importance and more bits for those which has higher visual importance. The duration of stability in the visual quality is reduced to less than 15 frames of the coded sequence. In the aspect of PSNR, the proposed scheme shows higher quality of more than 2d13 against the conventional schemes. Thus the coding scheme interfaced to human- eye proves an efficient video coder dealing with the multiple number of video objects.

Multiview Video Sequence CODEC with View Scalability (View Scalability를 고려한 다시점 동영상 코덱)

  • 임정은;손광훈
    • Journal of Broadcast Engineering
    • /
    • v.9 no.3
    • /
    • pp.236-245
    • /
    • 2004
  • A multiview sequence CODEC with view scaiability is proposed in this paper. We define a GGOP (Group of GOP) structure as a basic coding unit to efficiently code multiview sequences. 7he proposed CODEC provides flexible GGOP structures based on the number of views and baseline distances among cameras. Multiview sequences encode consists of disparity estimation/compensation, motion estimation/compensation, residual coding and rate control and generates multiview sequence bitstream. The main bitstream is the same as an MPEG-2 mono-sequence bitstream for MPEG-2 compatibility. The auxiliary bitstream contains information concerning the remaining multiview sequences except for the reference sequences. The proposed CODEC with view scalability provides that a number of view flints are selectively determined at the receiver according to the type of display modes. The proposed multiview sequence CODEC is tested with several multiview sequences to determine its flexibility. compatibility with MPEG-2 and view scaiability. In addition, we subjectively confirm that the decoded bitstreams with view scaiability can be Properly displayed by several types of display modes. including 3D monitors.