• Title/Summary/Keyword: Video sequence

Search Result 507, Processing Time 0.025 seconds

View synthesis with sparse light field for 6DoF immersive video

  • Kwak, Sangwoon;Yun, Joungil;Jeong, Jun-Young;Kim, Youngwook;Ihm, Insung;Cheong, Won-Sik;Seo, Jeongil
    • ETRI Journal
    • /
    • v.44 no.1
    • /
    • pp.24-37
    • /
    • 2022
  • Virtual view synthesis, which generates novel views similar to the characteristics of actually acquired images, is an essential technical component for delivering an immersive video with realistic binocular disparity and smooth motion parallax. This is typically achieved in sequence by warping the given images to the designated viewing position, blending warped images, and filling the remaining holes. When considering 6DoF use cases with huge motion, the warping method in patch unit is more preferable than other conventional methods running in pixel unit. Regarding the prior case, the quality of synthesized image is highly relevant to the means of blending. Based on such aspect, we proposed a novel blending architecture that exploits the similarity of the directions of rays and the distribution of depth values. By further employing the proposed method, results showed that more enhanced view was synthesized compared with the well-designed synthesizers used within moving picture expert group (MPEG-I). Moreover, we explained the GPU-based implementation synthesizing and rendering views in the level of real time by considering the applicability for immersive video service.

Generation of high cadence SDO/AIA images using a video frame interpolation method, SuperSloMo

  • Sung, Suk-Kyung;Shin, Seungheon;Kim, TaeYoung;Lee, Jin-Yi;Park, Eunsu;Moon, Yong-Jae;Kim, Il-Hoon
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.44 no.2
    • /
    • pp.44.1-44.1
    • /
    • 2019
  • We generate new intermediate images between observed consecutive solar images using NVIDIA's SuperSloMo that is a novel video interpolation method. This technique creates intermediate frames between two successive frames to form a coherent video sequence for both spatially and temporally. By using SuperSloMo, we create 600 images (12-second interval) using the observed 121 SDO/AIA 304 Å images (1-minute interval) of a filament eruption event on December 3, 2012. We compare the generated images with the original 12-second images. For the generated 480 images the correlation coefficient (CC), the relative error (R1), and the normalized mean square error (R2) are 0.99, 0.40, and 0.86, respectively. We construct a video made of the generated images and find a smoother erupting movement. In addition, we generate nonexistent 2.4-second interval images using the original 12-second interval images, showing slow motions in the eruption. We will discuss possible applications of this method.

  • PDF

Violent crowd flow detection from surveillance cameras using deep transfer learning-gated recurrent unit

  • Elly Matul Imah;Riskyana Dewi Intan Puspitasari
    • ETRI Journal
    • /
    • v.46 no.4
    • /
    • pp.671-682
    • /
    • 2024
  • Violence can be committed anywhere, even in crowded places. It is hence necessary to monitor human activities for public safety. Surveillance cameras can monitor surrounding activities but require human assistance to continuously monitor every incident. Automatic violence detection is needed for early warning and fast response. However, such automation is still challenging because of low video resolution and blind spots. This paper uses ResNet50v2 and the gated recurrent unit (GRU) algorithm to detect violence in the Movies, Hockey, and Crowd video datasets. Spatial features were extracted from each frame sequence of the video using a pretrained model from ResNet50V2, which was then classified using the optimal trained model on the GRU architecture. The experimental results were then compared with wavelet feature extraction methods and classification models, such as the convolutional neural network and long short-term memory. The results show that the proposed combination of ResNet50V2 and GRU is robust and delivers the best performance in terms of accuracy, recall, precision, and F1-score. The use of ResNet50V2 for feature extraction can improve model performance.

Construction of the Facilities Management System by Video Structuring (동영상자료 구조화에 의한 시설물관리시스템 구축)

  • Yoo, Hwan-Hee;Choi, Kyoung-Ho;Koo, Heung-Dae
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.12 no.4 s.31
    • /
    • pp.69-74
    • /
    • 2004
  • By the expanding of infrastructure caused by urbanization, new technologies are required to manage various kinds of facilities. GIS has been appraised as valuable technology for facilities management since the 1990s. Therefore, the long and mid term GIS construction plan has been established by the national government and the local government. Some facilities management systems have been built and developed for suppling user-friendly functions. From this point of view, the information system based on the video sequences is considered a more effective way to improve the defects of conventional GIS using the digital map or the image as the base map. Using the video sequences as a base map, the availability of the system ill be increased because the real world information can be furnished to the users. In this study, through the connection between the GIS data, the digital map and the attribute data, and the video sequences taken from the airship using the video geo-referencing and the object tracking, we developed the facilities management system as a prototype which can effectively manage the road utilities. We also presented potentialities of the suggested system for facility management based on the video sequences.

  • PDF

CPU Parallel Processing and GPU-accelerated Processing of UHD Video Sequence using HEVC (HEVC를 이용한 UHD 영상의 CPU 병렬처리 및 GPU가속처리)

  • Hong, Sung-Wook;Lee, Yung-Lyul
    • Journal of Broadcast Engineering
    • /
    • v.18 no.6
    • /
    • pp.816-822
    • /
    • 2013
  • The latest video coding standard HEVC was developed by the joint work of JCT-VC(Joint Collaborative Team on Video Coding) from ITU-T VCEG and ISO/IEC MPEG. The HEVC standard reduces the BD-Bitrate of about 50% compared with the H.264/AVC standard. However, using the various methods for obtaining the coding gains has increased complexity problems. The proposed method reduces the complexity of HEVC by using both CPU parallel processing and GPU-accelerated processing. The experiment result for UHD($3840{\times}2144$) video sequences achieves 15fps encoding/decoding performance by applying the proposed method. Sooner or later, we expect that the H/W speedup of data transfer rates between CPU and GPU will result in reducing the encoding/decoding times much more.

Digital Video Watermarking Using Block Reordering Algorithm Based on DCT (DCT 기반의 블록 재정렬 알고리즘을 이용한 디지털 비디오 워터마킹)

  • Kim Kyoung-Sik;Park Kyung-Jun;Ko Hyung Hwa
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.30 no.7C
    • /
    • pp.696-705
    • /
    • 2005
  • The rapid progress of the software has enabled individuals to copy and remake digital contents, which was only done by professionals. However, this rapid growth of software has brought many other defects. As a solution to the problems, contents producer needs to have certification and inspection of its contents text and hold the proprietary rights. In this paper, the new video watermarking scheme is proposed which is appropriate for the requirements mentioned for digital video contents and it is applied to MPEG-2 video encoder and decoder. In order to insert watermark sequence for digital video data, watermarking used blocks reordering algorithm and SCL(secret code list). It is satisfied two properties of cryptography that confidentiality and integrity. We test the proposed algorithm to see its performance in terms of watermark capacity, compression robustness and visual quality impact. According to the experiment of proposed watermarking algorithm, we obtained satisfactory results that watermark can still be extracted after MPEG-2 re-encoding at lower bit rates. It is important property for data hiding and user authentication in video data.

A Descriptor Design for the Video Retrieval Combining the Global Feature of an Image and the Local of a Moving Object (영상의 전역 특징과 이동객체의 지역 특징을 융합한 동영상 검색 디스크립터 설계)

  • Jung, Byung-Man;Lee, Kyu-Won
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.18 no.1
    • /
    • pp.142-148
    • /
    • 2014
  • A descriptor which is suitable for motion analysis by using the motion features of moving objects from the real time image sequence is proposed. To segment moving objects from the background, the background learning is performed. We extract motion trajectories of individual objects by using the sequence of the 1st order moment of moving objects. The center points of each object are managed by linked list. The descriptor includes the 1st order coordinates of moving object belong to neighbor of the pre-defined position in grid pattern, The start frame number which a moving object appeared in the scene and the end frame number which it disappeared. A video retrieval by the proposed descriptor combining global and local feature is more effective than conventional methods which adopt a single feature among global and local features.

A Framework for Object Detection by Haze Removal (안개 제거에 의한 객체 검출 성능 향상 방법)

  • Kim, Sang-Kyoon;Choi, Kyoung-Ho;Park, Soon-Young
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.51 no.5
    • /
    • pp.168-176
    • /
    • 2014
  • Detecting moving objects from a video sequence is a fundamental and critical task in video surveillance, traffic monitoring and analysis, and human detection and tracking. It is very difficult to detect moving objects in a video sequence degraded by the environmental factor such as fog. In particular, the color of an object become similar to the neighbor and it reduces the saturation, thus making it very difficult to distinguish the object from the background. For such a reason, it is shown that the performance and reliability of object detection and tracking are poor in the foggy weather. In this paper, we propose a novel method to improve the performance of object detection, combining a haze removal algorithm and a local histogram-based object tracking method. For the quantitative evaluation of the proposed system, information retrieval measurements, recall and precision, are used to quantify how well the performance is improved before and after the haze removal. As a result, the visibility of the image is enhanced and the performance of objects detection is improved.

Measurement of Flickering Artifact for H.264 with Periodic I-Frame Structure (주기적 I-프레임 구조의 H.264 부호화 동영상을 위한 플리커링 측정 알고리즘)

  • Lim, Jong-Min;Kang, Dong-Wook;Jung, Kyeong-Hoon
    • Journal of Broadcast Engineering
    • /
    • v.15 no.3
    • /
    • pp.321-331
    • /
    • 2010
  • Most of multimedia video coding algorithms are lossy schemes and several kinds of spatial and temporal artifacts are inevitable. Flickering, which is the most typical coding artifact in time domain, is mainly due to fact that the quality of coded sequence fluctuates as the quantization parameter is adjusted for rate control. In this paper, we analyzed the effect of quality variation according to the characteristics of video sequence when the I-frames are periodically inserted. And we proposed the FR(Full Reference)-based assessment algorithm to measure the amount of flickering artifacts in the coded video. It is discovered that the flickering becomes critical when the level of quality is intermediate and is affected by the amount of detail or movement, the size of object, and camera parameters. The proposed measurement algorithm shows is well consistent with HVS(Human Visual System).

Depth Estimation and Intermediate View Synthesis for Three-dimensional Video Generation (3차원 영상 생성을 위한 깊이맵 추정 및 중간시점 영상합성 방법)

  • Lee, Sang-Beom;Lee, Cheon;Ho, Yo-Sung
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.10B
    • /
    • pp.1070-1075
    • /
    • 2009
  • In this paper, we propose new depth estimation and intermediate view synthesis algorithms for three-dimensional video generation. In order to improve temporal consistency of the depth map sequence, we add a temporal weighting function to the conventional matching function when we compute the matching cost for estimating the depth information. In addition, we propose a boundary noise removal method in the view synthesis operation. after finding boundary noise areas using the depth map, we replace them with corresponding texture information from the other reference image. Experimental results showed that the proposed algorithm improved temporal consistency of the depth sequence and reduced flickering artifacts in the virtual view. It also improved visual quality of the synthesized virtual views by removing the boundary noise.