• Title/Summary/Keyword: Video Image Data

Search Result 897, Processing Time 0.065 seconds

PCA-Based MPEG Video Retrieval in Compressed Domain (PCA에 기반한 압축영역에서의 MPEG Video 검색기법)

  • 이경화;강대성
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.40 no.1
    • /
    • pp.28-33
    • /
    • 2003
  • This paper proposes a database index and retrieval method using the PCA(Principal Component Analysis). We perform a scene change detection and key frame extraction from the DC Image constructed by DCT DC coefficients in the compressed video stream that is video compression standard such as MPEG. In the extracted key frame, we use the PCA, then we can make codebook that has a statistical data as a codeword, which is saved as a database index. We also provide retrieval image that are similar to user's query image in a video database. As a result of experiments, we confirmed that the proposed method clearly showed superior performance in video retrieval and reduced computation time and memory space.

Soft-$\alpha$ Filter Technology for image enhancement of MPEG-2 Video (MPEG-2 비디오의 화질 향상을 위한 소프트-$\alpha$ 필터 기법)

  • 심비연;박영배
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2002.04b
    • /
    • pp.109-111
    • /
    • 2002
  • Visual organs play an important role in human information recognition processes. If they are expressed in a way of digital information, it makes much bigger amount of visual information among any other information. For that reason, MPEG-2 has been taken use of to represent information compressing technology in multi-media. Although the imported data would basically contain noises, when original video images are encoded into MPET-2. Accordingly, we propose soft- $\alpha$ filter to improve image quality of digital image received from the actual image and to reduce noises from them. We also propose a method combining vertical/horizontal filter and soft- $\alpha$ filter on MPEG-2 video image. We can get two kinds of effects from the advantages of this kind of combination. Firstly, it will reduce processing time ducting horizontal and vetical filtering process. It will cover time for soft- $\alpha$ filter. Secondly, it will simplify the colors in horizontal and vertical filter. Therefore we can get clearer quality without noises from soft- $\alpha$ filter.

  • PDF

2-D MMFF Model and Performance Analysis of 2-layer coded Video Traffic Sources (2-차원 MMFF 모델을 이용한 2-계층 부호화 영상 트래픽의 모델링 및 성능 분석)

  • 안희준;노병희;김재균
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.21 no.1
    • /
    • pp.17-32
    • /
    • 1996
  • In this paper, a model for two-layered video traffic is proposed. The performance analysis of the proposed model and the effects of two-layer coding scehemes in ATM networks are also studied. ATM-based networks give the possibility to support image codingat variable bit rate(VBR). Two layer coding is one of the very promising methods among many proposed methods to compensate the cell loss, the major drawback in ATM networks. From the experimental data of the 2-layer coded video traffics, it is observed that traffic patterns of base layer and enhanced layer are highly correlate to each other, when constant image quality is kept. With this observation, coded two layered video traffic can be modeled as 2-dimensional Markov chain. The model well fit the real experimental data. The model was used for the analysis of the performance of statistical multiplexer with priorites in ATM networks.

  • PDF

Distributed Video Compressive Sensing Reconstruction by Adaptive PCA Sparse Basis and Nonlocal Similarity

  • Wu, Minghu;Zhu, Xiuchang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.8
    • /
    • pp.2851-2865
    • /
    • 2014
  • To improve the rate-distortion performance of distributed video compressive sensing (DVCS), the adaptive sparse basis and nonlocal similarity of video are proposed to jointly reconstruct the video signal in this paper. Due to the lack of motion information between frames and the appearance of some noises in the reference frames, the sparse dictionary, which is constructed using the examples directly extracted from the reference frames, has already not better obtained the sparse representation of the interpolated block. This paper proposes a method to construct the sparse dictionary. Firstly, the example-based data matrix is constructed by using the motion information between frames, and then the principle components analysis (PCA) is used to compute some significant principle components of data matrix. Finally, the sparse dictionary is constructed by these significant principle components. The merit of the proposed sparse dictionary is that it can not only adaptively change in terms of the spatial-temporal characteristics, but also has ability to suppress noises. Besides, considering that the sparse priors cannot preserve the edges and textures of video frames well, the nonlocal similarity regularization term has also been introduced into reconstruction model. Experimental results show that the proposed algorithm can improve the objective and subjective quality of video frame, and achieve the better rate-distortion performance of DVCS system at the cost of a certain computational complexity.

A STUDY ON PUPIL DETECTION AND TRACKING METHODS BASED ON IMAGE DATA ANALYSIS

  • CHOI, HANA;GIM, MINJUNG;YOON, SANGWON
    • Journal of the Korean Society for Industrial and Applied Mathematics
    • /
    • v.25 no.4
    • /
    • pp.327-336
    • /
    • 2021
  • In this paper, we will introduce the image processing methods for the remote pupillary light reflex measurement using the video taken by a general smartphone camera without a special device such as an infrared camera. We propose an algorithm for estimate the size of the pupil that changes with light using image data analysis without a learning process. In addition, we will introduce the results of visualizing the change in the pupil size by removing noise from the recorded data of the pupil size measured for each frame of the video. We expect that this study will contribute to the construction of an objective indicator for remote pupillary light reflex measurement in the situation where non-face-to-face communication has become common due to COVID-19 and the demand for remote diagnosis is increasing.

An Analysis of Big Video Data with Cloud Computing in Ubiquitous City (클라우드 컴퓨팅을 이용한 유시티 비디오 빅데이터 분석)

  • Lee, Hak Geon;Yun, Chang Ho;Park, Jong Won;Lee, Yong Woo
    • Journal of Internet Computing and Services
    • /
    • v.15 no.3
    • /
    • pp.45-52
    • /
    • 2014
  • The Ubiquitous-City (U-City) is a smart or intelligent city to satisfy human beings' desire to enjoy IT services with any device, anytime, anywhere. It is a future city model based on Internet of everything or things (IoE or IoT). It includes a lot of video cameras which are networked together. The networked video cameras support a lot of U-City services as one of the main input data together with sensors. They generate huge amount of video information, real big data for the U-City all the time. It is usually required that the U-City manipulates the big data in real-time. And it is not easy at all. Also, many times, it is required that the accumulated video data are analyzed to detect an event or find a figure among them. It requires a lot of computational power and usually takes a lot of time. Currently we can find researches which try to reduce the processing time of the big video data. Cloud computing can be a good solution to address this matter. There are many cloud computing methodologies which can be used to address the matter. MapReduce is an interesting and attractive methodology for it. It has many advantages and is getting popularity in many areas. Video cameras evolve day by day so that the resolution improves sharply. It leads to the exponential growth of the produced data by the networked video cameras. We are coping with real big data when we have to deal with video image data which are produced by the good quality video cameras. A video surveillance system was not useful until we find the cloud computing. But it is now being widely spread in U-Cities since we find some useful methodologies. Video data are unstructured data thus it is not easy to find a good research result of analyzing the data with MapReduce. This paper presents an analyzing system for the video surveillance system, which is a cloud-computing based video data management system. It is easy to deploy, flexible and reliable. It consists of the video manager, the video monitors, the storage for the video images, the storage client and streaming IN component. The "video monitor" for the video images consists of "video translater" and "protocol manager". The "storage" contains MapReduce analyzer. All components were designed according to the functional requirement of video surveillance system. The "streaming IN" component receives the video data from the networked video cameras and delivers them to the "storage client". It also manages the bottleneck of the network to smooth the data stream. The "storage client" receives the video data from the "streaming IN" component and stores them to the storage. It also helps other components to access the storage. The "video monitor" component transfers the video data by smoothly streaming and manages the protocol. The "video translator" sub-component enables users to manage the resolution, the codec and the frame rate of the video image. The "protocol" sub-component manages the Real Time Streaming Protocol (RTSP) and Real Time Messaging Protocol (RTMP). We use Hadoop Distributed File System(HDFS) for the storage of cloud computing. Hadoop stores the data in HDFS and provides the platform that can process data with simple MapReduce programming model. We suggest our own methodology to analyze the video images using MapReduce in this paper. That is, the workflow of video analysis is presented and detailed explanation is given in this paper. The performance evaluation was experiment and we found that our proposed system worked well. The performance evaluation results are presented in this paper with analysis. With our cluster system, we used compressed $1920{\times}1080(FHD)$ resolution video data, H.264 codec and HDFS as video storage. We measured the processing time according to the number of frame per mapper. Tracing the optimal splitting size of input data and the processing time according to the number of node, we found the linearity of the system performance.

A Heuristic Approach for Simulation of time-course Visual Adaptation for High Dynamic Image Streams

  • Kelvin, Bwalya;Yang, Seung-Ji;Choi, Jong-Soo;Park, Soo-Jun;Ro, Yong-Man
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2007.05a
    • /
    • pp.285-288
    • /
    • 2007
  • There is substantial evidence from earlier researches that older adults have difficult seeing under low illumination and at night, even in the absence of ocular diseases. During human aging, there is a rampant decrease in rod/cone-meditated adaptation which is caused by delayed rhodopsin regeneration and pigment depletion. This calls for a need to develop appropriate visual gadgets to effectively aid the aging generation. Our research culminates its approach from Pattanaik's model by making extensions to temporal visual filtering, thereby simulating a reduction of visual response which comes with age. Our filtering model paves way and lays a foundation for future research to develop a more effective adaptation model that may be further used in developing visual content adaptation aids and guidelines in MPEG 21 environment. We demonstrate our visual model using a High Dynamic Range image and the experiment results are in conversant with the psychophysical data from previous vision researches.

  • PDF

A Hadoop-based Multimedia Transcoding System for Processing Social Media in the PaaS Platform of SMCCSE

  • Kim, Myoungjin;Han, Seungho;Cui, Yun;Lee, Hanku;Jeong, Changsung
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.6 no.11
    • /
    • pp.2827-2848
    • /
    • 2012
  • Previously, we described a social media cloud computing service environment (SMCCSE). This SMCCSE supports the development of social networking services (SNSs) that include audio, image, and video formats. A social media cloud computing PaaS platform, a core component in a SMCCSE, processes large amounts of social media in a parallel and distributed manner for supporting a reliable SNS. Here, we propose a Hadoop-based multimedia system for image and video transcoding processing, necessary functions of our PaaS platform. Our system consists of two modules, including an image transcoding module and a video transcoding module. We also design and implement the system by using a MapReduce framework running on a Hadoop Distributed File System (HDFS) and the media processing libraries Xuggler and JAI. In this way, our system exponentially reduces the encoding time for transcoding large amounts of image and video files into specific formats depending on user-requested options (such as resolution, bit rate, and frame rate). In order to evaluate system performance, we measure the total image and video transcoding time for image and video data sets, respectively, under various experimental conditions. In addition, we compare the video transcoding performance of our cloud-based approach with that of the traditional frame-level parallel processing-based approach. Based on experiments performed on a 28-node cluster, the proposed Hadoop-based multimedia transcoding system delivers excellent speed and quality.

A Study on Generation of Free Stereo Mosaic Image Using Video Sequences (비디오 프레임 영상을 이용한 자유 입체 모자이크 영상 제작에 관한 연구)

  • Noh, Myoung-Jong;Cho, Woo-Sug;Park, June-Ku
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.27 no.4
    • /
    • pp.453-460
    • /
    • 2009
  • For constructing 3D information using aerial photograph or video sequences, left and right stereo images having different viewing angle should be prepared in overlapping area. In video sequences, left and right stereo images would be generated by mosaicing left and right slice images extracted in consecutive video sequences. Therefore, this paper is focused on generating left and right stereo mosaic images that are able to construct 3D information and video sequences could be made for the best use. In the stereo mosaic generation, motion parameters between video sequences should be firstly determined. In this paper, to determine motion parameters, free mosaic method using geometric relationship, such as relative orientation parameters, between consecutive frame images without GPS/INS geo-data have applied. After determining the motion parameters, the mosaic image have generated by 4 step processes: image registration, image slicing, determining on stitching line, and 3D image mosaicking. As the result of experiment, generated stereo mosaic image and analyzed result of x, y-parallax have showed.

Temporal Anti-aliasing of a Stereoscopic 3D Video

  • Kim, Wook-Joong;Kim, Seong-Dae;Hur, Nam-Ho;Kim, Jin-Woong
    • ETRI Journal
    • /
    • v.31 no.1
    • /
    • pp.1-9
    • /
    • 2009
  • Frequency domain analysis is a fundamental procedure for understanding the characteristics of visual data. Several studies have been conducted with 2D videos, but analysis of stereoscopic 3D videos is rarely carried out. In this paper, we derive the Fourier transform of a simplified 3D video signal and analyze how a 3D video is influenced by disparity and motion in terms of temporal aliasing. It is already known that object motion affects temporal frequency characteristics of a time-varying image sequence. In our analysis, we show that a 3D video is influenced not only by motion but also by disparity. Based on this conclusion, we present a temporal anti-aliasing filter for a 3D video. Since the human process of depth perception mainly determines the quality of a reproduced 3D image, 2D image processing techniques are not directly applicable to 3D images. The analysis presented in this paper will be useful for reducing undesirable visual artifacts in 3D video as well as for assisting the development of relevant technologies.

  • PDF