• Title/Summary/Keyword: video fusion

Search Result 118, Processing Time 0.023 seconds

Research of Vehicle Navigation Based Video-GIS

  • Feng, Jiang-Fan;Zhu, Guan-Yu;Liu, Zhao-Hong;Li, Yan
    • Journal of Korea Spatial Information System Society
    • /
    • v.11 no.2
    • /
    • pp.39-44
    • /
    • 2009
  • In order to make the effect of the navigation system more direct, the paper proposes a thought of vehicle navigation system based on Video-GIS. A semantic framework has been defined whose core is focused on the integration and interaction of video and spatial information, which supports full content retrieval based on multimodal metadata extraction and fusion, and supports kinds of wireless access mode. Furthermore, requirements of prototype system are discussed. Then the design and implementation of framework are discussed. Next, describe the key ideas and technologies involved. Finally, we point out its future research trend.

  • PDF

Video Palmprint Recognition System Based on Modified Double-line-single-point Assisted Placement

  • Wu, Tengfei;Leng, Lu
    • Journal of Multimedia Information System
    • /
    • v.8 no.1
    • /
    • pp.23-30
    • /
    • 2021
  • Palmprint has become a popular biometric modality; however, palmprint recognition has not been conducted in video media. Video palmprint recognition (VPR) has some advantages that are absent in image palmprint recognition. In VPR, the registration and recognition can be automatically implemented without users' manual manipulation. A good-quality image can be selected from the video frames or generated from the fusion of multiple video frames. VPR in contactless mode overcomes several problems caused by contact mode; however, contactless mode, especially mobile mode, encounters with several revere challenges. Double-line-single-point (DLSP) assisted placement technique can overcome the challenges as well as effectively reduce the localization error and computation complexity. This paper modifies DLSP technique to reduce the invalid area in the frames. In addition, the valid frames, in which users place their hands correctly, are selected according to finger gap judgement, and then some key frames, which have good quality, are selected from the valid frames as the gallery samples that are matched with the query samples for authentication decision. The VPR algorithm is conducted on the system designed and developed on mobile device.

Action Recognition Method in Sports Video Shear Based on Fish Swarm Algorithm

  • Jie Sun;Lin Lu
    • Journal of Information Processing Systems
    • /
    • v.19 no.4
    • /
    • pp.554-562
    • /
    • 2023
  • This research offers a sports video action recognition approach based on the fish swarm algorithm in light of the low accuracy of existing sports video action recognition methods. A modified fish swarm algorithm is proposed to construct invariant features and decrease the dimension of features. Based on this algorithm, local features and global features can be classified. The experimental findings on the typical sports action data set demonstrate that the key details of sports action can be successfully retained by the dimensionality-reduced fusion invariant characteristics. According to this research, the average recognition time of the proposed method for walking, running, squatting, sitting, and bending is less than 326 seconds, and the average recognition rate is higher than 94%. This proves that this method can significantly improve the performance and efficiency of online sports video motion recognition.

Two-Stream Convolutional Neural Network for Video Action Recognition

  • Qiao, Han;Liu, Shuang;Xu, Qingzhen;Liu, Shouqiang;Yang, Wanggan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.10
    • /
    • pp.3668-3684
    • /
    • 2021
  • Video action recognition is widely used in video surveillance, behavior detection, human-computer interaction, medically assisted diagnosis and motion analysis. However, video action recognition can be disturbed by many factors, such as background, illumination and so on. Two-stream convolutional neural network uses the video spatial and temporal models to train separately, and performs fusion at the output end. The multi segment Two-Stream convolutional neural network model trains temporal and spatial information from the video to extract their feature and fuse them, then determine the category of video action. Google Xception model and the transfer learning is adopted in this paper, and the Xception model which trained on ImageNet is used as the initial weight. It greatly overcomes the problem of model underfitting caused by insufficient video behavior dataset, and it can effectively reduce the influence of various factors in the video. This way also greatly improves the accuracy and reduces the training time. What's more, to make up for the shortage of dataset, the kinetics400 dataset was used for pre-training, which greatly improved the accuracy of the model. In this applied research, through continuous efforts, the expected goal is basically achieved, and according to the study and research, the design of the original dual-flow model is improved.

A Virtual Environment for Optimal use of Video Analytic of IP Cameras and Feasibility Study (IP 카메라의 VIDEO ANALYTIC 최적 활용을 위한 가상환경 구축 및 유용성 분석 연구)

  • Ryu, Hong-Nam;Kim, Jong-Hun;Yoo, Gyeong-Mo;Hong, Ju-Yeong;Choi, Byoung-Wook
    • Journal of the Korean Institute of Illuminating and Electrical Installation Engineers
    • /
    • v.29 no.11
    • /
    • pp.96-101
    • /
    • 2015
  • In recent years, researches regarding optimal placement of CCTV(Closed-circuit Television) cameras via architecture modeling has been conducted. However, for analyzing surveillance coverage through actual human movement, the application of VA(Video Analytics) function of IP(Internet Protocol) cameras has not been studied. This paper compares two methods using data captured from real-world cameras and data acquired from a virtual environment. In using real cameras, we develop GUI(Graphical User Interface) to be used as a logfile which is stored hourly and daily through VA functions and to be used commercially for placement of products inside a shop. The virtual environment was constructed to emulate an real world such as the building structure and the camera with its specifications. Moreover, suitable placement of the camera is done by recognizing obstacles and the number of people counted within the camera's range of view. This research aims to solve time and economic constraints of actual installation of surveillance cameras in real-world environment and to do feasibility study of virtual environment.

Real-Time Visible-Infrared Image Fusion using Multi-Guided Filter

  • Jeong, Woojin;Han, Bok Gyu;Yang, Hyeon Seok;Moon, Young Shik
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.6
    • /
    • pp.3092-3107
    • /
    • 2019
  • Visible-infrared image fusion is a process of synthesizing an infrared image and a visible image into a fused image. This process synthesizes the complementary advantages of both images. The infrared image is able to capture a target object in dark or foggy environments. However, the utility of the infrared image is hindered by the blurry appearance of objects. On the other hand, the visible image clearly shows an object under normal lighting conditions, but it is not ideal in dark or foggy environments. In this paper, we propose a multi-guided filter and a real-time image fusion method. The proposed multi-guided filter is a modification of the guided filter for multiple guidance images. Using this filter, we propose a real-time image fusion method. The speed of the proposed fusion method is much faster than that of conventional image fusion methods. In an experiment, we compare the proposed method and the conventional methods in terms of quantity, quality, fusing speed, and flickering artifacts. The proposed method synthesizes 57.93 frames per second for an image size of $320{\times}270$. Based on our experiments, we confirmed that the proposed method is able to perform real-time processing. In addition, the proposed method synthesizes flicker-free video.

A Study on in the Context of Audiovisual Art (<백-아베 비디오 신디사이저>의 오디오 비주얼아트적 고찰)

  • Yoon, Ji Won
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.4
    • /
    • pp.615-624
    • /
    • 2020
  • By enabling musicians to freely control the elements involved in sound production and tone generation with a variety of timbre, synthesizers have revolutionized and permanently changed music since the 1960s. Paik-Abe Video Synthesizer, a masterpiece of video art maestro Nam June Paik, is a prominent example of re-interpretation of this new musical instrument in the realm of video and audio. This article examines Paik-Abe Video Synthesizer as an innovative instrument to play videos from the perspective of audiovisual art, and establishes its aesthetic value and significance through both artistic and technical analysis. The instrument, which embodied the concept of image sampling and real-time interactive video as an image-based multi-channel music production tool, contributed to establishing a new relationship between sound and image within the realm of audiovisual art. The fact that his video synthesizer not only adds image to sound, but also presents a complete fusion of image and sound as an image instrument with musical characteristics, becomes highly meaningful in this age of synesthesia.

Motion Estimation Using Feature Matching and Strongly Coupled Recurrent Module Fusion (특징정합과 순환적 모듈융합에 의한 움직임 추정)

  • 심동규;박래홍
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.31B no.12
    • /
    • pp.59-71
    • /
    • 1994
  • This paper proposes a motion estimation method in video sequences based on the feature based matching and anistropic propagation. It measures translation and rotation parameters using a relaxation scheme at feature points and object orinted anistropic propagation in continuous and discontinuous regions. Also an iterative improvement motion extimation based on the strongly coupled module fusion and adaptive smoothing is proposed. Computer simulation results show the effectiveness of the proposed algorithm.

  • PDF

Clinical Analysis of Video-assisted Thoracoscopic Spinal Surgery in the Thoracic or Thoracolumbar Spinal Pathologies

  • Kim, Sung-Jin;Sohn, Moon-Jun;Ryoo, Ji-Yoon;Kim, Yeon-Soo;Whang, Choong-Jin
    • Journal of Korean Neurosurgical Society
    • /
    • v.42 no.4
    • /
    • pp.293-299
    • /
    • 2007
  • Objective : Thoracoscopic spinal surgery provides minimally invasive approaches for effective vertebral decompression and reconstruction of the thoracic and thoracolumbar spine, while surgery related morbidity can be significantly lowered. This study analyzes clinical results of thoracoscopic spinal surgery performed at our institute. Methods : Twenty consecutive patients underwent video-assisted thoracosopic surgery (VATS) to treat various thoracic and thoracolumbar pathologies from April 2000 to July 2006. The lesions consisted of spinal trauma (13 cases), thoracic disc herniation (4 cases), tuberculous spondylitis (1 case), post-operative thoracolumbar kyphosis (1 case) and thoracic tumor (1 case). The level of operation included upper thoracic lesions (3 cases), midthoracic lesions (6 cases) and thoracolumbar lesions (11 cases). We classified the procedure into three groups: stand-alone thoracoscopic discectomy (3 cases), thoracoscopic fusion (11 cases) and video assisted mini-thoracotomy (6 cases). Results : Analysis on the Frankel performance scale in spinal trauma patients (13 cases), showed a total of 7 patients who had neurological impairment preoperatively : Grade D (2 cases), Grade C (2 cases), Grade B (1 case), and Grade A (2 cases). Four patients were neurologically improved postoperatively, two patients were improved from C to E, one improved from grade D to E and one improved from grade B to grade D. The preoperative Cobb's and kyphotic angle were measured in spinal trauma patients and were $18.9{\pm}4.4^{\circ}$ and $18.8{\pm}4.6^{\circ}$, respectively. Postoperatively, the angles showed statistically significant improvement, $15.1{\pm}3.7^{\circ}$ and $11.3{\pm}2.4^{\circ}$, respectively(P<0.001). Conclusion : Although VATS requires a steep learning curve, it is an effective and minimally invasive procedure which provides biomechanical stability in terms of anterior column decompression and reconstruction for anterior load bearing, and preservation of intercostal muscles and diaphragm.

X3D Based Web Visualization by Data Fusion of 3D Spatial Information and Video Sequence (3D 공간정보와 비디오 융합에 의한 X3D기반 웹 가시화)

  • Sohn, Hong-Gyoo;Kim, Seong-Sam;Yoo, Byoung-Hyun;Kim, Sang-Min
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.17 no.4
    • /
    • pp.95-103
    • /
    • 2009
  • Global interests for construction of 3 dimensional spatial information has risen due to development of measurement sensors and data processing technologies. In spite of criticism for the violation of personal privacy, CCTV cameras equipped in outdoor public space of urban area are used as a fundamental sensor for traffic management, crime prevention or hazard monitoring. For safety guarantee in urban environment and disaster prevention, a surveillance system integrating pre-constructed 3 dimensional spatial information with CCTV data or video sequence is needed for monitoring and observing emergent situation interactively in real time. In this study, we proposed applicability of the prototype system for web visualization based on X3D, an international standard of real time web visualization, by integrating 3 dimensional spatial information with video sequence.

  • PDF