• 제목/요약/키워드: video transfer

검색결과 269건 처리시간 0.024초

딥러닝 스타일 전이 기반의 무대 탐방 콘텐츠 생성 기법 (Generation of Stage Tour Contents with Deep Learning Style Transfer)

  • 김동민;김현식;봉대현;최종윤;정진우
    • 한국정보통신학회논문지
    • /
    • 제24권11호
    • /
    • pp.1403-1410
    • /
    • 2020
  • 최근, 비대면 경험 및 서비스에 관한 관심이 증가하면서 스마트폰이나 태블릿과 같은 모바일 기기를 이용하여 손쉽게 이용할 수 있는 웹 동영상 콘텐츠에 대한 수요가 급격히 증가하고 있다. 이와 같은 요구사항에 대응하기 위하여, 본 논문에서는 애니메이션이나 영화에 등장하는 명소를 방문하는 무대 탐방 경험을 제공할 수 있는 영상 콘텐츠를 보다 효율적으로 제작하기 위한 기법을 제안한다. 이를 위하여, Google Maps와 Google Street View API를 이용하여 무대탐방 지역에 해당하는 이미지를 수집하여 이미지 데이터셋을 구축하였다. 그 후, 딥러닝 기반의 style transfer 기술을 접목시켜 애니메이션의 독특한 화풍을 실사 이미지에 적용한 후 동영상화하기 위한 방법을 제시하였다. 마지막으로, 다양한 실험을 통해 제안하는 기법을 이용하여 보다 재미있고 흥미로운 형태의 무대탐방 영상 콘텐츠를 생성할 수 있음을 보였다.

DTV 화질향상을 위한 자막데이터 전송방법 (Caption Data Transmission Method for HDTV Picture Quality Improvement)

  • 한찬호
    • 한국멀티미디어학회논문지
    • /
    • 제20권10호
    • /
    • pp.1628-1636
    • /
    • 2017
  • Such as closed caption, ancillary data, electronic program guide(EPG), data broadcasting, and etc, increased data for service convenience cause to degrade video quality of high definition contents. This article propose a method to transfer the closed caption data of video contents without video quality degradation. Video quality degradation does not cause in video compression by the block image insertion of caption data in DTV essential hidden area. Additionally the proposed methods have advantage to synchronize video, audio, and caption from preinserted script without time delay.

Vision Transformer를 활용한 비디오 분류 성능 향상을 위한 Fine-tuning 신경망 (Fine-tuning Neural Network for Improving Video Classification Performance Using Vision Transformer)

  • 이광엽;이지원;박태룡
    • 전기전자학회논문지
    • /
    • 제27권3호
    • /
    • pp.313-318
    • /
    • 2023
  • 본 논문은 Vision Transformer를 기반으로 하는 Video Classification의 성능을 개선하는 방법으로 fine-tuning를 적용한 신경망을 제안한다. 최근 딥러닝 기반 실시간 비디오 영상 분석의 필요성이 대두되고 있다. Image Classification에 사용되는 기존 CNN 모델의 특징상 연속된 Frame에 대한 연관성을 분석하기 어렵다는 단점이 있다. 이와 같은 문제를 Attention 메커니즘이 적용된 Vistion Transformer와 Non-local 신경망 모델을 비교 분석하여 최적의 모델을 찾아 해결하고자 한다. 또한, 전이 학습 방법으로 fine-tuning의 다양한 방법을 적용하여 최적의 fine-tuning 신경망 모델을 제안한다. 실험은 UCF101 데이터셋으로 모델을 학습시킨 후, UTA-RLDD 데이터셋에 전이 학습 방법을 적용하여 모델의 성능을 검증하였다.

Video Phone Tube用 형광박의 제조에 관한 연구 (A Study on Manufacture of Phosphor Screen for Video Phone Tube)

  • 우진호
    • 한국인쇄학회지
    • /
    • 제22권2호
    • /
    • pp.123-138
    • /
    • 2004
  • The video phone tube (VPT) of monochrome CRT have utilized home door phone, fish-finder and the rear watch monitors. Phosphor screen formation is made by electrodeposition spin coating and thermal transfer methods etc. Recently, thermal transfer method was developed, as a novel method, to form the phosphor surface for mnonchrom VPT. This method have advantages of simple process, automatization, clean environment, saving raw material and saving running-cost. In this study, it was developed new phosphor of VPT, and tested about phosphor paste properties. An experimental studies of VPT as a new phosphor property and improved VPT's manufacturing process shortening and brightness. As thermal transper method is a paste processing, it is important that rheology of phosphor effects on the formation of phosphor screen. Hence this paper was studied rheology properties of phosphor paste and the formation of phosphor screen had looked most suitable condition. Experimented thermal separation properties of low calcination temperature resin and the result analyzed comparison by TGA. Also, examined calcination properties to reduce remaining binder phosphor.

  • PDF

초광각 무선 디지털 전자 내시경 모듈에 관한 연구 (A Study on a Ultra-wide-angle Wireless Digital Electronic Endoscope Modules)

  • 심동하;김형오;이봉주;홍승철;이재선;차재상
    • 전기학회논문지
    • /
    • 제63권4호
    • /
    • pp.570-574
    • /
    • 2014
  • This paper proposes a wireless digital endoscope with a ultra-wide-angle view. Two key components are implemented to demonstrate the feasibility of the proposed endoscope. First, a ultra-wide-angle lens module with the field of view of 144 degree and F-number of 2.2 is designed and manufactured. Second, a wireless module for a high-speed video transfer is implemented using a USB device server and wireless LAN router. The wireless module can directly transfer a streaming video to a computer with the resolution of 1920x1080, frame rate of 30 fps, and data rate of 53.3 Mbps without an internet connection. Since the wireless module supports two USB devices, two spots can be simultaneously observed using the proposed endoscope.

Adaptive Temporal Rate Control of Video Objects for Scalable Transmission

  • Chang, Hee-Dong;Lim, Young-Kwon;Lee, Myoung-Ho;Ahan, Chieteuk
    • 한국방송∙미디어공학회:학술대회논문집
    • /
    • 한국방송공학회 1997년도 Proceedings International Workshop on New Video Media Technology
    • /
    • pp.43-48
    • /
    • 1997
  • The video transmission for real-time viewing over the Internet is a core operation for the multimedia services. However, its realization is very difficult because the Internet has two major problems, namely, very narrow endpoint-bandwidth and the network jitter. We already proposed a scalable video transmission method in [8] which used MPEG-4 video VM(Verification Model) 2.0[3] for very low bit rate coding and an adaptive temporal rate control of video objects to overcome the network jitter problem. In this paper, we present the improved adaptive temporal rate control scheme for the scalable transmission. Experimental results for three test video sequences show that the adaptive temporal rate control can transfer the video bitstream at source frame rate under variable network condition.

  • PDF

A Novel Video Stitching Method for Multi-Camera Surveillance Systems

  • Yin, Xiaoqing;Li, Weili;Wang, Bin;Liu, Yu;Zhang, Maojun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제8권10호
    • /
    • pp.3538-3556
    • /
    • 2014
  • This paper proposes a novel video stitching method that improves real-time performance and visual quality of a multi-camera video surveillance system. A two-stage seam searching algorithm based on enhanced dynamic programming is proposed. It can obtain satisfactory result and achieve better real-time performance than traditional seam-searching methods. The experiments show that the computing time is reduced by 66.4% using the proposed algorithm compared with enhanced dynamic programming, while the seam-searching accuracy is maintained. A real-time local update scheme reduces the deformation effect caused by moving objects passing through the seam, and a seam-based local color transfer model is constructed and applied to achieve smooth transition in the overlapped area, and overcome the traditional pixel blending methods. The effectiveness of the proposed method is proved in the experiements.

Violent crowd flow detection from surveillance cameras using deep transfer learning-gated recurrent unit

  • Elly Matul Imah;Riskyana Dewi Intan Puspitasari
    • ETRI Journal
    • /
    • 제46권4호
    • /
    • pp.671-682
    • /
    • 2024
  • Violence can be committed anywhere, even in crowded places. It is hence necessary to monitor human activities for public safety. Surveillance cameras can monitor surrounding activities but require human assistance to continuously monitor every incident. Automatic violence detection is needed for early warning and fast response. However, such automation is still challenging because of low video resolution and blind spots. This paper uses ResNet50v2 and the gated recurrent unit (GRU) algorithm to detect violence in the Movies, Hockey, and Crowd video datasets. Spatial features were extracted from each frame sequence of the video using a pretrained model from ResNet50V2, which was then classified using the optimal trained model on the GRU architecture. The experimental results were then compared with wavelet feature extraction methods and classification models, such as the convolutional neural network and long short-term memory. The results show that the proposed combination of ResNet50V2 and GRU is robust and delivers the best performance in terms of accuracy, recall, precision, and F1-score. The use of ResNet50V2 for feature extraction can improve model performance.

철도차량 시뮬레이션의 디지털 영상제어 시스템 연구 (A study on the Digital Video control system for train simulator)

  • 김봉택;최성
    • 한국철도학회:학술대회논문집
    • /
    • 한국철도학회 1999년도 추계학술대회 논문집
    • /
    • pp.259-266
    • /
    • 1999
  • A study on the static type train simulator will include the training of new drives requires that the environment of the cab, controls placement, etc. must highly realistic so that driver can readily transfer his training experience to the real world. The simulator computer sends video disc speed command to a Video PC processor. A video switcher select the output of the on-line player. This selection is done with loss of vertical synchronization, meaning the picture will not noticeable roll or jump as the simulation mover from disc to disc. The video image quality remain contestant through the simulated speed range from zero to 100km/h. Flicker is avoided in the scene by the use of a TBC(Time Base Corrector) which causes the display of one video field at a time. Thus, no interfield jitter is present when the scene is stopped.

  • PDF

Decision on Compression Ratios for Real-Time Transfer of Ultrasound Sequences

  • Lee, Jae-Hoon;Sung, Min-Mo;Kim, Hee-Joung;Yoo, Sun-Kwook;Kim, Eun-Kyung;Kim, Dong-Keun;Jung, Suk-Myung;Yoo, Hyung-Sik
    • 한국의학물리학회:학술대회논문집
    • /
    • 한국의학물리학회 2002년도 Proceedings
    • /
    • pp.489-491
    • /
    • 2002
  • The need for video diagnosis in medicine has been increased and real-time transfer of digital video will be an important component in PACS and telemedicine. But, Network environment has certain limitations that the required throughput can not satisfy quality of service (QoS). MPEG-4 ratified as a moving video standard by the ISO/IEC provides very efficient video coding covering the various ranges of low bit-rate in network environment. We implemented MPEG-4 CODEC (coder/decoder) and applied various compression ratios to moving ultrasound images. These images were displayed in random order on a client monitor passed through network. Radiologists determined subjective opinion scores for evaluating clinically acceptable image quality and then these were statistically processed in the t-Test method. Moreover the MPEG-4 decoded images were quantitatively analyzed by computing peak signal-to-noise ratio (PSNR) to objectively evaluate image quality. The bit-rate to maintain clinically acceptable image quality was up to 0.8Mbps. We successfully implemented the adaptive throughput or bit-rate relative to the image quality of ultrasound sequences used MPEG-4 that can be applied for diagnostic performance in real-time.

  • PDF