• Title/Summary/Keyword: Video/Audio Transfer

Search Result 35, Processing Time 0.02 seconds

An Advanced Coding for Video Streaming System: Hardware and Software Video Coding

  • Le, Tuan Thanh;Ryu, Eun-Seok
    • Journal of Internet Computing and Services
    • /
    • v.21 no.4
    • /
    • pp.51-57
    • /
    • 2020
  • Currently, High-efficient video coding (HEVC) has become the most promising video coding technology. However, the implementation of HEVC in video streaming systems is restricted by factors such as cost, design complexity, and compatibility with existing systems. While HEVC is considering deploying to various systems with different reached methods, H264/AVC can be one of the best choices for current video streaming systems. This paper presents an adaptive method for manipulating video streams using video coding on an integrated circuit (IC) designed with a private network processor. The proposed system allows to transfer multimedia data from cameras or other video sources to client. For this work, a series of video or audio packages from the video source are forwarded to the designed IC via HDMI cable, called Tx transmitter. The Tx processes input data into a real-time stream using its own protocol according to the Real-Time Transmission Protocol for both video and audio, then Tx transmits output packages to the video client though internet. The client includes hardware or software video/audio decoders to decode the received packages. Tx uses H264/AVC or HEVC video coding to encode video data, and its audio coding is PCM format. By handling the message exchanges between Tx and the client, the transmitted session can be set up quickly. Output results show that transmission's throughput can be achieved about 50 Mbps with approximately 80 msec latency.

A study on remote video transmit technique of mobile phone (모바일폰에서의 원격 영상 전송 기술에 관한 연구)

  • Jeong, Jong-Geun;Kim, Chul-Won
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.10 no.10
    • /
    • pp.1914-1919
    • /
    • 2006
  • Video transfer problem on mobile is transfer speed and controls. Compression technique is needed to transfer videos and H.263 codec is used for compression, effectively controls camera on remote places, increased the real time connecting users. In this paper, we could solve the problem that use existent RF, and could transfer the most suitable image and audio.

Caption Data Transmission Method for HDTV Picture Quality Improvement (DTV 화질향상을 위한 자막데이터 전송방법)

  • Han, Chan-Ho
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.10
    • /
    • pp.1628-1636
    • /
    • 2017
  • Such as closed caption, ancillary data, electronic program guide(EPG), data broadcasting, and etc, increased data for service convenience cause to degrade video quality of high definition contents. This article propose a method to transfer the closed caption data of video contents without video quality degradation. Video quality degradation does not cause in video compression by the block image insertion of caption data in DTV essential hidden area. Additionally the proposed methods have advantage to synchronize video, audio, and caption from preinserted script without time delay.

Robot Vision to Audio Description Based on Deep Learning for Effective Human-Robot Interaction (효과적인 인간-로봇 상호작용을 위한 딥러닝 기반 로봇 비전 자연어 설명문 생성 및 발화 기술)

  • Park, Dongkeon;Kang, Kyeong-Min;Bae, Jin-Woo;Han, Ji-Hyeong
    • The Journal of Korea Robotics Society
    • /
    • v.14 no.1
    • /
    • pp.22-30
    • /
    • 2019
  • For effective human-robot interaction, robots need to understand the current situation context well, but also the robots need to transfer its understanding to the human participant in efficient way. The most convenient way to deliver robot's understanding to the human participant is that the robot expresses its understanding using voice and natural language. Recently, the artificial intelligence for video understanding and natural language process has been developed very rapidly especially based on deep learning. Thus, this paper proposes robot vision to audio description method using deep learning. The applied deep learning model is a pipeline of two deep learning models for generating natural language sentence from robot vision and generating voice from the generated natural language sentence. Also, we conduct the real robot experiment to show the effectiveness of our method in human-robot interaction.

A feasibility study of audio-video communication application using mobile telecommunication in inter-hospital transfer situations (병원 간 전원 상황에서 이동통신망을 이용한 음성화상정보통신 애플리케이션의 유용성에 관한 연구)

  • Jeong, Joo;Kim, Tae-Han;Kang, Saee-Byel
    • The Korean Journal of Emergency Medical Services
    • /
    • v.23 no.1
    • /
    • pp.125-134
    • /
    • 2019
  • Purpose: The aim of this study was to evaluate the usefulness of video communication with medical staff located at a remote location, through a communication application connected to a long term evolution (LTE) mobile communication network in a moving ambulance. Methods: In this study, we recruited patients who were transferred by mobile intensive care unit ambulance from one hospital to another. In the moving ambulance, the information of the patient was transferred to a physician using the application in real time. Recorded video files were evaluated by emergency physicians with experience in video direct medical control. Results: A total of 18 patients were evaluated, and the average score was 5.9 out of a possible 9. It was expected that applying the use of the technology to actual clinical sites would enable the sites to provide assistance. Conclusion: In this study, we confirmed the possibility of benefiting from the clinical field when using the video-audio communication application which is connected to the remote location in real time through the current LTE mobile communication network in the ambulance.

A Study on the Development of Hard Disk Recoder and Remote Control Using Embedded Linux (임베디드 리눅스를 이용한 하드디스크 레코더 및 원격 제어 구현에 관한 연구)

  • Park, Seung-Ho;Lee, Jong-Su
    • Proceedings of the KIEE Conference
    • /
    • 2004.07d
    • /
    • pp.2429-2431
    • /
    • 2004
  • In this paper, we have designed a remote controlable HDR System using all embedded linux board. The system is composed of three parts - a HDR System, a PC client program for remote control and a Nameserver for registering and aquisition of the IP address. The system is built in an embedded board using a linux kernel. With the Linux the system can support networking and file system for a hard disk management In addition, the system embeds a web-server and a ftp-server for remote manipulation and file transfer. And the hardwares of the system are controlled by the linux device driver mechanism. MPEG1/2 technique is used to compress TV tuner signal and external analog video/audio signal. And compressed data is stored in a hard disk. The data stored in the system is accesable through lan or internet. And RTP protocol is used to enable the system to service live stream of instant video/audio input.

  • PDF

Visual Telephone System of Differential Task Interrupt Method (차등 태스크 인터럽트 방식의 영상단말 시스템)

  • 박배욱;정하재;오창석
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.6 no.5
    • /
    • pp.739-746
    • /
    • 2002
  • In this paper, a new visual telephone system which has a differential task interrupt transfer feature for real time video phone service is presented. Owing to the result of Interrupt transfer of different speed according to the time critical degree of tasks, the flow of audio and video data stream can be kept as constant speed in other word that means video phone services are carried out in real time. The ITU-T H.32x visual telephone recommendations are first analyzed, and the unsatisfactory items of existing systems are second inquired the cause, such as performance, quality. And then the design concept and ideas which enable it to solve them are third devised, the next, the new architecture of visual telephone system for real time video phone source are designed, which make it possible to solve the existing problems by means of different tasks interrupt transfer method.

Multicontents Integrated Image Animation within Synthesis for Hiqh Quality Multimodal Video (고화질 멀티 모달 영상 합성을 통한 다중 콘텐츠 통합 애니메이션 방법)

  • Jae Seung Roh;Jinbeom Kang
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.4
    • /
    • pp.257-269
    • /
    • 2023
  • There is currently a burgeoning demand for image synthesis from photos and videos using deep learning models. Existing video synthesis models solely extract motion information from the provided video to generate animation effects on photos. However, these synthesis models encounter challenges in achieving accurate lip synchronization with the audio and maintaining the image quality of the synthesized output. To tackle these issues, this paper introduces a novel framework based on an image animation approach. Within this framework, upon receiving a photo, a video, and audio input, it produces an output that not only retains the unique characteristics of the individuals in the photo but also synchronizes their movements with the provided video, achieving lip synchronization with the audio. Furthermore, a super-resolution model is employed to enhance the quality and resolution of the synthesized output.

Method of Automatically Generating Metadata through Audio Analysis of Video Content (영상 콘텐츠의 오디오 분석을 통한 메타데이터 자동 생성 방법)

  • Sung-Jung Young;Hyo-Gyeong Park;Yeon-Hwi You;Il-Young Moon
    • Journal of Advanced Navigation Technology
    • /
    • v.25 no.6
    • /
    • pp.557-561
    • /
    • 2021
  • A meatadata has become an essential element in order to recommend video content to users. However, it is passively generated by video content providers. In the paper, a method for automatically generating metadata was studied in the existing manual metadata input method. In addition to the method of extracting emotion tags in the previous study, a study was conducted on a method for automatically generating metadata for genre and country of production through movie audio. The genre was extracted from the audio spectrogram using the ResNet34 artificial neural network model, a transfer learning model, and the language of the speaker in the movie was detected through speech recognition. Through this, it was possible to confirm the possibility of automatically generating metadata through artificial intelligence.

The Development of Virtual Reality Telemedicine System for Treatment of Acrophobia (고소공포증 치료를 위한 가상현실 원격진료 시스템의 개발)

  • Ryu Jong Hyun;Beack Seung Hwa;Paek Seung Eun;Hong Sung Chan
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.52 no.4
    • /
    • pp.252-257
    • /
    • 2003
  • Acrophobia is an abnormal fear of heights. Medications or cognitive-behavior methods have been mainly used as a treatment. Lately the virtual reality technology has been applied to that kind of anxiety disorders. A virtual environment provides patient with stimuli which arouses phobia, and exposing to that environment makes him having ability to over come the fear. Recently, the patient can take diagnose from a medical doctor in distance with the telemedicine system. The hospital and doctors can get the medical data, audio, video, signals in the actual examination room or operating room via a live interactive system. Audio visual and multimedia conference service, online questionary, ECG signal transfer system, update system are needed in this system. Virtual reality simulation system that composed with a position sensor, head mount display, and audio system, is also included in this telemedicine system. In this study, we tried this system to the acrophobia patient in distance.