• Title/Summary/Keyword: Video Synthesis

Search Result 116, Processing Time 0.021 seconds

Implementation of the Frame Memory Hardware for MPEG-2 Video Encoder (MPEG-2 비디오 부호화기의 프레임 메모리 하드웨어 구현)

  • 고영기;강의성;이경훈;고성제
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.9A
    • /
    • pp.1442-1450
    • /
    • 1999
  • In this paper, we present an efficient hardware architecture for the frame memory of the MPEG-2 video encoder. Both the total size of internal buffers and the number of logic gates are reduced by the proposed memory map which can provide an effective interface between MPEG-2 video encoder and the external DRAM. Furthermore, the proposed scheme can reduce the DRAM access time. To realize the frame memory hardware,$0.5\mu\textrm{m}$, VTI, vemn5a3 standard cell library is used. VHDL simulator and logic synthesis tool are used for hardware design and RTL (register transfer level) function verification. The frame memory hardware emulator of the proposed architecture is designed for gate-level function verification. It is expected that the proposed frame memory hardware using VHDL can achieve suitable performance for MPEG-2 MP@ML.

  • PDF

The User Interface of Button Type for Stereo Video-See-Through (Stereo Video-See-Through를 위한 버튼형 인터페이스)

  • Choi, Young-Ju;Seo, Young-Duek
    • Journal of the Korea Computer Graphics Society
    • /
    • v.13 no.2
    • /
    • pp.47-54
    • /
    • 2007
  • This paper proposes a user interface based on video see-through environment which shows the images via stereo-cameras so that the user can control the computer systems or other various processes easily. We include an AR technology to synthesize virtual buttons; the graphic images are overlaid on the captured frames taken by the camera real-time. We search for the hand position in the frames to judge whether or not the user selects the button. The result of judgment is visualized through changing of the button color. The user can easily interact with the system by selecting the virtual button in the screen with watching the screen and moving her fingers at the air.

  • PDF

Reduced Reference Quality Metric for Synthesized Virtual Views in 3DTV

  • Le, Thanh Ha;Long, Vuong Tung;Duong, Dinh Trieu;Jung, Seung-Won
    • ETRI Journal
    • /
    • v.38 no.6
    • /
    • pp.1114-1123
    • /
    • 2016
  • Multi-view video plus depth (MVD) has been widely used owing to its effectiveness in three-dimensional data representation. Using MVD, color videos with only a limited number of real viewpoints are compressed and transmitted along with captured or estimated depth videos. Because the synthesized views are generated from decoded real views, their original reference views do not exist at either the transmitter or receiver. Therefore, it is challenging to define an efficient metric to evaluate the quality of synthesized images. We propose a novel metric-the reduced-reference quality metric. First, the effects of depth distortion on the quality of synthesized images are analyzed. We then employ the high correlation between the local depth distortions and local color characteristics of the decoded depth and color images, respectively, to achieve an efficient depth quality metric for each real view. Finally, the objective quality metric of the synthesized views is obtained by combining all the depth quality metrics obtained from the decoded real views. The experimental results show that the proposed quality metric correlates very well with full reference image and video quality metrics.

Robot Vision to Audio Description Based on Deep Learning for Effective Human-Robot Interaction (효과적인 인간-로봇 상호작용을 위한 딥러닝 기반 로봇 비전 자연어 설명문 생성 및 발화 기술)

  • Park, Dongkeon;Kang, Kyeong-Min;Bae, Jin-Woo;Han, Ji-Hyeong
    • The Journal of Korea Robotics Society
    • /
    • v.14 no.1
    • /
    • pp.22-30
    • /
    • 2019
  • For effective human-robot interaction, robots need to understand the current situation context well, but also the robots need to transfer its understanding to the human participant in efficient way. The most convenient way to deliver robot's understanding to the human participant is that the robot expresses its understanding using voice and natural language. Recently, the artificial intelligence for video understanding and natural language process has been developed very rapidly especially based on deep learning. Thus, this paper proposes robot vision to audio description method using deep learning. The applied deep learning model is a pipeline of two deep learning models for generating natural language sentence from robot vision and generating voice from the generated natural language sentence. Also, we conduct the real robot experiment to show the effectiveness of our method in human-robot interaction.

Method of Video Stitching based on Minimal Error Seam (최소 오류 경계를 활용한 동적 물체 기반 동영상 정합 방안)

  • Kang, Jeonho;Kim, Junsik;Kim, Sang-IL;Kim, Kyuheon
    • Journal of Broadcast Engineering
    • /
    • v.24 no.1
    • /
    • pp.142-152
    • /
    • 2019
  • There is growing interest in ultra-high-resolution content that gives a more realistic sense of presence than existing broadcast content. However, in order to provide ultra-high-resolution contents in existing broadcast services, there are limitations in view angle and resolution of the image acquisition device. In order to solve this problem, many researches on stitching, which is an image synthesis method using a plurality of input devices, have been conducted. In this paper, we propose method of dynamic object based video stitching using minimal error seam in order to overcome the temporal invariance degradation of moving objects in the stitching process of horizontally oriented videos.

Virtual Viewpoint Video Synthesis Using Symmetric Belief Propagation (대칭적 신뢰 전파 기법 기반의 가상 시점 비디오 생성)

  • Jung, Il-Lyong;Chung, Tae-Young;Kim, Chang-Su
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2008.11a
    • /
    • pp.113-116
    • /
    • 2008
  • 본 논문에서는 다시점 비디오(multi-view video)에서 보다 다양한 시점을 제공하기 위한 가상 시점 비디오 생성 기법을 제안한다. 제안하는 가상 시점 비디오 생성 기법은 우선적으로 대칭적 신뢰 전파 기법(symmetric belief propagation)을 기반으로, 각 시점의 깊이 정보 및 폐색 영역(occlusion region)을 추출하기 위해서 에너지를 최소화한다. 추출된 깊이 정보 및 에너지를 이용하여 참조하는 시점 간의 가중치를 적용하여, 새로운 가상 시점의 비디오를 생성하고, 추출된 폐색 영역의 값을 이용하여, 가상 시점의 비디오를 보정하는 가상 시점 비디오 생성 기법을 제안한다. 또한 제안하는 알고리즘을 한정된 중간 시점 영상에서 임의의 가상 시점으로 확장하여, 임의의 두 시점 간의 자유로운 시점(free-view point)을 제공함을 확인한다. 실험을 통하여 제안하는 기법이 다시점 비디오에서 높은 화질의 가상 시점 비디오를 제공함을 확인한다.

  • PDF

Acquisition Workflow of Multiview Stereoscopic Video at Real and CG Environment (실사 및 CG 환경에서의 다시점 입체영상 획득 기술)

  • Jeong, Jun Young;Yun, Kug Jin;Cheong, Won-Sik
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2022.06a
    • /
    • pp.51-53
    • /
    • 2022
  • 고정된 위치를 중심으로 회전운동만 체험할 수 있는 3 자유도(DoF: Degrees of Freedom)를 넘어 위치를 변경하며 운동시차까지 포함된 6 자유도를 지원하는 몰입형 미디어에 대한 연구가 지속해서 진행되고 있다. 특히 부드러운 시점 변경을 제공하기 위해 특정 위치에서 샘플링 된 여러 개의 텍스쳐(또는 컬러) 및 깊이맵 영상(MVD: Multiview Video plus Depth)으로 구성된 다시점 영상을 통해 실제로 획득되지 않은 위치에서의 영상을 만들어내는 가상시점 합성(virtual view synthesis) 기술이 많이 사용되고 있다. 본 논문에서는 몰입형 미디어의 대표적인 데이터 형식인 다시점 영상을 실사 및 컴퓨터 그래픽스(CG: Computer Graphics) 환경에서 획득하는 방법에 관해 설명한다.

  • PDF

Effectiveness of Education Program Using Video Recording and Feedback on Skill Competency for Students of Majors in Health Care: A Meta-Analysis (동영상 촬영과 피드백을 이용한 실습교육이 보건의료전공 대학생의 술기수행능력에 미치는 효과: 메타분석)

  • Shin, Yun Hee;Kim, Sun Kyung;Kim, Hyunjoo
    • Journal of Korean Academy of Fundamentals of Nursing
    • /
    • v.25 no.2
    • /
    • pp.120-133
    • /
    • 2018
  • Purpose: A meta-analysis was conducted to evaluate effectiveness of education programs using video recording and feedback in the improvement of competency in clinical skills for health care majors. Methods: Six databases were searched and inclusion criteria were randomized controlled trials (RCTs) or non-randomized controlled trials (NRTs) reporting level of skill competency using numerical measurements. Data analysis and synthesis were performed using Comprehensive Meta-Analysis software and Revman program. Results: Of 1,568 records, 11 studies met inclusion criteria. Statistically significant effectiveness of education programs using video recording and feedback was identified. A low risk of bias was detected among both RCTs and NRTs. Meta-analysis showed that the intervention groups had more effective improvements in skill competency (standardized mean difference [SMD]: 0.74; 95% CI: 0.33~1.16). Results of subgroup analysis showed higher effects when interventions dealt with one skill, used self-reflection with expert feedback, and included instruction from instructor in the education programs. Conclusion: Findings suggest that schools for health care majors should actively adopt video and feedback based skill training allowing educators to design effective programs. Potential is higher for students to achieve higher competency when they train with one skill at a time, use of instruction and receive feedback from experts.

Generating Augmented Lifting Player using Pose Tracking

  • Choi, Jong-In;Kim, Jong-Hyun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.25 no.5
    • /
    • pp.19-26
    • /
    • 2020
  • This paper proposes a framework for creating acrobatic scenes such as soccer ball lifting using various users' videos. The proposed method can generate a desired result within a few seconds using a general video of user recorded with a mobile phone. The framework of this paper is largely divided into three parts. The first is to analyze the posture by receiving the user's video. To do this, the user can calculate the pose of the user by analyzing the video using a deep learning technique, and track the movement of a selected body part. The second is to analyze the movement trajectory of the selected body part and calculate the location and time of hitting the object. Finally, the trajectory of the object is generated using the analyzed hitting information. Then, a natural object lifting scenes synchronized with the input user's video can be generated. Physical-based optimization was used to generate a realistic moving object. Using the method of this paper, we can produce various augmented reality applications.

Group-based Adaptive Rendering for 6DoF Immersive Video Streaming (6DoF 몰입형 비디오 스트리밍을 위한 그룹 분할 기반 적응적 렌더링 기법)

  • Lee, Soonbin;Jeong, Jong-Beom;Ryu, Eun-Seok
    • Journal of Broadcast Engineering
    • /
    • v.27 no.2
    • /
    • pp.216-227
    • /
    • 2022
  • The MPEG-I (Immersive) group is working on a standardization project for immersive video that provides 6 degrees of freedom (6DoF). The MPEG Immersion Video (MIV) standard technology is intended to provide limited 6DoF based on depth map-based image rendering (DIBR) technique. Many efficient coding methods have been suggested for MIV, but efficient transmission strategies have received little attention in MPEG-I. This paper proposes group-based adaptive rendering method for immersive video streaming. Each group can be transmitted independently using group-based encoding, enabling adaptive transmission depending on the user's viewport. In the rendering process, the proposed method derives weights of group for view synthesis and allocate high quality bitstream according to a given viewport. The proposed method is implemented through the Test Model for Immersive Video (TMIV) test model. The proposed method demonstrates 17.0% Bjontegaard-delta rate (BD-rate) savings on the peak signalto-noise ratio (PSNR) and 14.6% on the Immersive Video PSNR(IV-PSNR) in terms of various end-to-end evaluation metrics in the experiment.