• Title/Summary/Keyword: 2D Video

Search Result 910, Processing Time 0.046 seconds

Design and Implementation of LTE-TDD 2×2 MIMO Bidirectional RF Hybrid Beamforming System (LTE-TDD 2×2 MIMO 양방향 RF 하이브리드 빔포밍 시스템 설계 및 구현)

  • Lee, Kwang-Suk;Kim, Dong-Hyun;Oh, Hyuk-Jun
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.23 no.4
    • /
    • pp.23-31
    • /
    • 2018
  • This paper presented the implementation and design of the 2T-2R wireless HD video streaming systems over 1.7 GHz frequency band using 3GPP LTE-TDD standard on NI USRP RIO SDR platform. The baseband of the system used USRP RIO that are stored in Xilinx Kintex-7 chip to implement LTE-TDD transceiver modem, the signal that are transmitted from USRP RIO up or down converts to 1.7 GHz by using self-designed 1.7 GHz RF transceiver modules and it is finally communicated HD video data through self-designed 2x9 sub array antennas. It is that communication method between USRP RIO and Host PC use PCI express x4 to minimize delay of data to transmit and receive. The implemented system show high error vector magnitude performance above 32 dBc and to transmit and receive HD video in experiment environment anywhere. The proposed hybrid beam forming system could be used not only in the future 5G mobile communication systems under 6 GHz frequency band but also in the systems over 6 GHz frequency band like ones in mmWave frequency bands.

Perception based video anticipation generation (선택적 주의 기법 기반의 영상의 기대효과 자동생성)

  • Yoon, Jong-Chul;Lee, In-Kwon
    • Journal of the Korea Computer Graphics Society
    • /
    • v.13 no.3
    • /
    • pp.1-6
    • /
    • 2007
  • Anticipation effect has been used as a traditional skill to enhance the dynamic motion of the traditional 2D animation. Basically, anticipation means the action of opposite direction which performs before the real action step. In this paper, we propose the perception-based video anticipation method to guide a user's visual attention to the important region. Using the image based attention map, we calculate the visual attention region and then combine this map with temporal saliency of video. We apply the anticipation effect in these saliency regions using the blur kernel. Using our method, we can generate the dynamic video motion which has attentive guidance.

  • PDF

Producing Stereoscopic Video Contents Using Transformation of Character Objects (캐릭터 객체의 변환을 이용하는 입체 동영상 콘텐츠 제작)

  • Lee, Kwan-Wook;Won, Ji-Yeon;Choi, Chang-Yeol;Kim, Man-Bae
    • Journal of Broadcast Engineering
    • /
    • v.16 no.1
    • /
    • pp.33-43
    • /
    • 2011
  • Recently, 3D displays are supplied in the 3D markets so that the demand for 3D stereoscopic contents increases. In general, a simple method is to use a stereoscopic camera. As well, the production of 3D from 2D materials is regarded as an important technology. Such conversion works have gained much interest in the field of 3D converting. However, the stereoscopic image generation from a single 2D image is limited to simple 2D to 3D conversion so that the better realistic perception is difficult to deliver to the users. This paper presents a new stereoscopic content production method where foreground objects undergo alive action events. Further stereoscopic animation is viewed on 3D displays. Given a 2D image, the production is composed of background image generation, foreground object extraction, object/background depth maps and stereoscopic image generation The alive objects are made using the geometric transformation (e.g., translation, rotation, scaling, etc). The proposed method is performed on a Korean traditional painting, Danopungjung as well as Pixar's Up. The animated video showed that through the utilization of simple object transformations, more realistic perception can be delivered to the viewers.

The Clinical Study on the Relationship between State of the Tympanic Membrane and Treatment Period in the Acute Otitis Media by Video Otoscope (video otoscope로 관찰한 급성 중이염 소아 환자의 고막상태와 치료기간과의 연관성에 대한 연구)

  • Yu, Hyun-Jung;Choi, In-Hwa
    • The Journal of Korean Medicine Ophthalmology and Otolaryngology and Dermatology
    • /
    • v.18 no.3
    • /
    • pp.95-101
    • /
    • 2005
  • Objective: In children with acute otitis media, we investigated the relationship between state of tympanic membrane and treatment period by video otoscope. Methods: Ten children(6 boys, 4 girls) with acute otitis media were assigned to one of three groups. In the grade I, there were redness in the tympanic membrane. In the grade II, there were bulging, effusion with exudate, dark color in the tympanic membrane. In the grade III, there were effusion with purulent exudate or cholesteatoma with serous exudate. And we investigated the treatment period in each group. Results: Overall, the $mean{\pm}S.D.$ on the treatment period in the grade I was shorter than grade II. It was $12.6{\pm}5.18$ in the grade I versus $26{\pm}2.83$ in the grade II. That in the grade II was shorter than grade III. It was $26{\pm}2.83$ in the grade II versus $148{\pm}43.03$ in the grade III. Conclusions: Although we couldn't have enough cases in each group. we could predict the conclusion that there was the relationship between treatment period and state of tympanic membrane.

  • PDF

Fast Macroblock Mode Selection Algorithm for B Frames in Multiview Video Coding

  • Yu, Mei;He, Ping;Peng, Zongju;Zhang, Yun;Si, Yuehou;Jiang, Gangyi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.5 no.2
    • /
    • pp.408-427
    • /
    • 2011
  • Intensive computational complexity is an obstacle of enabling multiview video coding for real-time applications. In this paper, we present a fast macroblock (MB) mode selection algorithm for B frames which are based on the computational complexity analyses between the MB mode selection and reference frame selection. Three strategies are proposed to reduce the coding complexity jointly. First, the temporal correlation of MB modes between current MB and its temporal corresponding MBs is utilized to reduce computational complexity in determining the optimal MB mode. Secondly, Lagrangian cost of SKIP mode is compared with that of Inter $16{\times}16$ modes to early terminate the mode selection process. Thirdly, reference frame correlation among different Inter modes is exploited to reduce the number of reference frames. Experimental results show that the proposed algorithm can promote the encoding speed by 3.71~7.22 times with 0.08dB PSNR degradation and 2.03% bitrate increase on average compared with the joint multiview video model.

2D Interpolation 3D Point Cloud using Video-based Point Cloud Compression (비디오 기반 포인트 클라우드 압축을 사용한 차원 포인트 클라우드의 차원 보간 방안)

  • Hwang, Yonghae;Kim, Junsik;Kim, Kyuheon
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2021.06a
    • /
    • pp.147-150
    • /
    • 2021
  • 최근 컴퓨터 그래픽 기술이 발전함에 따라 가상으로 만들어낸 객체와 현실 객체 사이의 분간이 어려워지고 있으며, AR/VR/XR 등의 서비스를 위해 현실 객체를 컴퓨터 그래픽으로 표현하는 기술의 연구가 활발히 진행되고 있다. 포인트 클라우드는 현실 객체를 표현하는 기술 중의 하나로 객체의 표면을 수많은 3차원의 점으로 표현하며, 2차원 영상보다 더욱 거대한 데이터 크기를 가지게 된다. 이를 다양한 서비스에 응용하기 위해서는 3차원 데이터의 특징에 맞는 고효율의 압축 기술이 필요하며, 국제표준기구인 MPEG에서는 연속적인 움직임을 가지는 동적 포인트 클라우드를 2차원 평면으로 투영하여 비디오 코덱을 사용해 압축하는 Video-based Point Cloud Compression (V-PCC) 기술이 연구되고 있다. 포인트 클라우드를 2차원 평면에 투영하는 방식은 점유 맵 (Occupancy Map), 기하 영상 (Geometry Image), 속성 영상 (Attribute Image) 등의 2차원 정보와 보조 정보를 사용해 압축을 진행하고, 부호화 과정에서는 보조 정보와 2차원 영상들의 정보를 사용해 3차원 포인트 클라우드를 재구성한다. 2차원 영상을 사용해 포인트 클라우드를 생성하는 특징 때문에 압축 과정에서 발생하는 영상 정보의 열화는 포인트 클라우드의 품질에 영향을 미친다. 이와 마찬가지로 추가적인 기술을 사용한 2차원 영상 정보의 향상으로 포인트 클라우드의 품질을 향상할 수 있을 것으로 예상된다. 이에 본 논문은 V-PCC 기술에서 생성되는 영상 정보에 2차원 보간 (Interpolation) 기술을 적용하여 기존의 영상 정보에 포함되지 않은 추가적인 포인트를 생성하는 것으로 재구성되는 포인트 클라우드의 밀도를 증가시키고 그 영향을 분석하고자 한다.

  • PDF

Adaptive Spatio-Temporal Prediction for Multi-view Coding in 3D-Video (3차원 비디오 압축에서의 다시점 부호화를 위한 적응적 시공간적 예측 부호화)

  • 성우철;이영렬
    • Journal of Broadcast Engineering
    • /
    • v.9 no.3
    • /
    • pp.214-224
    • /
    • 2004
  • In this paper, an adaptive spatio-temporal predictive coding based on the H.264 is proposed for 3D immersive media encoding, such as 3D image processing, 3DTV, and 3D videoconferencing. First, we propose a spatio-temporal predictive coding using the same view and inter-view images for the two TPPP, IBBP GOP (group of picture) structures 4hat are different from the conventional simulcast method. Second, an 2D inter-view direct mode for the efficient prediction is proposed when the proposed spatio-temporal prediction uses the IBBP structure. The 2D inter-view direct mode is applied when the temporal direct mode in B(hi-Predictive) picture of the H.264 refers to an inter-view image, since the current temporal direct mode in the H.264 standard could no: be applied to the inter-view image. The proposed method is compared to the conventional simulcast method in terms of PSNR (peak signal to noise ratio) for the various 3D test video sequences. The proposed method shows better PSNR results than the conventional simulcast mode.

A Study on Kernel Size Variations in 1D Convolutional Layer for Single-Frame supervised Temporal Action Localization (단일 프레임 지도 시간적 행동 지역화에서 1D 합성곱 층의 커널 사이즈 변화 연구)

  • Hyejeong Jo;Huiwon Gwon;Sunhee Jo;Chanho Jung
    • Journal of IKEEE
    • /
    • v.28 no.2
    • /
    • pp.199-203
    • /
    • 2024
  • In this paper, we propose variations in the kernel size of 1D convolutional layers for single-frame supervised temporal action localization. Building upon the existing method, which utilizes two 1D convolutional layers with kernel sizes of 3 and 1, we introduce an approach that adjusts the kernel sizes of each 1D convolutional layer. To validate the efficiency of our proposed approach, we conducted comparative experiments using the THUMOS'14 dataset. Additionally, we use overall video classification accuracy, mAP (mean Average Precision), and Average mAP as performance metrics for evaluation. According to the experimental results, our proposed approach demonstrates higher accuracy in terms of mAP and Average mAP compared to the existing method. The method with variations in kernel size of 7 and 1 further demonstrates an 8.0% improvement in overall video classification accuracy.

Adaptive Correlation Noise Model for DC Coefficients in Wyner-Ziv Video Coding

  • Qin, Hao;Song, Bin;Zhao, Yue;Liu, Haihua
    • ETRI Journal
    • /
    • v.34 no.2
    • /
    • pp.190-198
    • /
    • 2012
  • An adaptive correlation noise model (CNM) construction algorithm is proposed in this paper to increase the efficiency of parity bits for correcting errors of the side information in transform domain Wyner-Ziv (WZ) video coding. The proposed algorithm introduces two techniques to improve the accuracy of the CNM. First, it calculates the mean of direct current (DC) coefficients of the original WZ frame at the encoder and uses it to assist the decoder to calculate the CNM parameters. Second, by considering the statistical property of the transform domain correlation noise and the motion characteristic of the frame, the algorithm adaptively models the DC coefficients of the correlation noise with the Gaussian distribution for the low motion frames and the Laplacian distribution for the high motion frames, respectively. With these techniques, the proposed algorithm is able to make a more accurate approximation to the real distribution of the correlation noise at the expense of a very slight increment to the coding complexity. The simulation results show that the proposed algorithm can improve the average peak signal-to-noise ratio of the decoded WZ frames by 0.5 dB to 1.5 dB.

Parametric Video Compression Based on Panoramic Image Modeling (파노라믹 영상 모델에 근거한 파라메트릭 비디오 압축)

  • Sim Dong-Gyu
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.4 s.310
    • /
    • pp.96-107
    • /
    • 2006
  • In this paper, a low bitrate video coding method based on new panoramic modeling is proposed for panning cameras. An input video frame from a panning camera is decomposed into a background image, rectangular moving object regions, and a residual image. In coding the background, we employ a panoramic model that can account for several image formation processes, such as perspective projection, lens distortion, vignetting and illumination effects. Moving objects aredetected, and their minimum bounding rectangular regions are coded with a JPEG-2000 coder. We have evaluated the effectiveness of the proposed algorithm with several indoor and outdoor sequences and found that the PSNR is improved by $1.3{\sim}4.4dB$ compared to that of JPEG-2000.