• Title/Summary/Keyword: Immersive Media

Search Result 231, Processing Time 0.031 seconds

Performance Evaluation of ARCore Anchors According to Camera Tracking

  • Shinhyup Lee;Leehwan Hwang;Seunghyun Lee;Taewook Kim;Soonchul Kwon
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.15 no.4
    • /
    • pp.215-222
    • /
    • 2023
  • Augmented reality (AR), which integrates virtual media into reality, is increasingly utilized across various industrial sectors, thanks to advancements in 3D graphics and mobile device technologies. The IT industry is thus carrying out active R&D activities about AR platforms. Google plays a significant role in the AR landscape, with a focus on ARCore services. An essential aspect of ARCore is the use of anchors, which serve as reference points that help maintain the position and orientation of virtual objects within the physical environment. However, if the accuracy of anchor positioning is suboptimal when running AR content, it can significantly diminish the user's immersive experience. We are to assess the performance of these anchors in this study. To conduct the performance evaluation, virtual 3D objects, matching the shape and size of real-world objects, we strategically positioned ourselves to overlap with their physical counterparts. Images of both real and virtual objects were captured from five distinct camera trajectories, and ARCore's performance was analyzed by examining the difference between these captured images.

Implementation of Immersive Interactive Content Using Face Recognition Technology - (Exhibition of ReneMagritte) Focused on 'ARPhotoZone' (얼굴 인식 기술을 활용한 실감형 인터랙티브 콘텐츠의 구현 - (르네마그리트 특별전) AR포토존을 중심으로)

  • Lee, Eun-Jin;Sung, Jung-Hwan
    • Journal of Korea Game Society
    • /
    • v.20 no.5
    • /
    • pp.13-20
    • /
    • 2020
  • Biometric technology with the advance of deep learning enabled the new types of content. Especially, face recognition can provide immersion in terms of convenience and non-compulsiveness, but most commercial content has limitations that are limited to application areas. In this paper, we attempted to overcome these limitations, implement content that can utilize face recognition technology based on realtime video feed. We used Unity engine for high quality graphics, but performance degradation and frame drop occurred. To solve them, we augmented Dlib toolkit and adjusted the resolution image.

Motion-based Controlling 4D Special Effect Devices to Activate Immersive Contents (실감형 콘텐츠 작동을 위한 모션 기반 4D 특수효과 장치 제어)

  • Kim, Kwang Jin;Lee, Chil Woo
    • Smart Media Journal
    • /
    • v.8 no.1
    • /
    • pp.51-58
    • /
    • 2019
  • This paper describes a gesture application to controlling the special effects of physical devices for 4D contents using the PWM (Pulse Width Modulation) method. The user operation recognized by the infrared sensor is interpreted as a command for 3D content control, several of which manipulate the device that generates the special effect to display the physical stimulus to the user. With the content controlled under the NUI (Natural User Interface) technique, the user can be directly put into an immersion experience, which leads to provision of the higher degree of interest and attention. In order to measure the efficiency of the proposed method, we implemented a PWM-based real-time linear control system that manages the parameters of the motion recognition and animation controller using the infrared sensor and transmits the event.

Camera pose estimation framework for array-structured images

  • Shin, Min-Jung;Park, Woojune;Kim, Jung Hee;Kim, Joonsoo;Yun, Kuk-Jin;Kang, Suk-Ju
    • ETRI Journal
    • /
    • v.44 no.1
    • /
    • pp.10-23
    • /
    • 2022
  • Despite the significant progress in camera pose estimation and structure-from-motion reconstruction from unstructured images, methods that exploit a priori information on camera arrangements have been overlooked. Conventional state-of-the-art methods do not exploit the geometric structure to recover accurate camera poses from a set of patch images in an array for mosaic-based imaging that creates a wide field-of-view image by sewing together a collection of regular images. We propose a camera pose estimation framework that exploits the array-structured image settings in each incremental reconstruction step. It consists of the two-way registration, the 3D point outlier elimination and the bundle adjustment with a constraint term for consistent rotation vectors to reduce reprojection errors during optimization. We demonstrate that by using individual images' connected structures at different camera pose estimation steps, we can estimate camera poses more accurately from all structured mosaic-based image sets, including omnidirectional scenes.

Fast Content-preserving Seam Estimation for Real-time High-resolution Video Stitching (실시간 고해상도 동영상 스티칭을 위한 고속 콘텐츠 보존 시접선 추정 방법)

  • Kim, Taeha;Yang, Seongyeop;Kang, Byeongkeun;Lee, Hee Kyung;Seo, Jeongil;Lee, Yeejin
    • Journal of Broadcast Engineering
    • /
    • v.25 no.6
    • /
    • pp.1004-1012
    • /
    • 2020
  • We present a novel content-preserving seam estimation algorithm for real-time high-resolution video stitching. Seam estimation is one of the fundamental steps in image/video stitching. It is to minimize visual artifacts in the transition areas between images. Typical seam estimation algorithms are based on optimization methods that demand intensive computations and large memory. The algorithms, however, often fail to avoid objects and results in cropped or duplicated objects. They also lack temporal consistency and induce flickering between frames. Hence, we propose an efficient and temporarily-consistent seam estimation algorithm that utilizes a straight line. The proposed method also uses convolutional neural network-based instance segmentation to locate seam at out-of-objects. Experimental results demonstrate that the proposed method produces visually plausible stitched videos with minimal visual artifacts in real-time.

A Study on Metaverse Hype for Sustainable Growth

  • Lee, Jee Young
    • International journal of advanced smart convergence
    • /
    • v.10 no.3
    • /
    • pp.72-80
    • /
    • 2021
  • Metaverse is an immersive 3D virtual environment, a true virtual artificial community in which avatars act as the user's alter ego and interact with each other. If we do not manage the hype for the metaverse, which has recently been receiving a surge in interest, the metaverse will fail to cross the chasm. In this study, to provide stakeholders with insights for the successful introduction and growth of the 3D immersive next-generation virtual world, metaverse, we analyzed user-side interest, media-side interest, and research-side interest. For this purpose, in this study, search traffic, news frequency and topic, and research article frequency and topic were analyzed. The methodology and results of this study are expected to provide insight for the stable success of metaverse transformation and the coexistence of the real world and the virtual world through hyper-connection and hyper-convergence.

Standardization Trend of 3DoF+ Video for Immersive Media (이머시브미디어를 3DoF+ 비디오 부호화 표준 동향)

  • Lee, G.S.;Jeong, J.Y.;Shin, H.C.;Seo, J.I.
    • Electronics and Telecommunications Trends
    • /
    • v.34 no.6
    • /
    • pp.156-163
    • /
    • 2019
  • As a primitive immersive video technology, a three degrees of freedom (3DoF) $360^{\circ}$ video can currently render viewport images that are dependent on the rotational movements of the viewer. However, rendering a flat $360^{\circ}$ video, that is supporting head rotations only, may generate visual discomfort especially when objects close to the viewer are rendered. 3DoF+ enables head movements for a seated person adding horizontal, vertical, and depth translations. The 3DoF+ $360^{\circ}$ video is positioned between 3DoF and six degrees of freedom, which can realize the motion parallax with relatively simple virtual reality software in head-mounted displays. This article introduces the standardization trends for the 3DoF+ video in the MPEG-I visual group.

A study on Metaverse keyword Consumer perception survey after Covid-19 using big Data

  • LEE, JINHO;Byun, Kwang Min;Ryu, Gi Hwan
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.14 no.4
    • /
    • pp.52-57
    • /
    • 2022
  • In this study, keywords from representative online portal sites such as Naver, Google, and Youtube were collected based on text mining analysis technique using Textom to check the changes in metqaverse after COVID-19. before Corona, it was confirmed that social media platforms such as Kakao Talk, Facebook, and Twitter were mentioned, and among the four metaverse, consumer awareness was still concentrated in the field of life logging. However, after Corona, keywords from Roblox, Fortnite, and Geppetto appeared, and keywords such as Universe, Space, Meta, and the world appeared, so Metaverse was recognized as a virtual world. As a result, it was confirmed that consumer perception changed from the life logging of Metaverse to the mirror world. Third, keywords such as cryptocurrency, cryptocurrency, coin, and exchange appeared before Corona, and the word frequency ranking for blockchain, which is an underlying technology, was high, but after Corona, the word frequency ranking fell significantly as mentioned above.

Asymmetric Threshold-Based Occupancy Map Correction for Efficient Coding of MPEG Immersive Video (MIV 의 효율적인 부호화를 위한 비대칭 임계값 기반 점유맵 보정)

  • Dong-Ha Kim;Sung-Gyun Lim;Jeong-yoon Kim;Jae-Gon Kim
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2022.11a
    • /
    • pp.51-53
    • /
    • 2022
  • MIV(MPEG Immersive Video)의 시험모델 TMIV 는 다시점의 비디오와 깊이(depth) 비디오를 입력 받아 시점 사이의 중복성을 제거한 후 남은 텍스처(texture)와 깊이로 텍스처 아틀라스(atlas)와 깊이 아틀라스를 각각 생성하고 이를 압축한다. 각 화소별 점유(occupancy) 정보는 깊이 아틀라스에 포함되어 압축되는데 압축 손실로 인한 점유맵 오류를 방지하기 위하여 임계값 T = 64 로 설정한 보호대역을 사용한다. 기존에 설정된 임계값을 낮추어 깊이 동적범위를 확대하면 보다 정확한 깊이값 표현으로 부호화 효율을 개선할 수 있지만 보호대역 축소로 점유맵 오류가 증가한다. 본 논문에서는 TMIV 의 부호화기와 보호화기에 비대칭 임계값을 사용하여 보호대역 축소로 인한 점유맵 오류를 보정하면서 보다 정확한 깊이 값 표현을 통하여 부호화 효율을 개선하는 기법을 제안한다. 제안기법은 깊이 동적범위 확대와 비대칭 임계값 기반의 점유맵 오류 보정을 통하여 CG 시퀀스에서 2.2% BD-rate 이득과 주관적 화질 개선을 보인다.

  • PDF

Performance Analysis of 3DoF+ Video Coding Using V3C (V3C 기반 3DoF+ 비디오 부호화 성능 분석)

  • Lee, Ye-Jin;Yoon, Yong-Uk;Kim, Jae-Gon
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2020.11a
    • /
    • pp.166-168
    • /
    • 2020
  • MPEG 비디오 그룹은 MPEG-I 표준의 일부로 포인트 클라우드(Point Cloud) 압축을 위한 비디오 기반 포인트 클라우드 부호화(V-PCC)와 몰입형(immersive) 비디오 압축을 위한 MPEG Immersive Video(MIV) 표준을 개발하고 있다. 최근에는 포인트 클라우드 및 몰입형 비디오와 같은 체적형(volumetric) 비디오를 모두 압축할 수 있도록 V-PCC 와 MIV 를 통합한 V3C(Visual Volumetric Video-based Coding) 표준화를 진행하고 있다. 본 논문에서는 V3C 코덱을 사용한 3DoF+(3 Degree of Freedom plus) 비디오 부호화 방안을 분석한다. 또한 V3C 코덱의 2D 코덱으로 기존 HEVC 대신 VVC 를 사용할 경우의 부호화 성능 향상을 분석한다.

  • PDF