• Title/Summary/Keyword: Cloud Rendering

Search Result 34, Processing Time 0.032 seconds

A 2-Tier Server Architecture for Real-time Multiple Rendering (실시간 다중 렌더링을 위한 이중 서버 구조)

  • Lim, Choong-Gyoo
    • Journal of Korea Game Society
    • /
    • v.12 no.4
    • /
    • pp.13-22
    • /
    • 2012
  • The wide-spread use of the broadband Internet service makes the cloud computing-based gaming service possible. A game program is executed on a cloud node and its live image is fed into a remote user's display device via video streaming. The user's input is immediately transmitted and applied to the game. The minimization of the time to process remote user's input and transmit the live image back to the user and thus satisfying the requirement of instant responsiveness for gaming makes it possible. However, the cost to build its servers can be very expensive to provide high quality 3D games because a general purpose graphics system that cloud nodes are likely to have for the service supports a single 3D application at a time. Thus, the server must have a technology of 'realtime multiple rendering' to execute multiple 3D games simultaneously. This paper proposes a new architecture of 2-tier servers of clouds nodes of which one group executes multiple games and the other produces game's live images. It also performs a few experimentations to prove the feasibility of the new architecture.

Rendering Quality Improvement Method based on Depth and Inverse Warping (깊이정보와 역변환 기반의 포인트 클라우드 렌더링 품질 향상 방법)

  • Lee, Heejea;Yun, Junyoung;Park, Jong-Il
    • Journal of Broadcast Engineering
    • /
    • v.26 no.6
    • /
    • pp.714-724
    • /
    • 2021
  • The point cloud content is immersive content recorded by acquiring points and colors corresponding to the real environment and objects having three-dimensional location information. When a point cloud content consisting of three-dimensional points having position and color information is enlarged and rendered, the gap between the points widens and an empty hole occurs. In this paper, we propose a method for improving the quality of point cloud contents through inverse transformation-based interpolation using depth information for holes by finding holes that occur due to the gap between points when expanding the point cloud. The points on the back are rendered between the holes created by the gap between the points, acting as a hindrance to applying the interpolation method. To solve this, remove the points corresponding to the back side of the point cloud. Next, a depth map at the point in time when an empty hole is generated is extracted. Finally, inverse transform is performed to extract pixels from the original data. As a result of rendering content by the proposed method, the rendering quality improved by 1.2 dB in terms of average PSNR compared to the conventional method of increasing the size to fill the blank area.

Development of a Remote Rendering System using Direct3D API (Direct3D API의 원격 실시간 실행 시스템 개발)

  • Lim, Choong-Gyoo
    • Journal of Korea Game Society
    • /
    • v.14 no.5
    • /
    • pp.117-126
    • /
    • 2014
  • There are various kinds of applications if one can develop a remote execution system using for legacy 3D APIs. It can be used in implementing a cloud gaming service based on the real-time video streaming technology. Or, it can also be used in implementing a GPU virtualization for simultaneously rendering of many different 3D applications. The OpenGL API consists of independent global functions while the Direct3D API consists of Microsoft COM-based interfaces and their member functions, which makes the implementation of remote rendering system more difficult. The purpose of the paper is to show the applicability of the technology to any legacy 3D API by successfully designing and implementing a remote rendering system using the Direct3D API. It applies the implementation to a sample Direct3D application and also performs a few experimentations to show the technical feasibility.

Implementation of AR Remote Rendering Techniques for Real-time Volumetric 3D Video

  • Lee, Daehyeon;Lee, Munyong;Lee, Sang-ha;Lee, Jaehyun;Kwon, Soonchul
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.12 no.2
    • /
    • pp.90-97
    • /
    • 2020
  • Recently, with the growth of mixed reality industrial infrastructure, relevant convergence research has been proposed. For real-time mixed reality services such as remote video conferencing, the research on real-time acquisition-process-transfer methods is required. This paper aims to implement an AR remote rendering method of volumetric 3D video data. We have proposed and implemented two modules; one, the parsing module of the volumetric 3D video to a game engine, and two, the server rendering module. The result of the experiment showed that the volumetric 3D video sequence data of about 15 MB was compressed by 6-7%. The remote module was streamed at 27 fps at a 1200 by 1200 resolution. The results of this paper are expected to be applied to an AR cloud service.

Extracting Graphics Information for Better Video Compression

  • Hong, Kang Woon;Ryu, Won;Choi, Jun Kyun;Lim, Choong-Gyoo
    • ETRI Journal
    • /
    • v.37 no.4
    • /
    • pp.743-751
    • /
    • 2015
  • Cloud gaming services are heavily dependent on the efficiency of real-time video streaming technology owing to the limited bandwidths of wire or wireless networks through which consecutive frame images are delivered to gamers. Video compression algorithms typically take advantage of similarities among video frame images or in a single video frame image. This paper presents a method for computing and extracting both graphics information and an object's boundary from consecutive frame images of a game application. The method will allow video compression algorithms to determine the positions and sizes of similar image blocks, which in turn, will help achieve better video compression ratios. The proposed method can be easily implemented using function call interception, a programmable graphics pipeline, and off-screen rendering. It is implemented using the most widely used Direct3D API and applied to a well-known sample application to verify its feasibility and analyze its performance. The proposed method computes various kinds of graphics information with minimal overhead.

Effective Ray-tracing based Rendering Methods for Point Cloud Data in Mobile Environments (모바일 환경에서 점 구름 데이터에 대한 효과적인 광선 추적 기반 렌더링 기법)

  • Woong Seo;Youngwook Kim;Kiseo Park;Yerin Kim;Insung Ihm
    • Journal of the Korea Computer Graphics Society
    • /
    • v.29 no.3
    • /
    • pp.93-103
    • /
    • 2023
  • The problem of reconstructing three-dimensional models of people and objects from color and depth images captured by low-cost RGB-D cameras has long been an active research area in computer graphics. Color and depth images captured by low-cost RGB-D cameras are represented as point clouds in three-dimensional space, which correspond to discrete values in a continuous three-dimensional space and require additional surface reconstruction compared to rendering using polygonal models. In this paper, we propose an effective ray-tracing based technique for visualizing point clouds rather than polygonal models. In particular, our method shows the possibility of an effective rendering method even in mobile environment which has limited performance due to processor heat and lack of battery.

Development of an Integrated IaaS+PaaS Environment for Providing Cloud Computing Service in a BIM Platform for Harbor Facilities (항만 BIM 플랫폼의 클라우드 서비스를 위한 IaaS+PaaS 통합 환경 개발)

  • Moon, Hyoun-Seok;Hyun, Keun-Ju;Kim, Won-Sik
    • Journal of KIBIM
    • /
    • v.9 no.4
    • /
    • pp.62-74
    • /
    • 2019
  • Because the existing BIM platform is based on user services, the focus is on the development of SaaS (Software as a Service), which provides business services online. However, since a harbor is a security facility, the harbor BIM platform is preferably provided in a private form, rather than relying on the infrastructure environment provided by external cloud providers. Therefore, this study analyzes and reviews the main functions to be provided as SaaS services of the harbor BIM platform. The goal is to build a cloud-based harbor BIM platform that can provide this service to users. To this end, we built IaaS (Infrastructure as a Service) environment of the harbor BIM platform based on the open source Open Stack and integrate and develop PaaS environment with Open Shift applied with IaaS. We applied the GPU to the harbor BIM platform to verify the performance of the harbor BIM platform, and found that the rendering and loading times are improved. In particular, it is expected to reduce the cost of introduction and provide it as the basic cloud environment of similar BIM platform for infrastructure facilities.

Design of Mobile-based Security Agent for Contents Networking in Mixed Reality (융합현실에서 콘텐츠 네트워킹을 위한 모바일 기반 보안 중계 설계)

  • Kim, Donghyun;Lim, Jaehyun;Kim, Seoksoo
    • Journal of Convergence for Information Technology
    • /
    • v.9 no.3
    • /
    • pp.22-29
    • /
    • 2019
  • Due to the development of ICT technology, convergence reality contents are utilized as technology for providing services in various industrial fields by visualizing various information such as sensor information and shared information in a service platform showing only simple three-dimensional contents. Research is underway to reduce the weight of applications by transmitting the resources of the object to be enhanced to the network as the information and the contents to be provided increase. In order to provide resources through the network, servers for processing various information such as pattern information, content information, and sensor information must be constructed in a cloud environment. However, in order to authenticate data transmitted and received in real-time in a cloud environment, there is a problem in that the processing is delayed and a delay phenomenon occurs in the rendering process and QoS is lowered. In this paper, we propose a system to distribute cloud server which provides augmented contents of convergent reality service that provides various contents such as sensor information and three - dimensional model, and shorten the processing time of reliable data through distributed relay between servers Respectively.

3D Cloud Animation using Cloud Modeling Method of 2D Meteorological Satellite Images (2차원 기상 위성 영상의 구름 모델링 기법을 이용한 3차원 구름 애니메이션)

  • Lee, Jeong-Jin;Kang, Moon-Koo;Lee, Ho;Shin, Byeong-Seok
    • Journal of Korea Game Society
    • /
    • v.10 no.1
    • /
    • pp.147-156
    • /
    • 2010
  • In this paper, we propose 3D cloud animation by cloud modeling method of 2D images retrieved from a meteorological satellite. First, on the satellite images, we locate numerous control points to perform thin-plate spline warping analysis between consecutive frames for the modeling of cloud motion. In addition, the spectrum channels of visible and infrared wavelengths are used to determine the amount and altitude of clouds for 3D cloud image reconstruction. Pre-integrated volume rendering method is used to achieve seamless inter-laminar shades in real-time using small number of slices of the volume data. The proposed method could successfully construct continuously moving 3D clouds from 2D satellite images at an acceptable speed and image quality.

Enhancement Techniques for GPU-Based Rendering of Participating Media (GPU 기반 반투과 매체 렌더링의 향상 기법)

  • Cha, Deuk-Hyun;Yi, Yong-Il;Ihm, In-Sung
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.12
    • /
    • pp.1165-1176
    • /
    • 2010
  • In order to realistically visualize such participating media as cloud, smoke, and gas, the light transport process must be physically simulated inside the media. While it is known that this process is well described physically through the volume rendering equation, it usually takes a great deal of computation time for obtaining high-precision solutions. Recently, GPU-based, fast rendering methods have been proposed for the realistic simulation of participating media, however, there still remain several problems to be resolved. In this article, we describe our rendering techniques applied to enhance the performances and features of our GPU-assisted participating media renderer, and analyze how such efforts have actually improved the renderer. The presented techniques will be effectively used in volume renderers for creating various digital contents in the special effects industries.