• Title/Summary/Keyword: Real-time 3D graphics

Search Result 139, Processing Time 0.02 seconds

A study on the Performance of Hybrid Normal Mapping Techniques for Real-time Rendering

  • ZhengRan Liu;KiHong Kim;YuanZi Sang
    • International journal of advanced smart convergence
    • /
    • v.12 no.4
    • /
    • pp.361-369
    • /
    • 2023
  • Achieving realistic visual quality while maintaining optimal real-time rendering performance is a major challenge in evolving computer graphics and interactive 3D applications. Normal mapping, as a core technology in 3D, has matured through continuous optimization and iteration. Hybrid normal mapping as a new mapping model has also made significant progress and has been applied in the 3D asset production pipeline. This study comprehensively explores the hybrid normal techniques, analyzing Linear Blending, Overlay Blending, Whiteout Blending, UDN Blending, and Reoriented Normal Mapping, and focuses on how the various hybrid normal techniques can be used to achieve rendering performance and visual fidelity. performance and visual fidelity. Under the consideration of computational efficiency, visual coherence, and adaptability in different 3D production scenes, we design comparative experiments to explore the optimal solutions of the hybrid normal techniques by analyzing and researching the code, the performance of different hybrid normal mapping in the engine, and analyzing and comparing the data. The purpose of the research and summary of the hybrid normal technology is to find out the most suitable choice for the mainstream workflow based on the objective reality. Provide an understanding of the hybrid normal mapping technique, so that practitioners can choose how to apply different hybrid normal techniques to the corresponding projects. The purpose of our research and summary of mixed normal technology is to find the most suitable choice for mainstream workflows based on objective reality. We summarized the hybrid normal mapping technology and experimentally obtained the advantages and disadvantages of different technologies, so that practitioners can choose to apply different hybrid normal mapping technologies to corresponding projects in a reasonable manner.

DESIGN AND IMPLEMENTATION OF FEATURE-BASED 3D GEO-SPATIAL RENDERING SYSTEM USING OPENGL API

  • Kim Seung-Yeb;Lee Kiwon
    • Proceedings of the KSRS Conference
    • /
    • 2005.10a
    • /
    • pp.321-324
    • /
    • 2005
  • In these days, the management and visualization of 3D geo-spatial information is regarded as one of an important issue in GiS and remote sensing fields. 3D GIS is considered with the database issues such as handling and managing of 3D geometry/topology attributes, whereas 3D visualization is basically concerned with 3D computer graphics. This study focused on the design and implementation for the OpenGL API-based rendering system for the complex types of 3D geo-spatial features. In this approach 3D features can be separately processed with the functions of authoring and manipulation of terrain segments, building segments, road segments, and other geo-based things with texture mapping. Using this implementation, it is possible to the generation of an integrated scene with these complex types of 3D features. This integrated rendering system based on the feature-based 3D-GIS model can be extended and effectively applied to urban environment analysis, 3D virtual simulation and fly-by navigation in urban planning. Furthermore, we expect that 3D-GIS visualization application based on OpenGL API can be easily extended into a real-time mobile 3D-GIS system, soon after the release of OpenGLIES which stands for OpenGL for embedded system, though this topic is beyond the scope of this implementation.

  • PDF

Technology Trends for Motion Synthesis and Control of 3D Character

  • Choi, Jong-In
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.4
    • /
    • pp.19-26
    • /
    • 2019
  • In this study, we study the development and control of motion of 3D character animation and discuss the development direction of technology. Character animation has been developed as a data-based method and a physics-based method. The animation generation technique based on the keyframe method has been made possible by the development of the hardware technology, and the motion capture device has been used. Various techniques for effectively editing the motion data have appeared. At the same time, animation techniques based on physics have emerged, which realistically generate the motion of the character by physically optimized numerical computation. Recently, animation techniques using machine learning have shown new possibilities for creating characters that can be controlled by the user in real time and are expected to be developed in the future.

Oriental Black Ink Rendering System (수묵화 렌더링 시스템)

  • Jeong, Kyu-Man;Lee, Seung-Yong
    • Journal of the Korea Computer Graphics Society
    • /
    • v.15 no.2
    • /
    • pp.19-34
    • /
    • 2009
  • This paper presents an automatic rendering system for 3D meshes which generates images in the style of oriental black ink painting. In this paper, we analyze the characteristics of black ink painting and present some requirements for our black ink rendering system. Then we survey the steps used by artists to create black ink paintings. Based on the analysis of drawing steps, we propose a black ink rendering system which reproduces the key features of black ink painting through three sub-systems: the feature line layer, the interior shading layer, and the media layer. The main contribution of this paper is to introduce new methods which generate realistic and natural results very fast while any simulation methods are not used to satisfy real-time performance. We describe our implementation of each of these, and demonstrate our results in the major three styles of black ink painting: the outline style, the spontaneous style, and the outline-spontaneous style. Our system renders 3D models of moderate complexity at interactive frame rates. As a result, we expect that our system can be directly applied to real-time applications such as computer games and virtual environments.

  • PDF

Implementation of Virtual Architectural Engineering and Design of Real-time State Server for Internet Virtual Collaboration (가상건축엔지니어링의 구현과 인터넷 가상협동작업을 위한 실시간 상태서버의 설계)

  • 고동일;이범렬;김종성;오원근
    • Proceedings of the IEEK Conference
    • /
    • 2000.11c
    • /
    • pp.169-172
    • /
    • 2000
  • Recently, the advent of World-Wide-Web(WWW) and the explosive popularity of the Internet gave birth to collaborative applications which were enabled by computers and networks as their primary media. And the progress of 3D computer graphics enabled collaborative application with 3D virtual environments or distributed virtual environments. In this paper, we explain our implementation of Share collaboration engine and Virtual Architectural Engineering 2000 (VAE2000) system which is our pilot application implemented with Share collaboration engine. And we explain problems presented by our experiments with VAE2000 system. For those problems, we design our new middle-ware system, SHINE(SHared INternet Environment). The SHINE proposes new concepts and approaches for collaboration with 3D objects in a virtual world.

  • PDF

Image-based Collision Detection on GPU (GPU를 이용한 이미지 기반 충돌검사)

  • Jang, Han-Young;Jung, Taek-Sang;Han, Jung-Hyun
    • 한국HCI학회:학술대회논문집
    • /
    • 2006.02a
    • /
    • pp.812-817
    • /
    • 2006
  • This paper presents an image-space algorithm to real-time collision detection, which is run completely by GPU. For a single object or for multiple objects with no collision, the front and back faces appear alternately along the view direction. However, such alternation is violated when objects collide. Based on these observations, the algorithm has been devised, and the implementation utilizes the state-of-the-art functionalities of GPU such as framebuffer objects(FBO), vertex buffer object(VBO) and occlusion query. The experimental results show the feasibility of GPU-intensive collision detection and its performance gain in real-time applications such as 3D games.

  • PDF

Construction of Virtual Environment for a Vehicle Simulator (자동차 시뮬레이터의 가상환경 구성에 대한 연구)

  • Chang, Chea-Won;Son, Kwon;Choi, Kyung-Hyun
    • Transactions of the Korean Society of Automotive Engineers
    • /
    • v.8 no.4
    • /
    • pp.158-168
    • /
    • 2000
  • Vehicle driving simulators can provide engineers with benefits on the development and modification of vehicle models. One of the most important factors to realistic simulations is the fidelity given by a motion system and a real-time visual image generation system. Virtual reality technology has been widely used to achieve high fidelity. In this paper the virtual environment including a visual system like a head-mounted display is developed for a vehicle driving simulator system by employing the virtual reality technique. virtual vehicle and environment models are constructed using the object-oriented analysis and design approach. Accordint to the object model a three dimensional graphic model is developed with CAD tools such as Rhino and Pro/E. For the real-time image generation the optimized IRIS Performer 3D graphics library is embedded with the multi-thread methodology. Compared with the single loop apprach the proposed methodology yields an acceptable image generation speed 20 frames/sec for the simulator.

  • PDF

An Approximation Technique for Real-time Rendering of Phong Reflection Model with Image-based Lighting (영상 기반 조명을 적용한 퐁 반사 모델의 실시간 렌더링을위한 근사 기법)

  • Jeong, Taehong;Shin, Hyun Joon
    • Journal of the Korea Computer Graphics Society
    • /
    • v.20 no.1
    • /
    • pp.13-19
    • /
    • 2014
  • In this paper, we introduce a real-time method to render a 3D scene using image-based lighting. Previous approaches for image-based lighting focused on diffuse reflection and mirror-like specular reflection. In this paper, we provide a simple preprocessing approach to efficiently approximate Phong reflection model, which has been used for computer graphics applications for several decades. For diffuse reflection, we generate a texture map for diffuse reflection by integrating the source image in preprocessing step, similarly to the previous approaches. We adopt the similar idea to produce a set of specular reflection maps for various material shininess. By doing this, we can render a dynamic scene without high computational complexity or numerous texture map access.

Video Augmentation by Image-based Rendering

  • Seo, Yong-Duek;Kim, Seung-Jin;Sang, Hong-Ki
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1998.06b
    • /
    • pp.147-153
    • /
    • 1998
  • This paper provides a method for video augmentation using image interpolation. In computer graphics or augmented reality, 3D information of a model object is necessary to generate 2D views of the model, which are then inserted into or overlayed on environmental views or real video frames. However, we do not require any three dimensional model but images of the model object at some locations to render views according to the motion of video camera which is calculated by an SFM algorithm using point matches under weak-perspective (scaled-orthographic) projection model. Thus, a linear view interpolation algorithm is applied rather than a 3D ray-tracing method to get a view of the model at different viewpoints from model views. In order to get novel views in a way that agrees with the camera motion the camera coordinate system is embedded into model coordinate system at initialization time on the basis of 3D information recovered from video images and model views, respectively. During the sequence, motion parameters from video frames are used to compute interpolation parameters, and rendered model views are overlayed on corresponding video frames. Experimental results for real video frames and model views are given. Finally, discussion on the limitations of the method and subjects for future research are provided.

  • PDF

MPEG-I RVS Software Speed-up for Real-time Application (실시간 렌더링을 위한 MPEG-I RVS 가속화 기법)

  • Ahn, Heejune;Lee, Myeong-jin
    • Journal of Broadcast Engineering
    • /
    • v.25 no.5
    • /
    • pp.655-664
    • /
    • 2020
  • Free viewpoint image synthesis technology is one of the important technologies in the MPEG-I (Immersive) standard. RVS (Reference View Synthesizer) developed by MPEG-I and in use in MPEG group is a DIBR (Depth Information-Based Rendering) program that generates an image at a virtual (intermediate) viewpoint from multiple viewpoints' inputs. RVS uses the mesh surface method based on computer graphics, and outperforms the pixel-based ones by 2.5dB or more compared to the previous pixel method. Even though its OpenGL version provides 10 times speed up over the non OpenGL based one, it still shows a non-real-time processing speed, i.e., 0.75 fps on the two 2k resolution input images. In this paper, we analyze the internal of RVS implementation and modify its structure, achieving 34 times speed up, therefore, real-time performance (22-26 fps), through the 3 key improvements: 1) the reuse of OpenGL buffers and texture objects 2) the parallelization of file I/O and OpenGL execution 3) the parallelization of GPU shader program and buffer transfer.