• Title/Summary/Keyword: interactive rendering

Search Result 100, Processing Time 0.025 seconds

Enhanced Image Mapping Method for Computer-Generated Integral Imaging System (집적 영상 시스템을 위한 향상된 이미지 매핑 방법)

  • Lee Bin-Na-Ra;Cho Yong-Joo;Park Kyoung-Shin;Min Sung-Wook
    • The KIPS Transactions:PartB
    • /
    • v.13B no.3 s.106
    • /
    • pp.295-300
    • /
    • 2006
  • The integral imaging system is an auto-stereoscopic display that allows users to see 3D images without wearing special glasses. In the integral imaging system, the 3D object information is taken from several view points and stored as elemental images. Then, users can see a 3D reconstructed image by the elemental images displayed through a lens array. The elemental images can be created by computer graphics, which is referred to the computer-generated integral imaging. The process of creating the elemental images is called image mapping. There are some image mapping methods proposed in the past, such as PRR(Point Retracing Rendering), MVR(Multi-Viewpoint Rendering) and PGR(Parallel Group Rendering). However, they have problems with heavy rendering computations or performance barrier as the number of elemental lenses in the lens array increases. Thus, it is difficult to use them in real-time graphics applications, such as virtual reality or real-time, interactive games. In this paper, we propose a new image mapping method named VVR(Viewpoint Vector Rendering) that improves real-time rendering performance. This paper describes the concept of VVR first and the performance comparison of image mapping process with previous methods. Then, it discusses possible directions for the future improvements.

Real-Time Hierarchical Techniques for Rendering of Translucent Materials and Screen-Space Interpolation (반투명 재질의 렌더링과 화면 보간을 위한 실시간 계층화 알고리즘)

  • Ki, Hyun-Woo;Oh, Kyoung-Su
    • Journal of Korea Game Society
    • /
    • v.7 no.1
    • /
    • pp.31-42
    • /
    • 2007
  • In the natural world, most materials such as skin, marble and cloth are translucent. Their appearance is smooth and soft compared with metals or mirrors. In this paper, we propose a new GPU based hierarchical rendering technique for translucent materials, based on the dipole diffusion approximation, at interactive rates. Information of incident light, position, normal, and irradiance, on the surfaces are stored into 2D textures by rendering from a primary light view. Huge numbers of pixel photons are clustered into quad-tree image pyramids. Each pixel, we select clusters (sets of photons), and then we approximate multiple subsurface scattering term with the clusters. We also introduce a novel hierarchical screen-space interpolation technique by exploiting spatial coherence with early-z culling on the GPU. We also build image pyramids of the screen using mipmap and pixel shader. Each pixel of the pyramids is stores position, normal and spatial similarity of children pixels. If a pixel's the similarity is high, we render the pixel and interpolate the pixel to multiple pixels. Result images show that our method can interactively render deformable translucent objects by approximating hundreds of thousand photons with only hundreds clusters without any preprocessing. We use an image-space approach for entire process on the GPU, thus our method is less dependent to scene complexity.

  • PDF

Realistic and Fast Depth-of-Field Rendering in Direct Volume Rendering (직접 볼륨 렌더링에서 사실적인 고속 피사계 심도 렌더링)

  • Kang, Jiseon;Lee, Jeongjin;Shin, Yeong-Gil;Kim, Bohyoung
    • The Journal of Korean Institute of Next Generation Computing
    • /
    • v.15 no.5
    • /
    • pp.75-83
    • /
    • 2019
  • Direct volume rendering is a widely used method for visualizing three-dimensional volume data such as medical images. This paper proposes a method for applying depth-of-field effects to volume ray-casting to enable more realistic depth-of-filed rendering in direct volume rendering. The proposed method exploits a camera model based on the human perceptual model and can obtain realistic images with a limited number of rays using jittered lens sampling. It also enables interactive exploration of volume data by on-the-fly calculating depth-of-field in the GPU pipeline without preprocessing. In the experiment with various data including medical images, we demonstrated that depth-of-field images with better depth perception were generated 2.6 to 4 times faster than the conventional method.

Real-time BCC Volume Isosurface Ray Casting on the GPU (GPU를 이용한 실시간 BCC 볼륨 등가면 레이 캐스팅)

  • Kim, Minho;Lee, Young-Joon
    • Journal of the Korea Computer Graphics Society
    • /
    • v.18 no.4
    • /
    • pp.25-34
    • /
    • 2012
  • This paper presents a real-time GPU (graphics processing unit) ray casting scheme for rendering isosurfaces of BCC (body-centered cubic) volume datasets. A quartic spline field is built using the 7-direction box-spline filter accompanied with a quasi-interpolation prefilter. To obtain an interactive rendering speed on the graphics hardware, the shader code was optimized to avoid lookup table and conditional branches and to minimize data fetch overhead. Compared to previous implementations, our work outperforms the comparable one by more than 20% and the rendering quality is superior than others.

Development of 3D Visualization Technology for Meteorological Data (기상자료 3차원 가시화 기술개발 연구)

  • Seo In Bum;Joh Min Su;Yun Ja Young
    • Journal of the Korean Society of Visualization
    • /
    • v.1 no.2
    • /
    • pp.58-70
    • /
    • 2003
  • Meteorological data contains observation and numerical weather prediction model output data. The computerized analysis and visualization of meteorological data often requires very high computing capability due to the large size and complex structure of the data. Because the meteorological data is frequently formed in multi-variables, 3-dimensional and time-series form, it is very important to visualize and analyze the data in 3D spatial domain in order to get more understanding about the meteorological phenomena. In this research, we developed interactive 3-dimensional visualization techniques for visualizing meteorological data on a PC environment such as volume rendering, iso-surface rendering or stream line. The visualization techniques developed in this research are expected to be effectively used as basic technologies not only for deeper understanding and more exact prediction about meteorological environments but also for scientific and spatial data visualization research in any field from which three dimensional data comes out such as oceanography, earth science, and aeronautical engineering.

  • PDF

Fog Rendering Using Distance-Altitude Scattering Model on 2D Images

  • Lee, Ho-Chang;Jang, Jaeni;Yoon, Kyung-Hyun
    • Journal of Korea Multimedia Society
    • /
    • v.14 no.12
    • /
    • pp.1528-1535
    • /
    • 2011
  • We present a fog generation algorithm in 2D images. The proposed algorithm provides a scattering model for the approximated calculation of fog density. The scattering model needs parameters of distance and altitude information. However, 2D images do not include that information, so that we calculate them from the depth information generated in an interactive manner, and estimate the scattering factor by using the scattering model. Then we generate fog effect on an input image using the scattering factor by distance-oriented selection blur and color blending. With the algorithm, we can easily create the fog-effected images and fog generated animation from 2D images.

An Efficient Algorithm for Real-Time 3D Terrain Walkthrough

  • Hesse, Michael;Gavrilova, Marina L.
    • International Journal of CAD/CAM
    • /
    • v.3 no.1_2
    • /
    • pp.111-117
    • /
    • 2003
  • The paper presents an efficient algorithm based on ROAM for visualization of large scale terrain models in real-time. The quality and smoothness of the terrain data visualization within a 3D interactive environment is preserved, while the complexity of the algorithm is kept on a reasonable level. The main contribution of the paper is an introduction of a number of efficient techniques such as implicit coordinates method within the patch array representing ROAM and the viewpoint dependent triangle rendering method for dynamic level of detail (LOD) updates. In addition, the paper presents experimental comparison of a variety of culling techniques, including a newly introduced method: relational position culling. These techniques are incorporated in the visualization software, which allows to achieve more realistic terrain representation and the real-time level of detail reduction.

A Data Structure for Real-time Volume Ray Casting (실시간 볼륨 광선 투사법을 위한 자료구조)

  • Lim, Suk-Hyun;Shin, Byeong-Seok
    • Journal of the Korea Computer Graphics Society
    • /
    • v.11 no.1
    • /
    • pp.40-49
    • /
    • 2005
  • Several optimization techniques have been proposed for volume ray casting, but these cannot achieve real-time frame rates. In addition, it is difficult to apply them to some applications that require perspective projection. Recently, hardware-based methods using 3D texture mapping are being used for real-time volume rendering. Although rendering speed approaches real time, the larger volumes require more swapping of volume bricks for the limited texture memory. Also, image quality deteriorates compared with that of conventional volume ray casting. In this paper, we propose a data structure for real-time volume ray casting named PERM (Precomputed dEnsity and gRadient Map). The PERM stores interpolated density and gradient vector for quantized cells. Since the information requiring time-consuming computations is stored in the PERM, our method can ensure interactive frame rates on a consumer PC platform. Our method normally produces high-quality images because it is based on conventional volume ray casting.

  • PDF

Panorama Field Rendering based on Depth Estimation (깊이 추정에 기반한 파노라마 필드 렌더링)

  • Jung, Myoungsook;Han, JungHyun
    • Journal of the Korea Computer Graphics Society
    • /
    • v.6 no.4
    • /
    • pp.15-22
    • /
    • 2000
  • One of the main research trends in image based modeling and rendering is how to implement plenoptic function. For this purpose, this paper proposes a novel approach based on a set of randomly placed panoramas. The proposed approach, first of all, adopts a simple computer vision technique to approximate omni-directional depth information of the surrounding scene, and then corrects/interpolates panorama images to generate an output image at a vantage viewpoint. Implementation results show that the proposed approach achieves smooth navigation at an interactive rate.

  • PDF

A Study on Real-time Graphic Workflow For Achieving The Photorealistic Virtual Influencer

  • Haitao Jiang
    • International journal of advanced smart convergence
    • /
    • v.12 no.1
    • /
    • pp.130-139
    • /
    • 2023
  • With the increasing popularity of computer-generated virtual influencers, the trend is rising especially on social media. Famous virtual influencer characters Lil Miquela and Imma were all created by CGI graphics workflows. The process is typically a linear affair. Iteration is challenging and costly. Development efforts are frequently siloed off from one another. Moreover, it does not provide a real-time interactive experience. In the previous study, a real-time graphic workflow was proposed for the Digital Actor Hologram project while the output graphic quality is less than the results obtained from the CGI graphic workflow. Therefore, a real-time engine graphic workflow for Virtual Influencers is proposed in this paper to facilitate the creation of real-time interactive functions and realistic graphic quality. The real-time graphic workflow is obtained from four processes: Facial Modeling, Facial Texture, Material Shader, and Look-Development. The analysis of performance with real-time graphical workflow for Digital Actor Hologram demonstrates the usefulness of this research result. Our research will be efficient in producing virtual influencers.