• Title/Summary/Keyword: Rendering quality

Search Result 247, Processing Time 0.03 seconds

An Efficient Perspective Projection using $\textrm{VolumePro}^{TM}$ Hardware (볼륨프로 하드웨어를 이용한 효율적인 투시투영 방법)

  • 임석현;신병석
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.31 no.3_4
    • /
    • pp.195-203
    • /
    • 2004
  • VolumePro is a real-time volume rendering hardware for consumer PCs. However it cannot be used for the applications requiring perspective projection such as virtual endoscopy since it provides only orthographic projection. Several methods have been presented to approximate perspective projection by decomposing a volume into slabs and applying successive parallel projection to thou. But it takes a lot of time since the entire region of every slab should be processed, which does not contribute to final image. In this paper, we propose an efficient perspective projection method that makes the use of several sub-volumes with cropping feature of VolumePro. It reduces the rendering time in comparison to slab-based method without image quality deterioration since it processes only the parts contained in the view frustum.

Real-time BCC Volume Isosurface Ray Casting on the GPU (GPU를 이용한 실시간 BCC 볼륨 등가면 레이 캐스팅)

  • Kim, Minho;Lee, Young-Joon
    • Journal of the Korea Computer Graphics Society
    • /
    • v.18 no.4
    • /
    • pp.25-34
    • /
    • 2012
  • This paper presents a real-time GPU (graphics processing unit) ray casting scheme for rendering isosurfaces of BCC (body-centered cubic) volume datasets. A quartic spline field is built using the 7-direction box-spline filter accompanied with a quasi-interpolation prefilter. To obtain an interactive rendering speed on the graphics hardware, the shader code was optimized to avoid lookup table and conditional branches and to minimize data fetch overhead. Compared to previous implementations, our work outperforms the comparable one by more than 20% and the rendering quality is superior than others.

A Study on Delta Image Composition Methods of the Depth-Image-Based Rendering for the Generation of Stereoscopic Images on Mobile Devices (모바일 장치에서 입체 영상 생성을 위한 깊이 영상 기반 렌더링의 부가 정보 영상 구성 방법에 관한 연구)

  • Kim, Min-Young;Park, Kyoung-Shin;Choo, Hyon-Gon;Kim, Jin-Woong;Cho, Yong-Joo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.16 no.7
    • /
    • pp.1428-1436
    • /
    • 2012
  • This paper presents the delta image composition methods using Depth-Image-Based Rendering (DIBR) for 3D stereoscopic broadcasting for the low bandwidth mobile DMB broadcasting system. With DIBR, a left and depth images are transmitted to a mobile device, which restores the right view, whose quality may be poor. This paper describes delta image composition methods for the restoration while minimizing the amount of the transmitted data.

Real-time 3D Audio Downmixing System based on Sound Rendering for the Immersive Sound of Mobile Virtual Reality Applications

  • Hong, Dukki;Kwon, Hyuck-Joo;Kim, Cheong Ghil;Park, Woo-Chan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.12
    • /
    • pp.5936-5954
    • /
    • 2018
  • Eight out of the top ten the largest technology companies in the world are involved in some way with the coming mobile VR revolution since Facebook acquired Oculus. This trend has allowed the technology related with mobile VR to achieve remarkable growth in both academic and industry. Therefore, the importance of reproducing the acoustic expression for users to experience more realistic is increasing because auditory cues can enhance the perception of the complicated surrounding environment without the visual system in VR. This paper presents a audio downmixing system for auralization based on hardware, a stage of sound rendering pipelines that can reproduce realiy-like sound but requires high computation costs. The proposed system is verified through an FPGA platform with the special focus on hardware architectural designs for low power and real-time. The results show that the proposed system on an FPGA can downmix maximum 5 sources in real-time rate (52 FPS), with 382 mW low power consumptions. Furthermore, the generated 3D sound with the proposed system was verified with satisfactory results of sound quality via the user evaluation.

Voxel-wise UV parameterization and view-dependent texture synthesis for immersive rendering of truncated signed distance field scene model

  • Kim, Soowoong;Kang, Jungwon
    • ETRI Journal
    • /
    • v.44 no.1
    • /
    • pp.51-61
    • /
    • 2022
  • In this paper, we introduced a novel voxel-wise UV parameterization and view-dependent texture synthesis for the immersive rendering of a truncated signed distance field (TSDF) scene model. The proposed UV parameterization delegates a precomputed UV map to each voxel using the UV map lookup table and consequently, enabling efficient and high-quality texture mapping without a complex process. By leveraging the convenient UV parameterization, our view-dependent texture synthesis method extracts a set of local texture maps for each voxel from the multiview color images and separates them into a single view-independent diffuse map and a set of weight coefficients for an orthogonal specular map basis. Furthermore, the view-dependent specular maps for an arbitrary view are estimated by combining the specular weights of each source view using the location of the arbitrary and source viewpoints to generate the view-dependent textures for arbitrary views. The experimental results demonstrate that the proposed method effectively synthesizes texture for an arbitrary view, thereby enabling the visualization of view-dependent effects, such as specularity and mirror reflection.

Rendering Quality Improvement Method based on Depth and Inverse Warping (깊이정보와 역변환 기반의 포인트 클라우드 렌더링 품질 향상 방법)

  • Lee, Heejea;Yun, Junyoung;Park, Jong-Il
    • Journal of Broadcast Engineering
    • /
    • v.26 no.6
    • /
    • pp.714-724
    • /
    • 2021
  • The point cloud content is immersive content recorded by acquiring points and colors corresponding to the real environment and objects having three-dimensional location information. When a point cloud content consisting of three-dimensional points having position and color information is enlarged and rendered, the gap between the points widens and an empty hole occurs. In this paper, we propose a method for improving the quality of point cloud contents through inverse transformation-based interpolation using depth information for holes by finding holes that occur due to the gap between points when expanding the point cloud. The points on the back are rendered between the holes created by the gap between the points, acting as a hindrance to applying the interpolation method. To solve this, remove the points corresponding to the back side of the point cloud. Next, a depth map at the point in time when an empty hole is generated is extracted. Finally, inverse transform is performed to extract pixels from the original data. As a result of rendering content by the proposed method, the rendering quality improved by 1.2 dB in terms of average PSNR compared to the conventional method of increasing the size to fill the blank area.

An Efficient Anisotropic Volume Rendering using an Intensity Interpolation and Adaptive Intermediate Voxel Insertion Method (광도 보간과 적응형 중간복셀 삽입법을 이용한 효율적인 비균등 볼륨 렌더링)

  • 김태영;신병석
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.30 no.11
    • /
    • pp.655-662
    • /
    • 2003
  • In some volume visualization fields such as medical imaging, anisotropic volume data are more common than isotropic ones. In this paper, we propose an efficient rendering method for anisotropic volume data, which directly computes the intensity of intermediate samples by interpolating the intensity of two corresponding voxels on consecutive slices. Unlike density interpolation method, it does not require a preprocessing step for generating intermediate slices or additional memory for storing them. Additionally, we propose an adaptive intermediate voxel insertion method that avoids overblurring on object surfaces. This may occur when we render high frequency areas using the intensity interpolation method. Using these methods, we can improve the rendering speed without sacrificing image quality.

Near-lossless Coding of Multiview Texture and Depth Information for Graphics Applications (그래픽스 응용을 위한 다시점 텍스처 및 깊이 정보의 근접 무손실 부호화)

  • Yoon, Seung-Uk;Ho, Yo-Sung
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.46 no.1
    • /
    • pp.41-48
    • /
    • 2009
  • This Paper introduces representation and coding schemes of multiview texture and depth data for complex three-dimensional scenes. We represent input color and depth images using compressed texture and depth map pairs. The proposed X-codec encodes them further to increase compression ratio in a near-lossless way. Our system resolves two problems. First, rendering time and output visual quality depend on input image resolutions rather than scene complexity since a depth image-based rendering techniques is used. Second, the random access problem of conventional image-based rendering could be effectively solved using our image block-based compression schemes. From experimental results, the proposed approach is useful to graphics applications because it provides multiview rendering, selective decoding, and scene manipulation functionalities.

Real-time Rendering of Realistic Grasses Using Fractal and Shader-Instancing (프랙탈과 셰이더 인스턴싱 기법을 이용한 자연스러운 잔디의 실시간 렌더링)

  • Kim, Jin-Mo;Cho, Hyung-Je
    • Journal of Korea Multimedia Society
    • /
    • v.13 no.2
    • /
    • pp.298-307
    • /
    • 2010
  • The grass is one of important components that cover the wide surfaces in the application such as game or real time simulation. Actually, it not easy to render effectively numerous grasses that grow over the wide terrain. To solve the difficulty, we must find a solution to the two contradictions in terms : quality and calculation cost. As a solution to the above-mentioned task, in this paper, we propose an efficient method to represent the natural grasses by introducing fractal theory and instancing technique. Although the existing grass representation methods make use of a simple rule of applying a basic grass model repeatedly in rendering process, on the contrary we take advantage of the basic property of fractal's self-similarity and we devise a natural representation method suited to the given environment by introducing two important growth factors such as nature of terrain and quantity of light, and finally we apply a GPU-based shader instancing technique to rendering numerous grass models in real-time.

Effective Image Sequence Format in 3D Animation Production Pipeline (3D 애니메이션 제작 공정에 있어서 효율적인 이미지 시퀀스 포맷)

  • Kim, Ho
    • The Journal of the Korea Contents Association
    • /
    • v.7 no.8
    • /
    • pp.134-141
    • /
    • 2007
  • In 3D animation rendering process, Although we can render the output as a movie file format, most productions use image sequences in their rendering pipelines. This Image Sequence rendering process is extremely important step in final compositing in movie industries. Although there are various type of making image rendering processes, TGA format Is one of most widely used bitmap file formats using in industries. People may ask TGA format is most suitable for in any case. As we know 3D softwares have their own image formats. so we need to testify on this. In this paper, we are going to focus on Alias' 3D package software called MAYA which we will analyze of compressing image sequence, Image quality, supporting Alpha channels in compositing, and Z-depth information. The purpose of this paper is providing to 3D Pipeline as a guideline about effective image sequence format.