• Title/Summary/Keyword: Point-based rendering

Search Result 73, Processing Time 0.026 seconds

Pose Transformation of a Frontal Face Image by Invertible Meshwarp Algorithm (역전가능 메쉬워프 알고리즘에 의한 정면 얼굴 영상의 포즈 변형)

  • 오승택;전병환
    • Journal of KIISE:Software and Applications
    • /
    • v.30 no.1_2
    • /
    • pp.153-163
    • /
    • 2003
  • In this paper, we propose a new technique of image based rendering(IBR) for the pose transformation of a face by using only a frontal face image and its mesh without a three-dimensional model. To substitute the 3D geometric model, first, we make up a standard mesh set of a certain person for several face sides ; front. left, right, half-left and half-right sides. For the given person, we compose only the frontal mesh of the frontal face image to be transformed. The other mesh is automatically generated based on the standard mesh set. And then, the frontal face image is geometrically transformed to give different view by using Invertible Meshwarp Algorithm, which is improved to tolerate the overlap or inversion of neighbor vertexes in the mesh. The same warping algorithm is used to generate the opening or closing effect of both eyes and a mouth. To evaluate the transformation performance, we capture dynamic images from 10 persons rotating their heads horizontally. And we measure the location error of 14 main features between the corresponding original and transformed facial images. That is, the average difference is calculated between the distances from the center of both eyes to each feature point for the corresponding original and transformed images. As a result, the average error in feature location is about 7.0% of the distance from the center of both eyes to the center of a mouth.

Real-Time Indirect Illumination using a Light Quad-Tree (광원 트리를 사용한 간접 조명의 실시간 렌더링)

  • Ki, Hyun-Woo;Oh, Kyoung-Su
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.34 no.4
    • /
    • pp.158-167
    • /
    • 2007
  • Indirect illumination plays an important role for realistic image synthesis. We present a novel realtime indirect illumination rendering technique using image pyramids. Hundreds of thousands of indirect point light sources are stored into images, and then they hierarchically clustered into quad-tree image pyramids. We also introduce a GPU based top-down and breadth-first traversal of the quad-trees to approximate the illumination with clusters (set of lights). All steps entirely run on the GPU in real-time. Result images demonstrate that our method represents diffuse interreflection, especially a color bleeding effect well. We achieved interactive frame rates of tens to hundreads, without any preprocessing. We can avoid artifacts caused by sampling, and our method is seven times faster than a recently proposed sampling based method.

A 3D Memory System Allowing Multi-Access (다중접근을 허용하는 3차원 메모리 시스템)

  • 이형
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.32 no.9
    • /
    • pp.457-464
    • /
    • 2005
  • In this paper a 3D memory system that allows 17 access types at an arbitrary position is introduced. The proposed memory system is based on two main functions: memory module assignment function and address assignment function. Based on them, the memory system supports 17 access types: 13 Lines, 3 Rectangles, and 1 Hexahedron. That is, the memory system allows simultaneous access to multiple data in any access types at an arbitrary position with a constant interval. In order to allow 17 access types the memory system consists of memory module selection circuitry, data routing circuitry for READ/WRITE, and address calculation/routing circuitry In the point of view of a developer and a programmer, the memory system proposed in this paper supports easy hardware extension according to the applications and both of them to deal with it as a logical three-dimensional away In addition, multiple data in various across types can be simultaneously accessed with a constant interval. Therefore, the memory system is suitable for building systems related to ,3D applications (e.g. volume rendering and volume clipping) and a frame buffer for multi-resolution.

3D Reconstruction of Structure Fusion-Based on UAS and Terrestrial LiDAR (UAS 및 지상 LiDAR 융합기반 건축물의 3D 재현)

  • Han, Seung-Hee;Kang, Joon-Oh;Oh, Seong-Jong;Lee, Yong-Chang
    • Journal of Urban Science
    • /
    • v.7 no.2
    • /
    • pp.53-60
    • /
    • 2018
  • Digital Twin is a technology that creates a photocopy of real-world objects on a computer and analyzes the past and present operational status by fusing the structure, context, and operation of various physical systems with property information, and predicts the future society's countermeasures. In particular, 3D rendering technology (UAS, LiDAR, GNSS, etc.) is a core technology in digital twin. so, the research and application are actively performed in the industry in recent years. However, UAS (Unmanned Aerial System) and LiDAR (Light Detection And Ranging) have to be solved by compensating blind spot which is not reconstructed according to the object shape. In addition, the terrestrial LiDAR can acquire the point cloud of the object more precisely and quickly at a short distance, but a blind spot is generated at the upper part of the object, thereby imposing restrictions on the forward digital twin modeling. The UAS is capable of modeling a specific range of objects with high accuracy by using high resolution images at low altitudes, and has the advantage of generating a high density point group based on SfM (Structure-from-Motion) image analysis technology. However, It is relatively far from the target LiDAR than the terrestrial LiDAR, and it takes time to analyze the image. In particular, it is necessary to reduce the accuracy of the side part and compensate the blind spot. By re-optimizing it after fusion with UAS and Terrestrial LiDAR, the residual error of each modeling method was compensated and the mutual correction result was obtained. The accuracy of fusion-based 3D model is less than 1cm and it is expected to be useful for digital twin construction.

GPU-based dynamic point light particles rendering using 3D textures for real-time rendering (실시간 렌더링 환경에서의 3D 텍스처를 활용한 GPU 기반 동적 포인트 라이트 파티클 구현)

  • Kim, Byeong Jin;Lee, Taek Hee
    • Journal of the Korea Computer Graphics Society
    • /
    • v.26 no.3
    • /
    • pp.123-131
    • /
    • 2020
  • This study proposes a real-time rendering algorithm for lighting when each of more than 100,000 moving particles exists as a light source. Two 3D textures are used to dynamically determine the range of influence of each light, and the first 3D texture has light color and the second 3D texture has light direction information. Each frame goes through two steps. The first step is to update the particle information required for 3D texture initialization and rendering based on the Compute shader. Convert the particle position to the sampling coordinates of the 3D texture, and based on this coordinate, update the colour sum of the particle lights affecting the corresponding voxels for the first 3D texture and the sum of the directional vectors from the corresponding voxels to the particle lights for the second 3D texture. The second stage operates on a general rendering pipeline. Based on the polygon world position to be rendered first, the exact sampling coordinates of the 3D texture updated in the first step are calculated. Since the sample coordinates correspond 1:1 to the size of the 3D texture and the size of the game world, use the world coordinates of the pixel as the sampling coordinates. Lighting process is carried out based on the color of the sampled pixel and the direction vector of the light. The 3D texture corresponds 1:1 to the actual game world and assumes a minimum unit of 1m, but in areas smaller than 1m, problems such as stairs caused by resolution restrictions occur. Interpolation and super sampling are performed during texture sampling to improve these problems. Measurements of the time taken to render a frame showed that 146 ms was spent on the forward lighting pipeline, 46 ms on the defered lighting pipeline when the number of particles was 262144, and 214 ms on the forward lighting pipeline and 104 ms on the deferred lighting pipeline when the number of particle lights was 1,024766.

Spectral-based rendering technique for lighting simulation (라이팅 시뮬레이션을 위한 분광특성기반의 랜더링 기법)

  • Lee, Myong-Young;Cho, Yang-Ho;Lee, Cheol-Hee;Ha, Yeong-Ho
    • Proceedings of the IEEK Conference
    • /
    • 2005.11a
    • /
    • pp.379-382
    • /
    • 2005
  • This study proposes an effective algorithm that can render a realistic image of a lighting environment, especially an automotive rear lamp, using the backward ray tracing method. To producea realistic image similar to that perceived by the human eye, the incident light energy at the eye point estimated by a ray tracing algorithm is represented by XYZ tri-stimulus values, which are then converted into RGB values considering the particular display device.

  • PDF

A Study of Efficient Access Method based upon the Spatial Locality of Multi-Dimensional Data

  • Yoon, Seong-young;Joo, In-hak;Choy, Yoon-chul
    • Proceedings of the Korea Database Society Conference
    • /
    • 1997.10a
    • /
    • pp.472-482
    • /
    • 1997
  • Multi-dimensional data play a crucial role in various fields, as like computer graphics, geographical information system, and multimedia applications. Indexing method fur multi-dimensional data Is a very Important factor in overall system performance. What is proposed in this paper is a new dynamic access method for spatial objects called HL-CIF(Hierarchically Layered Caltech Intermediate Form) tree which requires small amount of storage space and facilitates efficient query processing. HL-CIF tree is a combination of hierarchical management of spatial objects and CIF tree in which spatial objects and sub-regions are associated with representative points. HL-CIF tree adopts "centroid" of spatial objects as the representative point. By reflecting objects′sizes and positions in its structure, HL-CIF tree guarantees the high spatial locality of objects grouped in a sub-region rendering query processing more efficient.

  • PDF

Automatic Mask Generation for 3D Makeup Simulation (3차원 메이크업 시뮬레이션을 위한 자동화된 마스크 생성)

  • Kim, Hyeon-Joong;Kim, Jeong-Sik;Choi, Soo-Mi
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02a
    • /
    • pp.397-402
    • /
    • 2008
  • 본 논문에서는 햅틱 인터랙션 기반의 3차원 가상 얼굴 메이크업 시뮬레이션에서 메이크업 대상에 대한 정교한 페인팅을 적용하기 위한 자동화된 마스크 생성 방법을 개발한다. 본 연구에서는 메이크업 시뮬레이션 이전의 전처리 과정에서 마스크를 생성한다. 우선, 3차원 스캐너 장치로부터 사용자의 얼굴 텍스쳐 이미지와 3차원 기하 표면 모델을 획득한다. 획득된 얼굴 텍스쳐 이미지로부터 AdaBoost 알고리즘, Canny 경계선 검출 방법과 색 모델 변환 방법 등의 영상처리 알고리즘들을 적용하여 마스크 대상이 되는 주요 특정 영역(눈, 입술)들을 결정하고 얼굴 이미지로부터 2차원 마스크 영역을 결정한다. 이렇게 생성된 마스크 영역 이미지는 3차원 표면 기하 모델에 투영되어 최종적인 3차원 특징 영역의 마스크를 레이블링하는데 사용된다. 이러한 전처리 과정을 통하여 결정된 마스크는 햅틱 장치와 스테레오 디스플레이기반의 가상 인터페이스를 통해서 자연스러운 메이크업 시뮬레이션을 수행하는데 사용된다. 본 연구에서 개발한 방법은 사용자에게 전처리 과정에서의 어떠한 개입 없이 자동적으로 메이크업 대상이 되는 마스크 영역을 결정하여 정교하고 손쉬운 메이크업 페인팅 인터페이스를 제공한다.

  • PDF

An Efficient Real-Time Rendering of Large Molecular Models based on GPU (GPU 기반의 효율적인 거대 분자의 실시간 렌더링 기법)

  • Lee, Jun;Park, Sung-Jun;Kim, Jee-In
    • Journal of the Korea Computer Graphics Society
    • /
    • v.11 no.3
    • /
    • pp.19-22
    • /
    • 2005
  • 정보생물학 분야에 있어서 분자 구조를 3차원으로 렌더링하여 보여주는 것은 매우 중요한 작업이다. 특히 분자의 표면 렌더링은 분자의 3차원 구조 분석 등에 중요하게 사용된다. 그러나 분자 표면 렌더링을 수행하기 위해서는 많은 양의 폴리곤이 필요하게 된다. 특히 대장균 바이러스와 같은 분자량이 많은 거대 분자를 자연스럽게 렌더링 하기 위해서는 고가의 그래픽 전용 워크스테이션을 사용해야 한다. 본 논문에서는 저렴한 일반 PC 급 시스템에서도 거대 분자를 무리 없이 렌더링 할 수 있는 효율적인 알고리즘을 제안하였다. 제안하는 알고리즘은 높은 속도와 좋은 화질을 유지할 수 있는 Hybrid Point & Polygon 렌더링 기법이다. 이 알고리즘은 계층적인 자료구조인 옥트리(Octree)를 사용하였으며 최적의 성능을 내기 위하여 GPU가 작업을 처리한다. 제안된 알고리즘의 성능 평가는 일반 PC급에서 수행되었으며 특히 그래픽 카드 2개를 병렬로 연결하여 높은 성능을 낼 수 있는 SLI(Scalable Link Interface) 환경에서 평가를 수행하였다.

  • PDF

An Implementation of High Speed Rendering to Process Touch Screen Multiple Inputs based on FPGA (FPGA 기반의 터치스크린 다중입력처리를 위한 고속 렌더링 구현)

  • Yoon, Junhan;Kim, Jin Heon
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.11
    • /
    • pp.1803-1810
    • /
    • 2017
  • A large amount of processing time is required if the process of detecting the touch position on the touch screen and displaying it on the display panel is performed only by software. In this paper, we propose a method to output information touched on the screen using H/W method in order to improve the response speed delay. In the FPGA module designed for the HDMI signal output to the display module, the touch information is input to the serial data signal including touch coordinate information, point size, and color information. Then the module render the image using HDMI signal input to the module and the touch information. This method has a pipeline structure so it has effect of reducing the delay time that occurs in outputting the touch information compared with the conventional software processing method.