1. INTRODUCTION
3D content system is a technology which helps users accept a VR(Virtual Reality) as reality through 3D visual information and feel as if they are in the field, which are utilized in various fields such as education, advertising, home shopping, military training, and mock labs and medical industry. This kind of computer display system significantly contributes to the development of user interface, real-time object recognition, cuttingedge systems for the disabled. Various studies on 3D content including information about a sense of 3D effect and immersion are carried out home and abroad, increasing fields utilizing them[1,2].
The typical technology is a image-based modeling and rendering which establish 3D augmented environment based on images. This method lacks a sense of reality since it provides realistic images of the environment or objects at arbitrary times based on information from images taken at actual environments. However, it has merits in that it doesn't need a traditional 3D modeling process, which reduces costs in th establishment and a production speed of printing images. To overcome these problems, studies on 3D image production through image synthesis and processing are carried out actively home and abroad[3-7].
In this study, a multi-scale 3D panorama content augmented system using a depth-map of images is suggested. By finding matching features from images to add 3D features to an augmented environment, a depth-map is derived and embodied as panorama, producing high-quality augmented content system with a sense of reality. This system adds a high-quality panorama technology to userfriendly technology, providing 3D panorama contents which can be intuitively and easily manipulated, thereby improving a sense of immersion and reality of users.
2. RELATED WORKS
2.1 Panorama VR
Panorama VR is to show nearby sight at one's location. it gives an effect just like he feels in the place and it can reenact at certain days. To acquire this effect, panoramic technique is used. Images of real environment taken in 360° are called omni images or omnidirectional images. And images transformed them into cylinder forms, circular or rectangular forms are called panoramic images[8-10].
To establish panorama VR, two steps are needed. The first step is to acquire panoramic images and the second step is to display them by panorama VR viewer. When users designate a certain gaze direction after receiving panoramic images, the panorama VR viewer shows the image seen from the direction by producing it from panoramic images in real time. Since panorama VR is widely used in the Internet, various kinds of panorama VR viewers have appeared. The most currently used viewers are Apple's QuickTime Player, Macromedia's Shockwave Player,Anything 3D's viewer, iSeeMedia's viewer, Zoomify's viewer, iPIX's viewer, and MG System's Viewer. Fig 1. shows how to produce panoramic VR using a cylinder.
Fig. 1.Panorama VR process step.
2.2 Software method of production of panoramic images
Image mosaic is a method for matching several images and connecting them by general camera. First, set the starting point of photo-taking. And then rotate horizontally by overlapping 10%-50% of images and take 12-24 pictures according to the degrees of overlapping. Connect the overlapping parts to make a one piece of panorama, which is called a stitching process. For images to configure a cylinder by matching with neighboring ones, each image will be bent. Overlapped pictures should be bent in the same angle to fit perfectly. Each picture image is a flat rectangular image first, but to configure a cylinder its corner should be bent. The degrees of bending is different according to a filed of view of a used lens, and the images which become the cylinder part are more distorted. And when playing them next, the images will become flat as original. A final source image will have different dimension and ratio according to each images filed of view. The bigger the filed of view, the higher the height becomes compared to area. In other words, the smaller the filed of view, the larger the panorama. Pictures completing stitching have different tops and bottoms in length since they seem to connect pictures cut long. And then the blending process which make them a panorama picture by connecting them smoothly is car ried out. The process is shown in Fig 2.
Fig. 2.Image blending step.
To produce high-quality panoramic images effectively, the exact camera direction of each input image should be found out. To do so, after finding the relative camera rotations of neighboring images through image matching, the camera direction is calculated by adding them in order. This approach has demerits in that errors occurring in image matching are accumulated and final panoramic images have gaps or overlays[10-13].
3. 3D PANORAMA GENERATION ALGORITHM
Let x=(x,y) which is a point from projection of X=(X,Y,Z), a point in 3D(dimensional) space that is taken by lens distortion free camera, to 2D image, the relationship between X and x is modelled as follow[14].
u is a vector that represents x as Homogeneous Coordinates. K, as a Camera Calibration Matrix, shows the relationship between 3D point on camera coordinates and 2D point on image within which 3D point is projected. R and t, that represent Rotation Matrix and Translation Vector each, indicate the transformation relationship between world coordinates and camera coordinates. As we call parameters included in vector K as Internal Parameter of camera, f is Focal Length, Q is Aspect Ratio of pixel, S is Skew and (Xc, Yc) is Principal Point.
Rotation Matrix and Translation Vector of camera before rotating are R0 and t0 respectively and those of camera after rotating are R1 and t1. When conforming projection center to zero point of world coordinates for convenience's sake, it becomes t0=t1=0. If homogeneous vector of a point that is from the projection of certain point X at 3 dimensional space to I0 (original image) and I1 (rotated image) is each u0 and u1, then.
That is, as pixel correspondence relationship between two images that are from camera rotation, which is presented by a 3 × 3 matrix H, we call it Planar Homography. Considering that corresponding image point is the same even though multiplying any real number except 0 to Homogeneous Vector, H also can be defined regardless of its scale. Hence, Planar Homography is defined 8 parameters, not 9 parameters. Actually, as the number of parameter that is included in K from H=KR01K-1 is 5 and the number of parameter that is in R01 is 3, we can aware that independent parameters that are composed of H is 8. This Homography is called as 8-Parameter Homography.
Given two images, finding H which is pixel correspondence relationship between the two images is Image Registration. In general, H can be calculated with more than 4 correspondent points but it takes caution to get more accurate H. Therefore, we assume that camera satisfies this normalized model. Table 1. shows the conditions of camera based on normalized model.
Table 1.Camera condition of normalized model
Therefore, the number of independent parameter included in H on normalized model is 4(focal point distance f and three rotating angles within R 01), which we call as 4-Parameter Homography. General rotation matrix R is divided as follow.
In formula(4), R(ψ), R(θ), R(∅) are respective matrix that present rotation transformation from each X, Y, Z axis and ψ, θ, ∅ are called Euler Angle. Given Planar Homography H in normalized model, calculating K and R are as shown below. Matrix K can be divided as follow.
Substituting to formula(3)
Therefore,
As R is Orthogonal Matrix, it satisfies following formula. After calculating f from formula(8), it gets R by substituting to formula(7).
Camera models described above do not consider distortion of its lens. As it utilizes wide angle lens to get 360 angle panorama video, it is necessary to consider lens distortion to apply algorithm to wider area. (xd,yd) is coordinates of pixel in original image(that is, image that contains lens distortion) and (xu,yu) is corresponding coordinates of pixel in image that removes lens distortion. As (xc,yc) is the center coordinates, given (xd,yd), (xu,yu) is calculated as follow.
In this moment, k1,k2 are radial distortion coefficients. In common,
To the contrary when (xu,yu) is given, (xd,yd) is as follow.
In formula (12),
4. MULTI-SCALE 3D PANORAMA CONTENT AUGMENTED SYSTEM USING DEPTHMAP
Multi-scale 3D panorama content augmented system using depth map produces one panoramic image by analyzing imput images of multi-scale images. The image stitching technique is a method for making images taken in certain angles as a stretched out form without joints. Multiple images acquired from more than one cameras are produced as one panoramic image without joints through a image synthesis process. Since cameras used to acquire a panoramic image are set in different directions and locations, overlapping parts of images should be found out and geometrical adjustments should be carried out to produce one panoramic image without joints.
Multi-scale 3D panorama augmented reality content system produces a depth-map between images internally to produce panoramic images. To derive a depth-map, median filters are applied to remove noises from images. Edges are detected through the derivation of canny edges to input images through preprocessing, and vanishing lines are detected through Hough transformation. Through the intersection points of the detected vanishing lines, vanishing points are designated. Since the depth step of the vanishing points should be different depending on the locations of images, the range of locations of corresponding vanishing points between images and vanishing points are defined. By setting the defined range as standard, a standard depth step on a base level is produced, By linear interpolation of neighboring vanishing points, a depth-map is produced.
The production of a depth-map comprises three steps. The first step, the detection of edge, reduces noise as a preprocessing for input images by applying median filters and derives geometric features by applying canny edge. The second step, the production of vanishing lines and points, derives straight lines through Hough transformation, calculates the intersection points of the straight lines, and estimates the locations of vanishing points through the rage of locations of defined vanishing points. The final step, the estimation of vanishing point locations and production of a depth-map, estimates the locations of vanishing points through the range of locations of vanishing points, sets the steps of depth, and finally produces a depth-map based on the locations of vanishing points.
To produce multi-scale panorama image, the relationships between images are set internally. Multi-scale panorama image is produced through the blending process which makes input images into panoramic pictures. According to the change of image scale, the exact corresponding locations with neighboring images are automatically calculated, securing the exactness within sub pixels and stitching the images. Fig 3. shows a general configuration map of multi-scale 3D panorama augmented system using a depth-map suggested in this study. Fig 4. shows the steps of stitching images derived from a depth-map, producing depthmap panorama, and embodying multi-scale 3D panorama augmented system.
Fig. 3.Multi-scale 3D panorama augmented system configuration map.
Fig. 4.Embodiment of multi-scale 3D panorama augmented system using depth-map.
5. CONCLUSION AND FUTURE WORKS
In this study, features matching images found out to add a sense of 3D effect to the augmented environment and a depth-map is used to embody 3D panorama, producing the augmented system with a sense of reality. This system will overcome the limits of 2D panorama technologies and provide users with a natural navigation, providing an improved sense of reality and immersion. Multi-scale 3D panorama augmented system using depth-map makes users feel as if they are at the location which photo-taking is taken and seeing the surroundings, in other words of a sense of 3D effect. Multi-scale 3D panorama augmented system provides highquality 3D images with a sense of reality which has a free vision of close objects and distant backgrounds. This suggests a possibility in that it can be utilized in establishing augmented reality system based on images, attracting a attention from many people.
Further studies will be carried out based on computer vision technologies including image processing and interpretation technologies and computer graphics configuring media images with a sense of reality from the interpreted images to commercialize tourist attractions tours with a sense of reality for interaction with users and embodied images and high-quality digital content of public relations office.
References
- D. Burschka, D. Cobzas, Z. Dodds, G. Hager, M. Jagersand, and K. Yerex, "Recent Methods for Image-Based Modeling and Rendering," IEEE Virtual Reality 2003 Tutorial 1, pp. 1-85, 2003.
- H.K. Shum, S.B. Kang, and S.C. Chan, "Survey of Image-based Representations and Compression Techniques," IEEE Transaction On Circuits and Systems for Video Technology, Vol. 13, No. 11, pp. 1020-1037, 2003. https://doi.org/10.1109/TCSVT.2003.817360
- B.E. Shelton and N.R. Hedley, "Using Augmented Reality for Teaching Earth-Sun Relationships to Undergraduate Geography Students," Proceeding of First IEEE International Augmented Reality Toolkit Workshop, pp. 1-8, 2002.
- Z. Yan, X. Ma, and S. Tao, "The Virtual Display Study on 3D Panorama in Tourist Areas-Take Shilin World Geopark as an Example," Proceeding of the International Symposium on Information Processing, pp. 229-232, 2009.
- S.M. Chun, J.Y. Choi, S.H. Kim, Y.C. Cho, and K.S. Park, "A Study on Depth Map Quantization for Multiview Image Generation," Journal of Korean Society For Computer Game, Vol. 26, No. 4, pp. 219-226, 2013.
- S. Curti, D. Sirtori, and F. Vella, "3D Effect Generation from Monocular View," IEEE Proceeding of the 3D Data Processing Visualization and Transmission, pp. 550-553, 2002.
- T. Hassner and R. Basri, "Example Based 3D Reconstruction from Single 2D Images," Proceeding of the Conference on Computer Vision and Pattern Recognition Workshop, pp. 15-22, 2006.
- L. Juan and O. Gwun, "SURF Applied in Panorama Image Stitching," Proceeding of 2nd International Conference on Image Processing Theory Tools and Applications, pp. 495-499, 2010.
- M. Brown and D.G. Lowe, "Automatic Panoramic Image Stitching using Invariant Features," International Journal of Computer Vision, Vol. 74, No. 1, pp. 59-73, 2007. https://doi.org/10.1007/s11263-006-0002-3
- R. Szeliski, "Video Mosaics for Virtual Environments," IEEE Computer Graphics and Applications, Vol. 16, No. 2, pp. 22-30, 1996. https://doi.org/10.1109/38.486677
- R. Szeliski and H.Y. Shum, "Creating Full View Panoramic Image Mosaics and Environment Maps," Proceeding of the Conference on Computer Graphics and Interactive Techniques, pp. 251-258, 1997.
- N. Chiba, M. Minoh, and H. Kano, "Feature-Based Lens Distortion Correction for Image Mosaicing," Proceeding of the IAPR Workshop on Machine Vision Applications, pp. 607-610, 2000.
- J. Luo, S.S. Shin, H. J. Park, and O. B. Gwun, "Stitcing for Panorama based on SURF and Multi-band Blending," Journal of Korea Multimedia Society, Vol. 14, No. 2, pp. 201-209, 2011. https://doi.org/10.9717/kmms.2011.14.2.201
- S.I. Cho, K.J. Kim, K.J. Ban, K.W. Park, C.Y. Kim, and E.K. Kim, "3D Panorama Generation Using Depth-Map Stitching," Journal of Information and Communication Convergence Engineering, Vol. 9, No. 6, pp. 780-784, 2011. https://doi.org/10.6109/ijice.2011.9.6.780