Browse > Article
http://dx.doi.org/10.7848/ksgpc.2020.38.2.109

Investigation on the Applicability of Defocus Blur Variations to Depth Calculation Using Target Sheet Images Captured by a DSLR Camera  

Seo, Suyoung (Dept. of Civil Engineering, Kyungpook National University)
Publication Information
Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography / v.38, no.2, 2020 , pp. 109-121 More about this Journal
Abstract
Depth calculation of objects in a scene from images is one of the most studied processes in the fields of image processing, computer vision, and photogrammetry. Conventionally, depth is calculated using a pair of overlapped images captured at different view points. However, there have been studies to calculate depths from a single image. Theoretically, it is known to be possible to calculate depth using the diameter of CoC (Circle of Confusion) caused by defocus under the assumption of a thin lens model. Thus, this study aims to verify the validity of the thin lens model to calculate depth from edge blur amount which corresponds to the radius of CoC. For this study, a commercially available DSLR (Digital Single Lens Reflex) camera was used to capture a set of target sheets which had different edge contrasts. In order to find out the pattern of the variations of edge blur against varying combination of FD (Focusing Distance) and OD (Object Distance), the camera was set to varying FD and target sheet images were captured at varying OD under each FD. Then, the edge blur and edge displacement were estimated from edge slope profiles using a brute-force method. The experimental results show that the pattern of the variations of edge blur observed in the target images was apart from their corresponding theoretical amounts derived under the thin lens assumption but can still be utilized to calculate depth from a single image for the cases similar to the limited conditions experimented under which the tendency between FD and OD is manifest.
Keywords
Defocus; Blur; Target Sheet; Blur Variation; DSLR Camera; Depth Calculation;
Citations & Related Records
연도 인용수 순위
  • Reference
1 Cao, Y., Fang, S., and Wang, Z. (2013), Digital multifocusing from a single photograph taken with an uncalibrated conventional camera, IEEE Transactions on Image Processing, Vol. 22, No. 9, pp. 3703-3714.   DOI
2 Chen, S.-J. and Shen, H.-L. (2015), Multispectral image outof-focus deblurring using interchannel correlation, IEEE Transactions on Image Processing, Vol. 24, No. 11, pp. 4433-4445.   DOI
3 De, I., Chanda, B., and Chattopadhyay, B. (2006), Enhancing effective depth-of-field by image fusion using mathematical morphology, Image and Vision Computing, Vol. 24, pp. 1278-1287.   DOI
4 Gajjar, R. and Zaveri, T. (2017), Defocus blur parameter estimation using polynomial expression and signature based methods, 4th International Conference on Signal Processing and Integrated Networks, 2-3 February, Noida, India, pp. 71-75.
5 Green, P., Sun, W., Matusik, W., and Durand, F. (2007), Multi-aperture photography, ACM Transactions on Graphics, Vol. 26, No. 3, Article 68.
6 Hecht, E. (2001), Optics, 4th ed., Addison Wesley.
7 Hong, Y., Ren, G., Liu, E., and Sun J. (2015), A blur estimation and detection method for out-of-focus images, Multimedia Tools and Applications, Vol. 75, pp. 10807-10822.   DOI
8 Karaali, A. and Jung, C.R. (2018), Edge-based defocus blur estimation with adaptive scale selection, IEEE Transactions on Image Processing, Vol. 27, No. 3, pp. 1126-1137.   DOI
9 Kim, S., Lee, E., Hayes, M.H., and Paik, J. (2012), Multifocusing and depth estimation using a color shift model-based computational camera, IEEE Transactions on Image Processing, Vol. 21, No. 9, pp. 4152-4166.   DOI
10 Lu. Q., Zhou, W. Fang L., and Li, H. (2016), Robust blur kernel estimation for license plate images from fast moving vehicles, IEEE Transactions on Image Processing, Vol. 25, No. 5, pp. 2311-2323.   DOI
11 Moghaddam, M.E. (2007), A mathematical model to estimate out of focus blur, 5th International Symposium on Image and Signal Processing and Analysis, 27-29 September, Istanbul, Turkey, pp. 278-281.
12 Oliveira, J.P., Figueiredo, M.A.T., and Bioucas-Dias, J.M. (2014), Parametric blur estimation for blind restoration of natural images: linear motion and out-of-focus, IEEE Transactions on Image Processing, Vol. 23, No. 1, pp. 466-477.   DOI
13 Sun, T.-Y., Ciou, S.-J., Liu, C.-C., and Huo, C.-L. (2009), Outof-focus blur estimation for blind image deconvolution: using particle swarm optimization, Proceedings of the 2009 IEEE International Conference on Systems, Man, and Cybernetics, 11-14 October, San Antonio, Texas, pp. 1627-1632.
14 Reuter, A., Seidel, H.-P., and Ihrke, I. (2012), BlurTags: spatially varying PSF estimation with out-of-focus patterns, 20th International Conference on Computer Graphics, Visualization and Computer Vision, June, Plenz, Czech Republic, pp. 239-247.
15 Seo, S. (2017a), Prediction of edge displacement due to image contrast, The Photogrammetric Record, Vol. 32, No. 158, pp. 119-140.   DOI
16 Seo, S. (2017b), Estimation of edge displacement against brightness and camera-to-object distance, IET Image Processing, Vol. 11, No. 8, pp. 568-577.   DOI
17 Seo, S. (2018a), Edge modeling by two blur parameters in varying contrasts, IEEE Transactions on Image Processing, Vol. 27, No. 6, pp. 2701-2714.   DOI
18 Seo, S. (2018b), Subpixel edge localization based on adaptive weighting of gradients, IEEE Transactions on Image Processing, Vol. 27, No. 11, pp. 5501-5513.   DOI
19 Tang, C., Hou, C., and Song, Z. (2013), Defocus map estimation from a single image via spectrum contrast, Optics Letters, Vol. 38, No. 10, pp. 1706-1708.   DOI
20 Wu, S. and Lin, W. (2008), Defocus estimation from a single image, Proceedings of 17th International Conference on Computer Communications and Networks, 3-7 August, St. Thomas, US Virgin Islands, pp. 1-5.
21 Yuan, L., Sun, J., Quan, L., and Shum, H.-Y. (2007), Image deblurring with blurred/noisy image pairs, SIGGRAPH-07: Special interest group on computer graphics and interactive techniques conference, August, San Diego, California, pp. 1-es.
22 Zhou, C., Lin, S., and Nayar, S. (2009), Coded aperture pairs for depth from defocus, IEEE 12th International Conference on Computer Vision-2009, 29 September-2 October, Kyoto, Japan, pp. 325-332.
23 Zhu, X., Cohen, S., Schiller, S., and Milnafar, P. (2013), Estimating spatially varying defocus blur from a single image, IEEE Transactions on Image Processing, Vol. 22, No. 12, pp. 4879-4891.   DOI
24 Zhuo, S. and Sim, T. (2011), Defocus map estimation from a single image, Pattern Recognition, Vol. 44, pp. 1852-1858.   DOI
25 Zhang, Y. and Hirakawa, K. (2016) Blind deblurring and denoising of images corrupted by unidirectional object motion blur and sensor noise, IEEE Transactions on Image Processing, Vol. 25, No. 9, pp. 4129-4144.   DOI
26 Zhang, Y. and Hirakawa, K. (2015) Fast spatially varying object motion blur estimation, IEEE International Conference on Image Processing, 27-30 September, Quebec City, Canada, pp. 646-650.