• Title/Summary/Keyword: cloud image

Search Result 448, Processing Time 0.024 seconds

A Comparison of 3D Reconstruction through the Passive and Pseudo-Active Acquisition of Images (수동 및 반자동 영상획득을 통한 3차원 공간복원의 비교)

  • Jeona, MiJeong;Kim, DuBeom;Chai, YoungHo
    • Journal of Broadcast Engineering
    • /
    • v.21 no.1
    • /
    • pp.3-10
    • /
    • 2016
  • In this paper, two reconstructed point cloud sets with the information of 3D features are analyzed. For a certain 3D reconstruction of the interior of a building, the first image set is taken from the sequential passive camera movement along the regular grid path and the second set is from the application of the laser scanning process. Matched key points over all images are obtained by the SIFT(Scale Invariant Feature Transformation) algorithm and are used for the registration of the point cloud data. The obtained results are point cloud number, average density of point cloud and the generating time for point cloud. Experimental results show the necessity of images from the additional sensors as well as the images from the camera for the more accurate 3D reconstruction of the interior of a building.

Prerequisite Research for the Development of an End-to-End System for Automatic Tooth Segmentation: A Deep Learning-Based Reference Point Setting Algorithm (자동 치아 분할용 종단 간 시스템 개발을 위한 선결 연구: 딥러닝 기반 기준점 설정 알고리즘)

  • Kyungdeok Seo;Sena Lee;Yongkyu Jin;Sejung Yang
    • Journal of Biomedical Engineering Research
    • /
    • v.44 no.5
    • /
    • pp.346-353
    • /
    • 2023
  • In this paper, we propose an innovative approach that leverages deep learning to find optimal reference points for achieving precise tooth segmentation in three-dimensional tooth point cloud data. A dataset consisting of 350 aligned maxillary and mandibular cloud data was used as input, and both end coordinates of individual teeth were used as correct answers. A two-dimensional image was created by projecting the rendered point cloud data along the Z-axis, where an image of individual teeth was created using an object detection algorithm. The proposed algorithm is designed by adding various modules to the Unet model that allow effective learning of a narrow range, and detects both end points of the tooth using the generated tooth image. In the evaluation using DSC, Euclid distance, and MAE as indicators, we achieved superior performance compared to other Unet-based models. In future research, we will develop an algorithm to find the reference point of the point cloud by back-projecting the reference point detected in the image in three dimensions, and based on this, we will develop an algorithm to divide the teeth individually in the point cloud through image processing techniques.

A Deep Learning Approach for Classification of Cloud Image Patches on Small Datasets

  • Phung, Van Hiep;Rhee, Eun Joo
    • Journal of information and communication convergence engineering
    • /
    • v.16 no.3
    • /
    • pp.173-178
    • /
    • 2018
  • Accurate classification of cloud images is a challenging task. Almost all the existing methods rely on hand-crafted feature extraction. Their limitation is low discriminative power. In the recent years, deep learning with convolution neural networks (CNNs), which can auto extract features, has achieved promising results in many computer vision and image understanding fields. However, deep learning approaches usually need large datasets. This paper proposes a deep learning approach for classification of cloud image patches on small datasets. First, we design a suitable deep learning model for small datasets using a CNN, and then we apply data augmentation and dropout regularization techniques to increase the generalization of the model. The experiments for the proposed approach were performed on SWIMCAT small dataset with k-fold cross-validation. The experimental results demonstrated perfect classification accuracy for most classes on every fold, and confirmed both the high accuracy and the robustness of the proposed model.

A Performance Test of Mobile Cloud Service for Bayesian Image Fusion (베이지안 영상융합을 적용한 모바일 클라우드 성능실험)

  • Kang, Sanggoo;Lee, Kiwon
    • Korean Journal of Remote Sensing
    • /
    • v.30 no.4
    • /
    • pp.445-454
    • /
    • 2014
  • In recent days, trend technologies for cloud, bigdata, or mobile, as the important marketable keywords or paradigm in Information Communication Technology (ICT), are widely used and interrelated each other in the various types of platforms and web-based services. Especially, the combination of cloud and mobile is recognized as one of a profitable business models, holding benefits of their own. Despite these challenging aspects, there are a few application cases of this model dealing with geo-based data sets or imageries. Among many considering points for geo-based cloud application on mobile, this study focused on a performance test of mobile cloud of Bayesian image fusion algorithm with satellite images. Two kinds of cloud platform of Amazon and OpenStack were built for performance test by CPU time stamp. In fact, the scheme for performance test of mobile cloud is not established yet, so experiment conditions applied in this study are to check time stamp. As the result, it is revealed that performance in two platforms is almost same level. It is implied that open source mobile cloud services based on OpenStack are enough to apply further applications dealing with geo-based data sets.

Cloud Removal Using Gaussian Process Regression for Optical Image Reconstruction

  • Park, Soyeon;Park, No-Wook
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.4
    • /
    • pp.327-341
    • /
    • 2022
  • Cloud removal is often required to construct time-series sets of optical images for environmental monitoring. In regression-based cloud removal, the selection of an appropriate regression model and the impact analysis of the input images significantly affect the prediction performance. This study evaluates the potential of Gaussian process (GP) regression for cloud removal and also analyzes the effects of cloud-free optical images and spectral bands on prediction performance. Unlike other machine learning-based regression models, GP regression provides uncertainty information and automatically optimizes hyperparameters. An experiment using Sentinel-2 multi-spectral images was conducted for cloud removal in the two agricultural regions. The prediction performance of GP regression was compared with that of random forest (RF) regression. Various combinations of input images and multi-spectral bands were considered for quantitative evaluations. The experimental results showed that using multi-temporal images with multi-spectral bands as inputs achieved the best prediction accuracy. Highly correlated adjacent multi-spectral bands and temporally correlated multi-temporal images resulted in an improved prediction accuracy. The prediction performance of GP regression was significantly improved in predicting the near-infrared band compared to that of RF regression. Estimating the distribution function of input data in GP regression could reflect the variations in the considered spectral band with a broader range. In particular, GP regression was superior to RF regression for reproducing structural patterns at both sites in terms of structural similarity. In addition, uncertainty information provided by GP regression showed a reasonable similarity to prediction errors for some sub-areas, indicating that uncertainty estimates may be used to measure the prediction result quality. These findings suggest that GP regression could be beneficial for cloud removal and optical image reconstruction. In addition, the impact analysis results of the input images provide guidelines for selecting optimal images for regression-based cloud removal.

Effect of All Sky Image Correction on Observations in Automatic Cloud Observation (자동 운량 관측에서 전천 영상 보정이 관측치에 미치는 효과)

  • Yun, Han-Kyung
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.15 no.2
    • /
    • pp.103-108
    • /
    • 2022
  • Various studies have been conducted on cloud observation using all-sky images acquired with a wide-angle camera system since the early 21st century, but it is judged that an automatic observation system that can completely replace the eye observation has not been obtained. In this study, to verify the quantification of cloud observation, which is the final step of the algorithm proposed to automate the observation, the cloud distribution of the all-sky image and the corrected image were compared and analyzed. The reason is that clouds are formed at a certain height depending on the type, but like the retina image, the center of the lens is enlarged and the edges are reduced, but the effect of human learning ability and spatial awareness on cloud observation is unknown. As a result of this study, the average cloud observation error of the all-sky image and the corrected image was 1.23%. Therefore, when compared with the eye observation in the decile, the error due to correction is 1.23% of the observed amount, which is very less than the allowable error of the eye observation, and it does not include human error, so it is possible to collect accurately quantified data. Since the change in cloudiness due to the correction is insignificant, it was confirmed that accurate observations can be obtained even by omitting the unnecessary correction step and observing the cloudiness in the pre-correction image.

An Open Source Mobile Cloud Service: Geo-spatial Image Filtering Tools Using R (오픈소스 모바일 클라우드 서비스: R 기반 공간영상정보 필터링 사례)

  • Kang, Sanggoo;Lee, Kiwon
    • Spatial Information Research
    • /
    • v.22 no.5
    • /
    • pp.1-8
    • /
    • 2014
  • Globally, mobile, cloud computing or big data are the recent marketable key terms. These trend technologies or paradigm in the ICT (Information Communication Technology) fields exert large influence on the most application fields including geo-spatial applications. Among them, cloud computing, though the early stage in Korea now, plays a important role as a platform for other trend technologies uses. Especially, mobile cloud, an integrated platform with mobile device and cloud computing can be considered as a good solution to overcome well known limitations of mobile applications and to provide more information processing functionalities to mobile users. This work is a case study to design and implement the mobile application system for geo-spatial image filtering processing operated on mobile cloud platform built using OpenStack and various open sources. Filtering processing is carried out using R environment, recently being recognized as one of big data analysis technologies. This approach is expected to be an element linking geo-spatial information for new service model development and the geo-spatial analysis service development using R.

2D Interpolation of 3D Points using Video-based Point Cloud Compression (비디오 기반 포인트 클라우드 압축을 사용한 3차원 포인트의 2차원 보간 방안)

  • Hwang, Yonghae;Kim, Junsik;Kim, Kyuheon
    • Journal of Broadcast Engineering
    • /
    • v.26 no.6
    • /
    • pp.692-703
    • /
    • 2021
  • Recently, with the development of computer graphics technology, research on technology for expressing real objects as more realistic virtual graphics is being actively conducted. Point cloud is a technology that uses numerous points, including 2D spatial coordinates and color information, to represent 3D objects, and they require huge data storage and high-performance computing devices to provide various services. Video-based Point Cloud Compression (V-PCC) technology is currently being studied by the international standard organization MPEG, which is a projection based method that projects point cloud into 2D plane, and then compresses them using 2D video codecs. V-PCC technology compresses point cloud objects using 2D images such as Occupancy map, Geometry image, Attribute image, and other auxiliary information that includes the relationship between 2D plane and 3D space. When increasing the density of point cloud or expanding an object, 3D calculation is generally used, but there are limitations in that the calculation method is complicated, requires a lot of time, and it is difficult to determine the correct location of a new point. This paper proposes a method to generate additional points at more accurate locations with less computation by applying 2D interpolation to the image on which the point cloud is projected, in the V-PCC technology.

Combining Conditional Generative Adversarial Network and Regression-based Calibration for Cloud Removal of Optical Imagery (광학 영상의 구름 제거를 위한 조건부 생성적 적대 신경망과 회귀 기반 보정의 결합)

  • Kwak, Geun-Ho;Park, Soyeon;Park, No-Wook
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.6_1
    • /
    • pp.1357-1369
    • /
    • 2022
  • Cloud removal is an essential image processing step for any task requiring time-series optical images, such as vegetation monitoring and change detection. This paper presents a two-stage cloud removal method that combines conditional generative adversarial networks (cGANs) with regression-based calibration to construct a cloud-free time-series optical image set. In the first stage, the cGANs generate initial prediction results using quantitative relationships between optical and synthetic aperture radar images. In the second stage, the relationships between the predicted results and the actual values in non-cloud areas are first quantified via random forest-based regression modeling and then used to calibrate the cGAN-based prediction results. The potential of the proposed method was evaluated from a cloud removal experiment using Sentinel-2 and COSMO-SkyMed images in the rice field cultivation area of Gimje. The cGAN model could effectively predict the reflectance values in the cloud-contaminated rice fields where severe changes in physical surface conditions happened. Moreover, the regression-based calibration in the second stage could improve the prediction accuracy, compared with a regression-based cloud removal method using a supplementary image that is temporally distant from the target image. These experimental results indicate that the proposed method can be effectively applied to restore cloud-contaminated areas when cloud-free optical images are unavailable for environmental monitoring.

EEIRI: Efficient Encrypted Image Retrieval in IoT-Cloud

  • Abduljabbar, Zaid Ameen;Ibrahim, Ayad;Hussain, Mohammed Abdulridha;Hussien, Zaid Alaa;Al Sibahee, Mustafa A.;Lu, Songfeng
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.11
    • /
    • pp.5692-5716
    • /
    • 2019
  • One of the best means to safeguard the confidentiality, security, and privacy of an image within the IoT-Cloud is through encryption. However, looking through encrypted data is a difficult process. Several techniques for searching encrypted data have been devised, but certain security solutions may not be used in IoT-Cloud because such solutions are not lightweight. We propose a lightweight scheme that can perform a content-based search of encrypted images, namely EEIRI. In this scheme, the images are represented using local features. We develop and validate a secure scheme for measuring the Euclidean distance between two descriptor sets. To improve the search efficiency, we employ the k-means clustering technique to construct a searchable tree-based index. Our index construction process ensures the privacy of the stored data and search requests. When compared with more familiar techniques of searching images over plaintexts, EEIRI is considered to be more efficient, demonstrating a higher search cost of 7% and a decrease in search accuracy of 1.7%. Numerous empirical investigations are carried out in relation to real image collections so as to evidence our work.