• Title/Summary/Keyword: Improved deep learning

Search Result 571, Processing Time 0.03 seconds

Detection of Zebra-crossing Areas Based on Deep Learning with Combination of SegNet and ResNet (SegNet과 ResNet을 조합한 딥러닝에 기반한 횡단보도 영역 검출)

  • Liang, Han;Seo, Suyoung
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.39 no.3
    • /
    • pp.141-148
    • /
    • 2021
  • This paper presents a method to detect zebra-crossing using deep learning which combines SegNet and ResNet. For the blind, a safe crossing system is important to know exactly where the zebra-crossings are. Zebra-crossing detection by deep learning can be a good solution to this problem and robotic vision-based assistive technologies sprung up over the past few years, which focused on specific scene objects using monocular detectors. These traditional methods have achieved significant results with relatively long processing times, and enhanced the zebra-crossing perception to a large extent. However, running all detectors jointly incurs a long latency and becomes computationally prohibitive on wearable embedded systems. In this paper, we propose a model for fast and stable segmentation of zebra-crossing from captured images. The model is improved based on a combination of SegNet and ResNet and consists of three steps. First, the input image is subsampled to extract image features and the convolutional neural network of ResNet is modified to make it the new encoder. Second, through the SegNet original up-sampling network, the abstract features are restored to the original image size. Finally, the method classifies all pixels and calculates the accuracy of each pixel. The experimental results prove the efficiency of the modified semantic segmentation algorithm with a relatively high computing speed.

A Study on A Deep Learning Algorithm to Predict Printed Spot Colors (딥러닝 알고리즘을 이용한 인쇄된 별색 잉크의 색상 예측 연구)

  • Jun, Su Hyeon;Park, Jae Sang;Tae, Hyun Chul
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.45 no.2
    • /
    • pp.48-55
    • /
    • 2022
  • The color image of the brand comes first and is an important visual element that leads consumers to the consumption of the product. To express more effectively what the brand wants to convey through design, the printing market is striving to print accurate colors that match the intention. In 'offset printing' mainly used in printing, colors are often printed in CMYK (Cyan, Magenta, Yellow, Key) colors. However, it is possible to print more accurate colors by making ink of the desired color instead of dotting CMYK colors. The resulting ink is called 'spot color' ink. Spot color ink is manufactured by repeating the process of mixing the existing inks. In this repetition of trial and error, the manufacturing cost of ink increases, resulting in economic loss, and environmental pollution is caused by wasted inks. In this study, a deep learning algorithm to predict printed spot colors was designed to solve this problem. The algorithm uses a single DNN (Deep Neural Network) model to predict printed spot colors based on the information of the paper and the proportions of inks to mix. More than 8,000 spot color ink data were used for learning, and all color was quantified by dividing the visible light wavelength range into 31 sections and the reflectance for each section. The proposed algorithm predicted more than 80% of spot color inks as very similar colors. The average value of the calculated difference between the actual color and the predicted color through 'Delta E' provided by CIE is 5.29. It is known that when Delta E is less than 10, it is difficult to distinguish the difference in printed color with the naked eye. The algorithm of this study has a more accurate prediction ability than previous studies, and it can be added flexibly even when new inks are added. This can be usefully used in real industrial sites, and it will reduce the attempts of the operator by checking the color of ink in a virtual environment. This will reduce the manufacturing cost of spot color inks and lead to improved working conditions for workers. In addition, it is expected to contribute to solving the environmental pollution problem by reducing unnecessarily wasted ink.

A Deep Learning-based Real-time Deblurring Algorithm on HD Resolution (HD 해상도에서 실시간 구동이 가능한 딥러닝 기반 블러 제거 알고리즘)

  • Shim, Kyujin;Ko, Kangwook;Yoon, Sungjoon;Ha, Namkoo;Lee, Minseok;Jang, Hyunsung;Kwon, Kuyong;Kim, Eunjoon;Kim, Changick
    • Journal of Broadcast Engineering
    • /
    • v.27 no.1
    • /
    • pp.3-12
    • /
    • 2022
  • Image deblurring aims to remove image blur, which can be generated while shooting the pictures by the movement of objects, camera shake, blurring of focus, and so forth. With the rise in popularity of smartphones, it is common to carry portable digital cameras daily, so image deblurring techniques have become more significant recently. Originally, image deblurring techniques have been studied using traditional optimization techniques. Then with the recent attention on deep learning, deblurring methods based on convolutional neural networks have been actively proposed. However, most of them have been developed while focusing on better performance. Therefore, it is not easy to use in real situations due to the speed of their algorithms. To tackle this problem, we propose a novel deep learning-based deblurring algorithm that can be operated in real-time on HD resolution. In addition, we improved the training and inference process and could increase the performance of our model without any significant effect on the speed and the speed without any significant effect on the performance. As a result, our algorithm achieves real-time performance by processing 33.74 frames per second at 1280×720 resolution. Furthermore, it shows excellent performance compared to its speed with a PSNR of 29.78 and SSIM of 0.9287 with the GoPro dataset.

Development of Deep Learning Based Ensemble Land Cover Segmentation Algorithm Using Drone Aerial Images (드론 항공영상을 이용한 딥러닝 기반 앙상블 토지 피복 분할 알고리즘 개발)

  • Hae-Gwang Park;Seung-Ki Baek;Seung Hyun Jeong
    • Korean Journal of Remote Sensing
    • /
    • v.40 no.1
    • /
    • pp.71-80
    • /
    • 2024
  • In this study, a proposed ensemble learning technique aims to enhance the semantic segmentation performance of images captured by Unmanned Aerial Vehicles (UAVs). With the increasing use of UAVs in fields such as urban planning, there has been active development of techniques utilizing deep learning segmentation methods for land cover segmentation. The study suggests a method that utilizes prominent segmentation models, namely U-Net, DeepLabV3, and Fully Convolutional Network (FCN), to improve segmentation prediction performance. The proposed approach integrates training loss, validation accuracy, and class score of the three segmentation models to enhance overall prediction performance. The method was applied and evaluated on a land cover segmentation problem involving seven classes: buildings,roads, parking lots, fields, trees, empty spaces, and areas with unspecified labels, using images captured by UAVs. The performance of the ensemble model was evaluated by mean Intersection over Union (mIoU), and the results of comparing the proposed ensemble model with the three existing segmentation methods showed that mIoU performance was improved. Consequently, the study confirms that the proposed technique can enhance the performance of semantic segmentation models.

Quantitative Evaluation of Super-resolution Drone Images Generated Using Deep Learning (딥러닝을 이용하여 생성한 초해상화 드론 영상의 정량적 평가)

  • Seo, Hong-Deok;So, Hyeong-Yoon;Kim, Eui-Myoung
    • Journal of Cadastre & Land InformatiX
    • /
    • v.53 no.2
    • /
    • pp.5-18
    • /
    • 2023
  • As the development of drones and sensors accelerates, new services and values are created by fusing data acquired from various sensors mounted on drone. However, the construction of spatial information through data fusion is mainly constructed depending on the image, and the quality of data is determined according to the specification and performance of the hardware. In addition, it is difficult to utilize it in the actual field because expensive equipment is required to construct spatial information of high-quality. In this study, super-resolution was performed by applying deep learning to low-resolution images acquired through RGB and THM cameras mounted on a drone, and quantitative evaluation and feature point extraction were performed on the generated high-resolution images. As a result of the experiment, the high-resolution image generated by super-resolution was maintained the characteristics of the original image, and as the resolution was improved, more features could be extracted compared to the original image. Therefore, when generating a high-resolution image by applying a low-resolution image to an super-resolution deep learning model, it is judged to be a new method to construct spatial information of high-quality without being restricted by hardware.

Deep Learning-Assisted Diagnosis of Pediatric Skull Fractures on Plain Radiographs

  • Jae Won Choi;Yeon Jin Cho;Ji Young Ha;Yun Young Lee;Seok Young Koh;June Young Seo;Young Hun Choi;Jung-Eun Cheon;Ji Hoon Phi;Injoon Kim;Jaekwang Yang;Woo Sun Kim
    • Korean Journal of Radiology
    • /
    • v.23 no.3
    • /
    • pp.343-354
    • /
    • 2022
  • Objective: To develop and evaluate a deep learning-based artificial intelligence (AI) model for detecting skull fractures on plain radiographs in children. Materials and Methods: This retrospective multi-center study consisted of a development dataset acquired from two hospitals (n = 149 and 264) and an external test set (n = 95) from a third hospital. Datasets included children with head trauma who underwent both skull radiography and cranial computed tomography (CT). The development dataset was split into training, tuning, and internal test sets in a ratio of 7:1:2. The reference standard for skull fracture was cranial CT. Two radiology residents, a pediatric radiologist, and two emergency physicians participated in a two-session observer study on an external test set with and without AI assistance. We obtained the area under the receiver operating characteristic curve (AUROC), sensitivity, and specificity along with their 95% confidence intervals (CIs). Results: The AI model showed an AUROC of 0.922 (95% CI, 0.842-0.969) in the internal test set and 0.870 (95% CI, 0.785-0.930) in the external test set. The model had a sensitivity of 81.1% (95% CI, 64.8%-92.0%) and specificity of 91.3% (95% CI, 79.2%-97.6%) for the internal test set and 78.9% (95% CI, 54.4%-93.9%) and 88.2% (95% CI, 78.7%-94.4%), respectively, for the external test set. With the model's assistance, significant AUROC improvement was observed in radiology residents (pooled results) and emergency physicians (pooled results) with the difference from reading without AI assistance of 0.094 (95% CI, 0.020-0.168; p = 0.012) and 0.069 (95% CI, 0.002-0.136; p = 0.043), respectively, but not in the pediatric radiologist with the difference of 0.008 (95% CI, -0.074-0.090; p = 0.850). Conclusion: A deep learning-based AI model improved the performance of inexperienced radiologists and emergency physicians in diagnosing pediatric skull fractures on plain radiographs.

Adaptation of Deep Learning Image Reconstruction for Pediatric Head CT: A Focus on the Image Quality (소아용 두부 컴퓨터단층촬영에서 딥러닝 영상 재구성 적용: 영상 품질에 대한 고찰)

  • Nim Lee;Hyun-Hae Cho;So Mi Lee;Sun Kyoung You
    • Journal of the Korean Society of Radiology
    • /
    • v.84 no.1
    • /
    • pp.240-252
    • /
    • 2023
  • Purpose To assess the effect of deep learning image reconstruction (DLIR) for head CT in pediatric patients. Materials and Methods We collected 126 pediatric head CT images, which were reconstructed using filtered back projection, iterative reconstruction using adaptive statistical iterative reconstruction (ASiR)-V, and all three levels of DLIR (TrueFidelity; GE Healthcare). Each image set group was divided into four subgroups according to the patients' ages. Clinical and dose-related data were reviewed. Quantitative parameters, including the signal-to-noise ratio (SNR) and contrast-to-noise ratio (CNR), and qualitative parameters, including noise, gray matter-white matter (GM-WM) differentiation, sharpness, artifact, acceptability, and unfamiliar texture change were evaluated and compared. Results The SNR and CNR of each level in each age group increased among strength levels of DLIR. High-level DLIR showed a significantly improved SNR and CNR (p < 0.05). Sequential reduction of noise, improvement of GM-WM differentiation, and improvement of sharpness was noted among strength levels of DLIR. Those of high-level DLIR showed a similar value as that with ASiR-V. Artifact and acceptability did not show a significant difference among the adapted levels of DLIR. Conclusion Adaptation of high-level DLIR for the pediatric head CT can significantly reduce image noise. Modification is needed while processing artifacts.

Blended-Transfer Learning for Compressed-Sensing Cardiac CINE MRI

  • Park, Seong Jae;Ahn, Chang-Beom
    • Investigative Magnetic Resonance Imaging
    • /
    • v.25 no.1
    • /
    • pp.10-22
    • /
    • 2021
  • Purpose: To overcome the difficulty in building a large data set with a high-quality in medical imaging, a concept of 'blended-transfer learning' (BTL) using a combination of both source data and target data is proposed for the target task. Materials and Methods: Source and target tasks were defined as training of the source and target networks to reconstruct cardiac CINE images from undersampled data, respectively. In transfer learning (TL), the entire neural network (NN) or some parts of the NN after conducting a source task using an open data set was adopted in the target network as the initial network to improve the learning speed and the performance of the target task. Using BTL, an NN effectively learned the target data while preserving knowledge from the source data to the maximum extent possible. The ratio of the source data to the target data was reduced stepwise from 1 in the initial stage to 0 in the final stage. Results: NN that performed BTL showed an improved performance compared to those that performed TL or standalone learning (SL). Generalization of NN was also better achieved. The learning curve was evaluated using normalized mean square error (NMSE) of reconstructed images for both target data and source data. BTL reduced the learning time by 1.25 to 100 times and provided better image quality. Its NMSE was 3% to 8% lower than with SL. Conclusion: The NN that performed the proposed BTL showed the best performance in terms of learning speed and learning curve. It also showed the highest reconstructed-image quality with the lowest NMSE for the test data set. Thus, BTL is an effective way of learning for NNs in the medical-imaging domain where both quality and quantity of data are always limited.

Enhancement of durability of tall buildings by using deep-learning-based predictions of wind-induced pressure

  • K.R. Sri Preethaa;N. Yuvaraj;Gitanjali Wadhwa;Sujeen Song;Se-Woon Choi;Bubryur Kim
    • Wind and Structures
    • /
    • v.36 no.4
    • /
    • pp.237-247
    • /
    • 2023
  • The emergence of high-rise buildings has necessitated frequent structural health monitoring and maintenance for safety reasons. Wind causes damage and structural changes on tall structures; thus, safe structures should be designed. The pressure developed on tall buildings has been utilized in previous research studies to assess the impacts of wind on structures. The wind tunnel test is a primary research method commonly used to quantify the aerodynamic characteristics of high-rise buildings. Wind pressure is measured by placing pressure sensor taps at different locations on tall buildings, and the collected data are used for analysis. However, sensors may malfunction and produce erroneous data; these data losses make it difficult to analyze aerodynamic properties. Therefore, it is essential to generate missing data relative to the original data obtained from neighboring pressure sensor taps at various intervals. This study proposes a deep learning-based, deep convolutional generative adversarial network (DCGAN) to restore missing data associated with faulty pressure sensors installed on high-rise buildings. The performance of the proposed DCGAN is validated by using a standard imputation model known as the generative adversarial imputation network (GAIN). The average mean-square error (AMSE) and average R-squared (ARSE) are used as performance metrics. The calculated ARSE values by DCGAN on the building model's front, backside, left, and right sides are 0.970, 0.972, 0.984 and 0.978, respectively. The AMSE produced by DCGAN on four sides of the building model is 0.008, 0.010, 0.015 and 0.014. The average standard deviation of the actual measures of the pressure sensors on four sides of the model were 0.1738, 0.1758, 0.2234 and 0.2278. The average standard deviation of the pressure values generated by the proposed DCGAN imputation model was closer to that of the measured actual with values of 0.1736,0.1746,0.2191, and 0.2239 on four sides, respectively. In comparison, the standard deviation of the values predicted by GAIN are 0.1726,0.1735,0.2161, and 0.2209, which is far from actual values. The results demonstrate that DCGAN model fits better for data imputation than the GAIN model with improved accuracy and fewer error rates. Additionally, the DCGAN is utilized to estimate the wind pressure in regions of buildings where no pressure sensor taps are available; the model yielded greater prediction accuracy than GAIN.

The Prediction of Durability Performance for Chloride Ingress in Fly Ash Concrete by Artificial Neural Network Algorithm (인공 신경망 알고리즘을 활용한 플라이애시 콘크리트의 염해 내구성능 예측)

  • Kwon, Seung-Jun;Yoon, Yong-Sik
    • Journal of the Korea institute for structural maintenance and inspection
    • /
    • v.26 no.5
    • /
    • pp.127-134
    • /
    • 2022
  • In this study, RCPTs (Rapid Chloride Penetration Test) were performed for fly ash concrete with curing age of 4 ~ 6 years. The concrete mixtures were prepared with 3 levels of water to binder ratio (0.37, 0.42, and 0.47) and 2 levels of substitution ratio of fly ash (0 and 30%), and the improved passed charges of chloride ion behavior were quantitatively analyzed. Additionally, the results were trained through the univariate time series models consisted of GRU (Gated Recurrent Unit) algorithm and those from the models were evaluated. As the result of the RCPT, fly ash concrete showed the reduced passed charges with period and an more improved resistance to chloride penetration than OPC concrete. At the final evaluation period (6 years), fly ash concrete showed 'Very low' grade in all W/B (water to binder) ratio, however OPC concrete showed 'Moderate' grade in the condition with the highest W/B ratio (0.47). The adopted algorithm of GRU for this study can analyze time series data and has the advantage like operation efficiency. The deep learning model with 4 hidden layers was designed, and it provided a reasonable prediction results of passed charge. The deep learning model from this study has a limitation of single consideration of a univariate time series characteristic, but it is in the developing process of providing various characteristics of concrete like strength and diffusion coefficient through additional studies.