• Title/Summary/Keyword: Image Augmentation

Search Result 213, Processing Time 0.022 seconds

Measurement of Construction Material Quantity through Analyzing Images Acquired by Drone And Data Augmentation (드론 영상 분석과 자료 증가 방법을 통한 건설 자재 수량 측정)

  • Moon, Ji-Hwan;Song, Nu-Lee;Choi, Jae-Gab;Park, Jin-Ho;Kim, Gye-Young
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.9 no.1
    • /
    • pp.33-38
    • /
    • 2020
  • This paper proposes a technique for counting construction materials by analyzing an image acquired by a Drone. The proposed technique use drone log which includes drone and camera information, RCNN for predicting construction material type, dummy area and Photogrammetry for counting the number of construction material. The existing research has large error ranges for predicting construction material detection and material dummy area, because of a lack of training data. To reduce the error ranges and improve prediction stability, this paper increases the training data with a method of data augmentation, but only uses rotated training data for data augmentation to prevent overfitting of the training model. For the quantity calculation, we use a drone log containing drones and camera information such as Yaw and FOV, RCNN model to find the pile of building materials in the image and to predict the type. And we synthesize all the information and apply it to the formula suggested in the paper to calculate the actual quantity of material pile. The superiority of the proposed method is demonstrated through experiments.

Evaluation of Sentimental Texts Automatically Generated by a Generative Adversarial Network (생성적 적대 네트워크로 자동 생성한 감성 텍스트의 성능 평가)

  • Park, Cheon-Young;Choi, Yong-Seok;Lee, Kong Joo
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.8 no.6
    • /
    • pp.257-264
    • /
    • 2019
  • Recently, deep neural network based approaches have shown a good performance for various fields of natural language processing. A huge amount of training data is essential for building a deep neural network model. However, collecting a large size of training data is a costly and time-consuming job. A data augmentation is one of the solutions to this problem. The data augmentation of text data is more difficult than that of image data because texts consist of tokens with discrete values. Generative adversarial networks (GANs) are widely used for image generation. In this work, we generate sentimental texts by using one of the GANs, CS-GAN model that has a discriminator as well as a classifier. We evaluate the usefulness of generated sentimental texts according to various measurements. CS-GAN model not only can generate texts with more diversity but also can improve the performance of its classifier.

Deep Learning Based On-Device Augmented Reality System using Multiple Images (다중영상을 이용한 딥러닝 기반 온디바이스 증강현실 시스템)

  • Jeong, Taehyeon;Park, In Kyu
    • Journal of Broadcast Engineering
    • /
    • v.27 no.3
    • /
    • pp.341-350
    • /
    • 2022
  • In this paper, we propose a deep learning based on-device augmented reality (AR) system in which multiple input images are used to implement the correct occlusion in a real environment. The proposed system is composed of three technical steps; camera pose estimation, depth estimation, and object augmentation. Each step employs various mobile frameworks to optimize the processing on the on-device environment. Firstly, in the camera pose estimation stage, the massive computation involved in feature extraction is parallelized using OpenCL which is the GPU parallelization framework. Next, in depth estimation, monocular and multiple image-based depth image inference is accelerated using the mobile deep learning framework, i.e. TensorFlow Lite. Finally, object augmentation and occlusion handling are performed on the OpenGL ES mobile graphics framework. The proposed augmented reality system is implemented as an application in the Android environment. We evaluate the performance of the proposed system in terms of augmentation accuracy and the processing time in the mobile as well as PC environments.

Deep Learning-based Pixel-level Concrete Wall Crack Detection Method (딥러닝 기반 픽셀 단위 콘크리트 벽체 균열 검출 방법)

  • Kang, Kyung-Su;Ryu, Han-Guk
    • Journal of the Korea Institute of Building Construction
    • /
    • v.23 no.2
    • /
    • pp.197-207
    • /
    • 2023
  • Concrete is a widely used material due to its excellent compressive strength and durability. However, depending on the surrounding environment and the characteristics of the materials used in the construction, various defects may occur, such as cracks on the surface and subsidence of the structure. The detects on the surface of the concrete structure occur after completion or over time. Neglecting these cracks may lead to severe structural damage, necessitating regular safety inspections. Traditional visual inspections of concrete walls are labor-intensive and expensive. This research presents a deep learning-based semantic segmentation model designed to detect cracks in concrete walls. The model addresses surface defects that arise from aging, and an image augmentation technique is employed to enhance feature extraction and generalization performance. A dataset for semantic segmentation was created by combining publicly available and self-generated datasets, and notable semantic segmentation models were evaluated and tested. The model, specifically trained for concrete wall fracture detection, achieved an extraction performance of 81.4%. Moreover, a 3% performance improvement was observed when applying the developed augmentation technique.

Classification of Leukemia Disease in Peripheral Blood Cell Images Using Convolutional Neural Network

  • Tran, Thanh;Park, Jin-Hyuk;Kwon, Oh-Heum;Moon, Kwang-Seok;Lee, Suk-Hwan;Kwon, Ki-Ryong
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.10
    • /
    • pp.1150-1161
    • /
    • 2018
  • Classification is widely used in medical images to categorize patients and non-patients. However, conventional classification requires a complex procedure, including some rigid steps such as pre-processing, segmentation, feature extraction, detection, and classification. In this paper, we propose a novel convolutional neural network (CNN), called LeukemiaNet, to specifically classify two different types of leukemia, including acute lymphoblastic leukemia (ALL) and acute myeloid leukemia (AML), and non-cancerous patients. To extend the limited dataset, a PCA color augmentation process is utilized before images are input into the LeukemiaNet. This augmentation method enhances the accuracy of our proposed CNN architecture from 96.9% to 97.2% for distinguishing ALL, AML, and normal cell images.

An Efficient Object Augmentation Scheme for Supporting Pervasiveness in a Mobile Augmented Reality

  • Jang, Sung-Bong;Ko, Young-Woong
    • Journal of Information Processing Systems
    • /
    • v.16 no.5
    • /
    • pp.1214-1222
    • /
    • 2020
  • Pervasive augmented reality (AR) technology can be used to efficiently search for the required information regarding products in stores through text augmentation in an Internet of Things (IoT) environment. The evolution of context awareness and image processing technologies are the main driving forces that realize this type of AR service. One of the problems to be addressed in the service is that augmented objects are fixed and cannot be replaced efficiently in real time. To address this problem, a real-time mobile AR framework is proposed. In this framework, an optimal object to be augmented is selected based on object similarity comparison, and the augmented objects are efficiently managed using distributed metadata servers to adapt to the user requirements, in a given situation. To evaluate the feasibility of the proposed framework, a prototype system was implemented, and a qualitative evaluation based on questionnaires was conducted. The experimental results show that the proposed framework provides a better user experience than existing features in smartphones, and through fast AR service, the users are able to conveniently obtain additional information on products or objects.

A Study on the Classification of Military Airplanes in Neighboring Countries Using Deep Learning and Various Data Augmentation Techniques (딥러닝과 다양한 데이터 증강 기법을 활용한 주변국 군용기 기종 분류에 관한 연구)

  • Chanwoo, Lee;Hajun, Hwang;Hyeok, Kwon;Seungryeong, Baik;Wooju, Kim
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.25 no.6
    • /
    • pp.572-579
    • /
    • 2022
  • The analysis of foreign aircraft appearing suddenly in air defense identification zones requires a lot of cost and time. This study aims to develop a pre-trained model that can identify neighboring military aircraft based on aircraft photographs available on the web and present a model that can determine which aircraft corresponds to based on aerial photographs taken by allies. The advantages of this model are to reduce the cost and time required for model classification by proposing a pre-trained model and to improve the performance of the classifier by data augmentation of edge-detected images, cropping, flipping and so on.

Video Augmentation by Image-based Rendering

  • Seo, Yong-Duek;Kim, Seung-Jin;Sang, Hong-Ki
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1998.06b
    • /
    • pp.147-153
    • /
    • 1998
  • This paper provides a method for video augmentation using image interpolation. In computer graphics or augmented reality, 3D information of a model object is necessary to generate 2D views of the model, which are then inserted into or overlayed on environmental views or real video frames. However, we do not require any three dimensional model but images of the model object at some locations to render views according to the motion of video camera which is calculated by an SFM algorithm using point matches under weak-perspective (scaled-orthographic) projection model. Thus, a linear view interpolation algorithm is applied rather than a 3D ray-tracing method to get a view of the model at different viewpoints from model views. In order to get novel views in a way that agrees with the camera motion the camera coordinate system is embedded into model coordinate system at initialization time on the basis of 3D information recovered from video images and model views, respectively. During the sequence, motion parameters from video frames are used to compute interpolation parameters, and rendered model views are overlayed on corresponding video frames. Experimental results for real video frames and model views are given. Finally, discussion on the limitations of the method and subjects for future research are provided.

  • PDF

Robust Deep Age Estimation Method Using Artificially Generated Image Set

  • Jang, Jaeyoon;Jeon, Seung-Hyuk;Kim, Jaehong;Yoon, Hosub
    • ETRI Journal
    • /
    • v.39 no.5
    • /
    • pp.643-651
    • /
    • 2017
  • Human age estimation is one of the key factors in the field of Human-Robot Interaction/Human-Computer Interaction (HRI/HCI). Owing to the development of deep-learning technologies, age recognition has recently been attempted. In general, however, deep learning techniques require a large-scale database, and for age learning with variations, a conventional database is insufficient. For this reason, we propose an age estimation method using artificially generated data. Image data are artificially generated through 3D information, thus solving the problem of shortage of training data, and helping with the training of the deep-learning technique. Augmentation using 3D has advantages over 2D because it creates new images with more information. We use a deep architecture as a pre-trained model, and improve the estimation capacity using artificially augmented training images. The deep architecture can outperform traditional estimation methods, and the improved method showed increased reliability. We have achieved state-of-the-art performance using the proposed method in the Morph-II dataset and have proven that the proposed method can be used effectively using the Adience dataset.

Study on the Surface Defect Classification of Al 6061 Extruded Material By Using CNN-Based Algorithms (CNN을 이용한 Al 6061 압출재의 표면 결함 분류 연구)

  • Kim, S.B.;Lee, K.A.
    • Transactions of Materials Processing
    • /
    • v.31 no.4
    • /
    • pp.229-239
    • /
    • 2022
  • Convolution Neural Network(CNN) is a class of deep learning algorithms and can be used for image analysis. In particular, it has excellent performance in finding the pattern of images. Therefore, CNN is commonly applied for recognizing, learning and classifying images. In this study, the surface defect classification performance of Al 6061 extruded material using CNN-based algorithms were compared and evaluated. First, the data collection criteria were suggested and a total of 2,024 datasets were prepared. And they were randomly classified into 1,417 learning data and 607 evaluation data. After that, the size and quality of the training data set were improved using data augmentation techniques to increase the performance of deep learning. The CNN-based algorithms used in this study were VGGNet-16, VGGNet-19, ResNet-50 and DenseNet-121. The evaluation of the defect classification performance was made by comparing the accuracy, loss, and learning speed using verification data. The DenseNet-121 algorithm showed better performance than other algorithms with an accuracy of 99.13% and a loss value of 0.037. This was due to the structural characteristics of the DenseNet model, and the information loss was reduced by acquiring information from all previous layers for image identification in this algorithm. Based on the above results, the possibility of machine vision application of CNN-based model for the surface defect classification of Al extruded materials was also discussed.