• Title/Summary/Keyword: Feature Augmentation

Search Result 46, Processing Time 0.027 seconds

Analysis of Small reservoir system by Flood control ability augmentation (치수능력 증대에 따른 저수지시스템 분석)

  • Park Ki-Bum;Lee Soon-Tak
    • Journal of Environmental Science International
    • /
    • v.14 no.11
    • /
    • pp.995-1004
    • /
    • 2005
  • As a research establish reservoir safety operation for small dam systems. This study presents hydrologic analysis conducted in the Duckdong and Bomun dam watershed based on various rainfall data and increase inflow. Especially the Duckdong dam without flood control feature are widely exposed to the risk of flooding, thus it is constructed emergency gate at present. In this study reservoir routing program was simulation for basin runoff estimating using HEC-HMS model, the model simulation the reservoir condition of emergency Sate with and without. At the reservoir analysis results is the Duckdong dam average storage decrease $20\%$ with emergency gate than without emergency gate. Also, the Bomun dam is not affected by the Duckdong flood control augmentation.

A Deep Learning Approach for Classification of Cloud Image Patches on Small Datasets

  • Phung, Van Hiep;Rhee, Eun Joo
    • Journal of information and communication convergence engineering
    • /
    • v.16 no.3
    • /
    • pp.173-178
    • /
    • 2018
  • Accurate classification of cloud images is a challenging task. Almost all the existing methods rely on hand-crafted feature extraction. Their limitation is low discriminative power. In the recent years, deep learning with convolution neural networks (CNNs), which can auto extract features, has achieved promising results in many computer vision and image understanding fields. However, deep learning approaches usually need large datasets. This paper proposes a deep learning approach for classification of cloud image patches on small datasets. First, we design a suitable deep learning model for small datasets using a CNN, and then we apply data augmentation and dropout regularization techniques to increase the generalization of the model. The experiments for the proposed approach were performed on SWIMCAT small dataset with k-fold cross-validation. The experimental results demonstrated perfect classification accuracy for most classes on every fold, and confirmed both the high accuracy and the robustness of the proposed model.

Facial Expression Classification Using Deep Convolutional Neural Network

  • Choi, In-kyu;Ahn, Ha-eun;Yoo, Jisang
    • Journal of Electrical Engineering and Technology
    • /
    • v.13 no.1
    • /
    • pp.485-492
    • /
    • 2018
  • In this paper, we propose facial expression recognition using CNN (Convolutional Neural Network), one of the deep learning technologies. The proposed structure has general classification performance for any environment or subject. For this purpose, we collect a variety of databases and organize the database into six expression classes such as 'expressionless', 'happy', 'sad', 'angry', 'surprised' and 'disgusted'. Pre-processing and data augmentation techniques are applied to improve training efficiency and classification performance. In the existing CNN structure, the optimal structure that best expresses the features of six facial expressions is found by adjusting the number of feature maps of the convolutional layer and the number of nodes of fully-connected layer. The experimental results show good classification performance compared to the state-of-the-arts in experiments of the cross validation and the cross database. Also, compared to other conventional models, it is confirmed that the proposed structure is superior in classification performance with less execution time.

Assembling three one-camera images for three-camera intersection classification

  • Marcella Astrid;Seung-Ik Lee
    • ETRI Journal
    • /
    • v.45 no.5
    • /
    • pp.862-873
    • /
    • 2023
  • Determining whether an autonomous self-driving agent is in the middle of an intersection can be extremely difficult when relying on visual input taken from a single camera. In such a problem setting, a wider range of views is essential, which drives us to use three cameras positioned in the front, left, and right of an agent for better intersection recognition. However, collecting adequate training data with three cameras poses several practical difficulties; hence, we propose using data collected from one camera to train a three-camera model, which would enable us to more easily compile a variety of training data to endow our model with improved generalizability. In this work, we provide three separate fusion methods (feature, early, and late) of combining the information from three cameras. Extensive pedestrian-view intersection classification experiments show that our feature fusion model provides an area under the curve and F1-score of 82.00 and 46.48, respectively, which considerably outperforms contemporary three- and one-camera models.

Video Augmentation of Virtual Object by Uncalibrated 3D Reconstruction from Video Frames (비디오 영상에서의 비보정 3차원 좌표 복원을 통한 가상 객체의 비디오 합성)

  • Park Jong-Seung;Sung Mee-Young
    • Journal of Korea Multimedia Society
    • /
    • v.9 no.4
    • /
    • pp.421-433
    • /
    • 2006
  • This paper proposes a method to insert virtual objects into a real video stream based on feature tracking and camera pose estimation from a set of single-camera video frames. To insert or modify 3D shapes to target video frames, the transformation from the 3D objects to the projection of the objects onto the video frames should be revealed. It is shown that, without a camera calibration process, the 3D reconstruction is possible using multiple images from a single camera under the fixed internal camera parameters. The proposed approach is based on the simplification of the camera matrix of intrinsic parameters and the use of projective geometry. The method is particularly useful for augmented reality applications to insert or modify models to a real video stream. The proposed method is based on a linear parameter estimation approach for the auto-calibration step and it enhances the stability and reduces the execution time. Several experimental results are presented on real-world video streams, demonstrating the usefulness of our method for the augmented reality applications.

  • PDF

Electrical Arc Detection using Convolutional Neural Network (합성곱 신경망을 이용한 전기 아크 신호 검출)

  • Lee, Sangik;Kang, Seokwoo;Kim, Taewon;Kim, Manbae
    • Journal of Broadcast Engineering
    • /
    • v.25 no.4
    • /
    • pp.569-575
    • /
    • 2020
  • The serial arc is one of factors causing electrical fires. Over past decades, various researches have been carried out to detect arc occurrences. Even though frequency analysis, wavelet, and statistical features have been used, additional steps such as transformation and feature extraction are required. On the contrary, deep learning models directly use the raw data without any feature extraction processes. Therefore, the usage of time-domain data is preferred, but the performance is not satisfactory. To solve this problem, subsequent 1-D signals are transformed into 2-D data that can feed into a convolutional neural network (CNN). Experiments validated that CNN model outperforms deep neural network (DNN) by the classification accuracy of 8.6%. In addition, data augmentation is utilized, resulting in the accuracy improvement by 14%.

Design of Pattern Array Method for Multi Data Augmentation of Power Equipment uisng Single Image Pattern (단일 이미지 패턴을 이용한 다수의 전력설비 데이터를 증강하기 위한 패턴 배열화 기법 설계)

  • Kim, Seoksoo
    • Journal of Convergence for Information Technology
    • /
    • v.10 no.11
    • /
    • pp.1-8
    • /
    • 2020
  • As power consumption is maximized, research on augmented reality-based monitoring systems for on-site facility managers to maintain and repair power facilities is being actively conducted as individual power brokerages and power production facilities increase. However, in the case of existing augmented reality-based monitoring systems, it is difficult to accurately detect patterns due to problems such as external environment, facility complexity, and interference with the lighting environment, and it is not possible to match various sensing information and service information for power facilities to one pattern. there is a problem. For this reason, since sensor information is matched using a single image pattern for each sensor of a power facility, a plurality of image patterns are required to augment and provide all information. In this paper, we propose a single image pattern arrangement method that matches and provides a plurality of information through an array combination of feature patterns in a single image composed of a plurality of feature patterns.

Constructing for Korean Traditional culture Corpus and Development of Named Entity Recognition Model using Bi-LSTM-CNN-CRFs (한국 전통문화 말뭉치구축 및 Bi-LSTM-CNN-CRF를 활용한 전통문화 개체명 인식 모델 개발)

  • Kim, GyeongMin;Kim, Kuekyeng;Jo, Jaechoon;Lim, HeuiSeok
    • Journal of the Korea Convergence Society
    • /
    • v.9 no.12
    • /
    • pp.47-52
    • /
    • 2018
  • Named Entity Recognition is a system that extracts entity names such as Persons(PS), Locations(LC), and Organizations(OG) that can have a unique meaning from a document and determines the categories of extracted entity names. Recently, Bi-LSTM-CRF, which is a combination of CRF using the transition probability between output data from LSTM-based Bi-LSTM model considering forward and backward directions of input data, showed excellent performance in the study of object name recognition using deep-learning, and it has a good performance on the efficient embedding vector creation by character and word unit and the model using CNN and LSTM. In this research, we describe the Bi-LSTM-CNN-CRF model that enhances the features of the Korean named entity recognition system and propose a method for constructing the traditional culture corpus. We also present the results of learning the constructed corpus with the feature augmentation model for the recognition of Korean object names.

Automatic proficiency assessment of Korean speech read aloud by non-natives using bidirectional LSTM-based speech recognition

  • Oh, Yoo Rhee;Park, Kiyoung;Jeon, Hyung-Bae;Park, Jeon Gue
    • ETRI Journal
    • /
    • v.42 no.5
    • /
    • pp.761-772
    • /
    • 2020
  • This paper presents an automatic proficiency assessment method for a non-native Korean read utterance using bidirectional long short-term memory (BLSTM)-based acoustic models (AMs) and speech data augmentation techniques. Specifically, the proposed method considers two scenarios, with and without prompted text. The proposed method with the prompted text performs (a) a speech feature extraction step, (b) a forced-alignment step using a native AM and non-native AM, and (c) a linear regression-based proficiency scoring step for the five proficiency scores. Meanwhile, the proposed method without the prompted text additionally performs Korean speech recognition and a subword un-segmentation for the missing text. The experimental results indicate that the proposed method with prompted text improves the performance for all scores when compared to a method employing conventional AMs. In addition, the proposed method without the prompted text has a fluency score performance comparable to that of the method with prompted text.

Subsurface anomaly detection utilizing synthetic GPR images and deep learning model

  • Ahmad Abdelmawla;Shihan Ma;Jidong J. Yang;S. Sonny Kim
    • Geomechanics and Engineering
    • /
    • v.33 no.2
    • /
    • pp.203-209
    • /
    • 2023
  • One major advantage of ground penetrating radar (GPR) over other field test methods is its ability to obtain subsurface images of roads in an efficient and non-intrusive manner. Not only can the strata of pavement structure be retrieved from the GPR scan images, but also various irregularities, such as cracks and internal cavities. This article introduces a deep learning-based approach, focusing on detecting subsurface cracks by recognizing their distinctive hyperbolic signatures in the GPR scan images. Given the limited road sections that contain target features, two data augmentation methods, i.e., feature insertion and generation, are implemented, resulting in 9,174 GPR scan images. One of the most popular real-time object detection models, You Only Learn One Representation (YOLOR), is trained for detecting the target features for two types of subsurface cracks: bottom cracks and full cracks from the GPR scan images. The former represents partial cracks initiated from the bottom of the asphalt layer or base layers, while the latter includes extended cracks that penetrate these layers. Our experiments show the test average precisions of 0.769, 0.803 and 0.735 for all cracks, bottom cracks, and full cracks, respectively. This demonstrates the practicality of deep learning-based methods in detecting subsurface cracks from GPR scan images.