• Title/Summary/Keyword: Deep Learning System

Search Result 1,745, Processing Time 0.034 seconds

Self-Supervised Long-Short Term Memory Network for Solving Complex Job Shop Scheduling Problem

  • Shao, Xiaorui;Kim, Chang Soo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.8
    • /
    • pp.2993-3010
    • /
    • 2021
  • The job shop scheduling problem (JSSP) plays a critical role in smart manufacturing, an effective JSSP scheduler could save time cost and increase productivity. Conventional methods are very time-consumption and cannot deal with complicated JSSP instances as it uses one optimal algorithm to solve JSSP. This paper proposes an effective scheduler based on deep learning technology named self-supervised long-short term memory (SS-LSTM) to handle complex JSSP accurately. First, using the optimal method to generate sufficient training samples in small-scale JSSP. SS-LSTM is then applied to extract rich feature representations from generated training samples and decide the next action. In the proposed SS-LSTM, two channels are employed to reflect the full production statues. Specifically, the detailed-level channel records 18 detailed product information while the system-level channel reflects the type of whole system states identified by the k-means algorithm. Moreover, adopting a self-supervised mechanism with LSTM autoencoder to keep high feature extraction capacity simultaneously ensuring the reliable feature representative ability. The authors implemented, trained, and compared the proposed method with the other leading learning-based methods on some complicated JSSP instances. The experimental results have confirmed the effectiveness and priority of the proposed method for solving complex JSSP instances in terms of make-span.

Attack Path and Intention Recognition System for detecting APT Attack (APT 공격 탐지를 위한 공격 경로 및 의도 인지 시스템)

  • Kim, Namuk;Eom, Jungho
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.16 no.1
    • /
    • pp.67-78
    • /
    • 2020
  • Typical security solutions such as intrusion detection system are not suitable for detecting advanced persistent attack(APT), because they cannot draw the big picture from trivial events of security solutions. Researches on techniques for detecting multiple stage attacks by analyzing the correlations between security events or alerts are being actively conducted in academic field. However, these studies still use events from existing security system, and there is insufficient research on the structure of the entire security system suitable for advanced persistent attacks. In this paper, we propose an attack path and intention recognition system suitable for multiple stage attacks like advanced persistent attack detection. The proposed system defines the trace format and overall structure of the system that detects APT attacks based on the correlation and behavior analysis, and is designed with a structure of detection system using deep learning and big data technology, etc.

IoT-Based Health Big-Data Process Technologies: A Survey

  • Yoo, Hyun;Park, Roy C.;Chung, Kyungyong
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.3
    • /
    • pp.974-992
    • /
    • 2021
  • Recently, the healthcare field has undergone rapid changes owing to the accumulation of health big data and the development of machine learning. Data mining research in the field of healthcare has different characteristics from those of other data analyses, such as the structural complexity of the medical data, requirement for medical expertise, and security of personal medical information. Various methods have been implemented to address these issues, including the machine learning model and cloud platform. However, the machine learning model presents the problem of opaque result interpretation, and the cloud platform requires more in-depth research on security and efficiency. To address these issues, this paper presents a recent technology for Internet-of-Things-based (IoT-based) health big data processing. We present a cloud-based IoT health platform and health big data processing technology that reduces the medical data management costs and enhances safety. We also present a data mining technology for health-risk prediction, which is the core of healthcare. Finally, we propose a study using explainable artificial intelligence that enhances the reliability and transparency of the decision-making system, which is called the black box model owing to its lack of transparency.

Tongue Image Segmentation Using CNN and Various Image Augmentation Techniques (콘볼루션 신경망(CNN)과 다양한 이미지 증강기법을 이용한 혀 영역 분할)

  • Ahn, Ilkoo;Bae, Kwang-Ho;Lee, Siwoo
    • Journal of Biomedical Engineering Research
    • /
    • v.42 no.5
    • /
    • pp.201-210
    • /
    • 2021
  • In Korean medicine, tongue diagnosis is one of the important diagnostic methods for diagnosing abnormalities in the body. Representative features that are used in the tongue diagnosis include color, shape, texture, cracks, and tooth marks. When diagnosing a patient through these features, the diagnosis criteria may be different for each oriental medical doctor, and even the same person may have different diagnosis results depending on time and work environment. In order to overcome this problem, recent studies to automate and standardize tongue diagnosis using machine learning are continuing and the basic process of such a machine learning-based tongue diagnosis system is tongue segmentation. In this paper, image data is augmented based on the main tongue features, and backbones of various famous deep learning architecture models are used for automatic tongue segmentation. The experimental results show that the proposed augmentation technique improves the accuracy of tongue segmentation, and that automatic tongue segmentation can be performed with a high accuracy of 99.12%.

Automated detection of panic disorder based on multimodal physiological signals using machine learning

  • Eun Hye Jang;Kwan Woo Choi;Ah Young Kim;Han Young Yu;Hong Jin Jeon;Sangwon Byun
    • ETRI Journal
    • /
    • v.45 no.1
    • /
    • pp.105-118
    • /
    • 2023
  • We tested the feasibility of automated discrimination of patients with panic disorder (PD) from healthy controls (HCs) based on multimodal physiological responses using machine learning. Electrocardiogram (ECG), electrodermal activity (EDA), respiration (RESP), and peripheral temperature (PT) of the participants were measured during three experimental phases: rest, stress, and recovery. Eleven physiological features were extracted from each phase and used as input data. Logistic regression (LoR), k-nearest neighbor (KNN), support vector machine (SVM), random forest (RF), and multilayer perceptron (MLP) algorithms were implemented with nested cross-validation. Linear regression analysis showed that ECG and PT features obtained in the stress and recovery phases were significant predictors of PD. We achieved the highest accuracy (75.61%) with MLP using all 33 features. With the exception of MLP, applying the significant predictors led to a higher accuracy than using 24 ECG features. These results suggest that combining multimodal physiological signals measured during various states of autonomic arousal has the potential to differentiate patients with PD from HCs.

Machine Learning Based BLE Indoor Positioning Performance Improvement (머신러닝 기반 BLE 실내측위 성능 개선)

  • Moon, Joon;Pak, Sang-Hyon;Hwang, Jae-Jeong
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.467-468
    • /
    • 2021
  • In order to improve the performance of the indoor positioning system using BLE beacons, a receiver that measures the angle of arrival among the direction finding technologies supported by BLE5.1 was manufactured and analyzed by machine learning to measure the optimal position. For the creation and testing of machine learning models, k-nearest neighbor classification and regression, logistic regression, support vector machines, decision tree artificial neural networks, and deep neural networks were used to learn and test. As a result, when the test set 4 produced in the study was used, the accuracy was up to 99%.

  • PDF

Prediction of Process-Induced Spring-Back of CFRP Composite Structure Using Deep Neural Network (심층신경망을 이용한 CFRP 복합재 구조의 공정 유도 스프링백 예측)

  • Yuseon Lee;Dong-Hyeop Kim;Sang-Woo Kim;Soo-Yong Lee
    • Journal of Aerospace System Engineering
    • /
    • v.18 no.5
    • /
    • pp.73-80
    • /
    • 2024
  • A deep neural network (DNN) was employed to predict the spring-back of a CFRP composite spar induced by the curing process. A total of 816 spring-back data points, derived from varying stacking angles, layer counts, and flange radii, were generated through finite element method (FEM)-based curing analysis to train the DNN model. The trained model demonstrated an R-squared value of 0.99 and a mean squared error of 0.00093, indicating excellent performance. For untrained flange radii, the spring-back predicted by the DNN exhibited a mean relative error of 2.18% when compared to FEM results. Additionally, while FEM analysis required approximately 20 minutes, the DNN-based prediction required only about 14 milliseconds. These results highlight the potential of using DNNs for the rapid prediction of process-induced deformation in CFRP composites.

A Study on Lightweight CNN-based Interpolation Method for Satellite Images (위성 영상을 위한 경량화된 CNN 기반의 보간 기술 연구)

  • Kim, Hyun-ho;Seo, Doochun;Jung, JaeHeon;Kim, Yongwoo
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.2
    • /
    • pp.167-177
    • /
    • 2022
  • In order to obtain satellite image products using the image transmitted to the ground station after capturing the satellite images, many image pre/post-processing steps are involved. During the pre/post-processing, when converting from level 1R images to level 1G images, geometric correction is essential. An interpolation method necessary for geometric correction is inevitably used, and the quality of the level 1G images is determined according to the accuracy of the interpolation method. Also, it is crucial to speed up the interpolation algorithm by the level processor. In this paper, we proposed a lightweight CNN-based interpolation method required for geometric correction when converting from level 1R to level 1G. The proposed method doubles the resolution of satellite images and constructs a deep learning network with a lightweight deep convolutional neural network for fast processing speed. In addition, a feature map fusion method capable of improving the image quality of multispectral (MS) bands using panchromatic (PAN) band information was proposed. The images obtained through the proposed interpolation method improved by about 0.4 dB for the PAN image and about 4.9 dB for the MS image in the quantitative peak signal-to-noise ratio (PSNR) index compared to the existing deep learning-based interpolation methods. In addition, it was confirmed that the time required to acquire an image that is twice the resolution of the 36,500×36,500 input image based on the PAN image size is improved by about 1.6 times compared to the existing deep learning-based interpolation method.

Grade Analysis and Two-Stage Evaluation of Beef Carcass Image Using Deep Learning (딥러닝을 이용한 소도체 영상의 등급 분석 및 단계별 평가)

  • Kim, Kyung-Nam;Kim, Seon-Jong
    • The Journal of the Convergence on Culture Technology
    • /
    • v.8 no.2
    • /
    • pp.385-391
    • /
    • 2022
  • Quality evaluation of beef carcasses is an important issue in the livestock industry. Recently, through the AI monitor system based on artificial intelligence, the quality manager can receive help in making accurate decisions based on the analysis of beef carcass images or result information. This artificial intelligence dataset is an important factor in judging performance. Existing datasets may have different surface orientation or resolution. In this paper, we proposed a two-stage classification model that can efficiently manage the grades of beef carcass image using deep learning. And to overcome the problem of the various conditions of the image, a new dataset of 1,300 images was constructed. The recognition rate of deep network for 5-grade classification using the new dataset was 72.5%. Two-stage evaluation is a method to increase reliability by taking advantage of the large difference between grades 1++, 1+, and grades 1 and 2 and 3. With two experiments using the proposed two stage model, the recognition rates of 73.7% and 77.2% were obtained. As this, The proposed method will be an efficient method if we have a dataset with 100% recognition rate in the first stage.

River Water Level Prediction Method based on LSTM Neural Network

  • Le, Xuan Hien;Lee, Giha
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2018.05a
    • /
    • pp.147-147
    • /
    • 2018
  • In this article, we use an open source software library: TensorFlow, developed for the purposes of conducting very complex machine learning and deep neural network applications. However, the system is general enough to be applicable in a wide variety of other domains as well. The proposed model based on a deep neural network model, LSTM (Long Short-Term Memory) to predict the river water level at Okcheon Station of the Guem River without utilization of rainfall - forecast information. For LSTM modeling, the input data is hourly water level data for 15 years from 2002 to 2016 at 4 stations includes 3 upstream stations (Sutong, Hotan, and Songcheon) and the forecasting-target station (Okcheon). The data are subdivided into three purposes: a training data set, a testing data set and a validation data set. The model was formulated to predict Okcheon Station water level for many cases from 3 hours to 12 hours of lead time. Although the model does not require many input data such as climate, geography, land-use for rainfall-runoff simulation, the prediction is very stable and reliable up to 9 hours of lead time with the Nash - Sutcliffe efficiency (NSE) is higher than 0.90 and the root mean square error (RMSE) is lower than 12cm. The result indicated that the method is able to produce the river water level time series and be applicable to the practical flood forecasting instead of hydrologic modeling approaches.

  • PDF