• Title/Summary/Keyword: Recurrent Training

Search Result 145, Processing Time 0.026 seconds

In-depth exploration of machine learning algorithms for predicting sidewall displacement in underground caverns

  • Hanan Samadi;Abed Alanazi;Sabih Hashim Muhodir;Shtwai Alsubai;Abdullah Alqahtani;Mehrez Marzougui
    • Geomechanics and Engineering
    • /
    • v.37 no.4
    • /
    • pp.307-321
    • /
    • 2024
  • This paper delves into the critical assessment of predicting sidewall displacement in underground caverns through the application of nine distinct machine learning techniques. The accurate prediction of sidewall displacement is essential for ensuring the structural safety and stability of underground caverns, which are prone to various geological challenges. The dataset utilized in this study comprises a total of 310 data points, each containing 13 relevant parameters extracted from 10 underground cavern projects located in Iran and other regions. To facilitate a comprehensive evaluation, the dataset is evenly divided into training and testing subset. The study employs a diverse array of machine learning models, including recurrent neural network, back-propagation neural network, K-nearest neighbors, normalized and ordinary radial basis function, support vector machine, weight estimation, feed-forward stepwise regression, and fuzzy inference system. These models are leveraged to develop predictive models that can accurately forecast sidewall displacement in underground caverns. The training phase involves utilizing 80% of the dataset (248 data points) to train the models, while the remaining 20% (62 data points) are used for testing and validation purposes. The findings of the study highlight the back-propagation neural network (BPNN) model as the most effective in providing accurate predictions. The BPNN model demonstrates a remarkably high correlation coefficient (R2 = 0.99) and a low error rate (RMSE = 4.27E-05), indicating its superior performance in predicting sidewall displacement in underground caverns. This research contributes valuable insights into the application of machine learning techniques for enhancing the safety and stability of underground structures.

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.

Load-deflection analysis prediction of CFRP strengthened RC slab using RNN

  • Razavi, S.V.;Jumaat, Mohad Zamin;El-Shafie, Ahmed H.;Ronagh, Hamid Reza
    • Advances in concrete construction
    • /
    • v.3 no.2
    • /
    • pp.91-102
    • /
    • 2015
  • In this paper, the load-deflection analysis of the Carbon Fiber Reinforced Polymer (CFRP) strengthened Reinforced Concrete (RC) slab using Recurrent Neural Network (RNN) is investigated. Six reinforced concrete slabs having dimension $1800{\times}400{\times}120mm$ with similar steel bar of 2T10 and strengthened using different length and width of CFRP were tested and compared with similar samples without CFRP. The experimental load-deflection results were normalized and then uploaded in MATLAB software. Loading, CFRP length and width were as neurons in input layer and mid-span deflection was as neuron in output layer. The network was generated using feed-forward network and a internal nonlinear condition space model to memorize the input data while training process. From 122 load-deflection data, 111 data utilized for network generation and 11 data for the network testing. The results of model on the testing stage showed that the generated RNN predicted the load-deflection analysis of the slabs in acceptable technique with a correlation of determination of 0.99. The ratio between predicted deflection by RNN and experimental output was in the range of 0.99 to 1.11.

Recognition of Characters Printed on PCB Components Using Deep Neural Networks (심층신경망을 이용한 PCB 부품의 인쇄문자 인식)

  • Cho, Tai-Hoon
    • Journal of the Semiconductor & Display Technology
    • /
    • v.20 no.3
    • /
    • pp.6-10
    • /
    • 2021
  • Recognition of characters printed or marked on the PCB components from images captured using cameras is an important task in PCB components inspection systems. Previous optical character recognition (OCR) of PCB components typically consists of two stages: character segmentation and classification of each segmented character. However, character segmentation often fails due to corrupted characters, low image contrast, etc. Thus, OCR without character segmentation is desirable and increasingly used via deep neural networks. Typical implementation based on deep neural nets without character segmentation includes convolutional neural network followed by recurrent neural network (RNN). However, one disadvantage of this approach is slow execution due to RNN layers. LPRNet is a segmentation-free character recognition network with excellent accuracy proved in license plate recognition. LPRNet uses a wide convolution instead of RNN, thus enabling fast inference. In this paper, LPRNet was adapted for recognizing characters printed on PCB components with fast execution and high accuracy. Initial training with synthetic images followed by fine-tuning on real text images yielded accurate recognition. This net can be further optimized on Intel CPU using OpenVINO tool kit. The optimized version of the network can be run in real-time faster than even GPU.

Induction of somatic embryogenesis from immature zygotic embryos and young apical leaves in cork oak (Quercus suber L.)

  • El Ansari, Zineb Nejjar;El Bouzdoudi, Brahim;Errabii, Tomader;Saidi, Rabah;El Kbiach, Mohammed L'bachir
    • Journal of Plant Biotechnology
    • /
    • v.48 no.1
    • /
    • pp.44-53
    • /
    • 2021
  • The present work aims to study the induction of somatic embryogenesis in cork oak (Quercus suber L.) from immature zygotic embryos and young apical leaves obtained from 2-month-old seedlings through acorn germination on sterilized peat. The immature zygotic embryos were grown for 1 month on the mineral solution of MS in the presence of 4.52 µM 2,4-D and 30 g/L sucrose. They were then transferred to the same mineral solution with no added growth regulators. In the third subculture, yellow somatic embryos, characterized by two voluminous cotyledons, were differentiated from the radicle of the immature zygotic embryos. The induction of somatic embryogenesis in young leaves required a series of transfers on different culture media containing 30 g/L sucrose and 100 mg/L myo-inositol. Secondary or recurrent somatic embryogenesis occurred within the immature somatic embryo radicles after 1 month of culture on growth regulator-free medium containing WPM macronutrients, MS micronutrients, and vitamins.

Prediction of Closed Quotient During Vocal Phonation using GRU-type Neural Network with Audio Signals

  • Hyeonbin Han;Keun Young Lee;Seong-Yoon Shin;Yoseup Kim;Gwanghyun Jo;Jihoon Park;Young-Min Kim
    • Journal of information and communication convergence engineering
    • /
    • v.22 no.2
    • /
    • pp.145-152
    • /
    • 2024
  • Closed quotient (CQ) represents the time ratio for which the vocal folds remain in contact during voice production. Because analyzing CQ values serves as an important reference point in vocal training for professional singers, these values have been measured mechanically or electrically by either inverse filtering of airflows captured by a circumferentially vented mask or post-processing of electroglottography waveforms. In this study, we introduced a novel algorithm to predict the CQ values only from audio signals. This has eliminated the need for mechanical or electrical measurement techniques. Our algorithm is based on a gated recurrent unit (GRU)-type neural network. To enhance the efficiency, we pre-processed an audio signal using the pitch feature extraction algorithm. Then, GRU-type neural networks were employed to extract the features. This was followed by a dense layer for the final prediction. The Results section reports the mean square error between the predicted and real CQ. It shows the capability of the proposed algorithm to predict CQ values.

Korean Sentence Generation Using Phoneme-Level LSTM Language Model (한국어 음소 단위 LSTM 언어모델을 이용한 문장 생성)

  • Ahn, SungMahn;Chung, Yeojin;Lee, Jaejoon;Yang, Jiheon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.71-88
    • /
    • 2017
  • Language models were originally developed for speech recognition and language processing. Using a set of example sentences, a language model predicts the next word or character based on sequential input data. N-gram models have been widely used but this model cannot model the correlation between the input units efficiently since it is a probabilistic model which are based on the frequency of each unit in the training set. Recently, as the deep learning algorithm has been developed, a recurrent neural network (RNN) model and a long short-term memory (LSTM) model have been widely used for the neural language model (Ahn, 2016; Kim et al., 2016; Lee et al., 2016). These models can reflect dependency between the objects that are entered sequentially into the model (Gers and Schmidhuber, 2001; Mikolov et al., 2010; Sundermeyer et al., 2012). In order to learning the neural language model, texts need to be decomposed into words or morphemes. Since, however, a training set of sentences includes a huge number of words or morphemes in general, the size of dictionary is very large and so it increases model complexity. In addition, word-level or morpheme-level models are able to generate vocabularies only which are contained in the training set. Furthermore, with highly morphological languages such as Turkish, Hungarian, Russian, Finnish or Korean, morpheme analyzers have more chance to cause errors in decomposition process (Lankinen et al., 2016). Therefore, this paper proposes a phoneme-level language model for Korean language based on LSTM models. A phoneme such as a vowel or a consonant is the smallest unit that comprises Korean texts. We construct the language model using three or four LSTM layers. Each model was trained using Stochastic Gradient Algorithm and more advanced optimization algorithms such as Adagrad, RMSprop, Adadelta, Adam, Adamax, and Nadam. Simulation study was done with Old Testament texts using a deep learning package Keras based the Theano. After pre-processing the texts, the dataset included 74 of unique characters including vowels, consonants, and punctuation marks. Then we constructed an input vector with 20 consecutive characters and an output with a following 21st character. Finally, total 1,023,411 sets of input-output vectors were included in the dataset and we divided them into training, validation, testsets with proportion 70:15:15. All the simulation were conducted on a system equipped with an Intel Xeon CPU (16 cores) and a NVIDIA GeForce GTX 1080 GPU. We compared the loss function evaluated for the validation set, the perplexity evaluated for the test set, and the time to be taken for training each model. As a result, all the optimization algorithms but the stochastic gradient algorithm showed similar validation loss and perplexity, which are clearly superior to those of the stochastic gradient algorithm. The stochastic gradient algorithm took the longest time to be trained for both 3- and 4-LSTM models. On average, the 4-LSTM layer model took 69% longer training time than the 3-LSTM layer model. However, the validation loss and perplexity were not improved significantly or became even worse for specific conditions. On the other hand, when comparing the automatically generated sentences, the 4-LSTM layer model tended to generate the sentences which are closer to the natural language than the 3-LSTM model. Although there were slight differences in the completeness of the generated sentences between the models, the sentence generation performance was quite satisfactory in any simulation conditions: they generated only legitimate Korean letters and the use of postposition and the conjugation of verbs were almost perfect in the sense of grammar. The results of this study are expected to be widely used for the processing of Korean language in the field of language processing and speech recognition, which are the basis of artificial intelligence systems.

Development of Prediction Model for Nitrogen Oxides Emission Using Artificial Intelligence (인공지능 기반 질소산화물 배출량 예측을 위한 연구모형 개발)

  • Jo, Ha-Nui;Park, Jisu;Yun, Yongju
    • Korean Chemical Engineering Research
    • /
    • v.58 no.4
    • /
    • pp.588-595
    • /
    • 2020
  • Prediction and control of nitrogen oxides (NOx) emission is of great interest in industry due to stricter environmental regulations. Herein, we propose an artificial intelligence (AI)-based framework for prediction of NOx emission. The framework includes pre-processing of data for training of neural networks and evaluation of the AI-based models. In this work, Long-Short-Term Memory (LSTM), one of the recurrent neural networks, was adopted to reflect the time series characteristics of NOx emissions. A decision tree was used to determine a time window of LSTM prior to training of the network. The neural network was trained with operational data from a heating furnace. The optimal model was obtained by optimizing hyper-parameters. The LSTM model provided a reliable prediction of NOx emission for both training and test data, showing an accuracy of 93% or more. The application of the proposed AI-based framework will provide new opportunities for predicting the emission of various air pollutants with time series characteristics.

Comparison of Rotational Strength in Shoulders with Anterior Instability and Normal Shoulders Using Isokinetic Testing (등속성 검사를 통한 견관절 전방 불안정 환자와 정상인의 회전력 비교)

  • Lee, Dong-Ki;Kim, Tae-Kwon;Lee, Jin-Hyuck;Lee, Dae-Hee;Jung, Woong-Kyo
    • Clinics in Shoulder and Elbow
    • /
    • v.15 no.2
    • /
    • pp.79-85
    • /
    • 2012
  • Objective: It has been expected that patient with posttraumatic recurrent anterior shoulder dislocation might have limited daily life activity because of pain and apprehension of dislocation. But there have been only a small number of investigations regarding the rotator strength in this patient. The aim of this study is to find the characteristics about rotator strength of patient with posttraumatic recurrent anterior shoulder dislocation using an isokinetic testing. Method: We enrolled thirteen patients with posttraumatic recurrent anterior shoulder dislocation and fifteen sex, age-matched healthy nonathletic subjects in this controlled study. All participants were male and there were no significant differences between the two groups in age, height, weight, BMI. Isokinetic internal rotator and external rotator strength was evaluated with a Biodex Isokinetic Testing machine (Biodex Medical Systems, Shirley, NY, USA), tests were performed at 60 deg/sec and 180 deg/sec for both sides. Peak torque normalized to body weight, external rotator to internal rotator ratio, total work and fatigue were calculated for each angular velocity. The association between internal rotator and external rotator strength and shoulder instability was analyzed by comparisons with a control group. Results: Any notable differences could not be found between the two groups given all data from no symptomatic left shoulder. There were no significant differences between the two groups statistically in internal rotation strength of right shoulder. However, there has been a tendency that at all angular velocities, external rotator peak torque to body weight, total work and external rotator to internal rotator ratio were significantly lower in the anterior instability group than the control group at all angular velocities. There was no substantial difference between those groups with respect to the fatigue of external rotator and internal rotator in our study. Conclusion: The prominent characteristics of posttraumatic recurrent anterior shoulder dislocation are external rotator weakness and loss of balance with external rotator and internal rotator. Therefore selective training using this information rotator might be helpful in conservative treatment and rehabilitation.

Estimation Model for Freight of Container Ships using Deep Learning Method (딥러닝 기법을 활용한 컨테이너선 운임 예측 모델)

  • Kim, Donggyun;Choi, Jung-Suk
    • Journal of the Korean Society of Marine Environment & Safety
    • /
    • v.27 no.5
    • /
    • pp.574-583
    • /
    • 2021
  • Predicting shipping markets is an important issue. Such predictions form the basis for decisions on investment methods, fleet formation methods, freight rates, etc., which greatly affect the profits and survival of a company. To this end, in this study, we propose a shipping freight rate prediction model for container ships using gated recurrent units (GRUs) and long short-term memory structure. The target of our freight rate prediction is the China Container Freight Index (CCFI), and CCFI data from March 2003 to May 2020 were used for training. The CCFI after June 2020 was first predicted according to each model and then compared and analyzed with the actual CCFI. For the experimental model, a total of six models were designed according to the hyperparameter settings. Additionally, the ARIMA model was included in the experiment for performance comparison with the traditional analysis method. The optimal model was selected based on two evaluation methods. The first evaluation method selects the model with the smallest average value of the root mean square error (RMSE) obtained by repeating each model 10 times. The second method selects the model with the lowest RMSE in all experiments. The experimental results revealed not only the improved accuracy of the deep learning model compared to the traditional time series prediction model, ARIMA, but also the contribution in enhancing the risk management ability of freight fluctuations through deep learning models. On the contrary, in the event of sudden changes in freight owing to the effects of external factors such as the Covid-19 pandemic, the accuracy of the forecasting model reduced. The GRU1 model recorded the lowest RMSE (69.55, 49.35) in both evaluation methods, and it was selected as the optimal model.