• Title/Summary/Keyword: Input layers

Search Result 431, Processing Time 0.027 seconds

Study on the Neural Network for Handwritten Hangul Syllabic Character Recognition (수정된 Neocognitron을 사용한 필기체 한글인식)

  • 김은진;백종현
    • Korean Journal of Cognitive Science
    • /
    • v.3 no.1
    • /
    • pp.61-78
    • /
    • 1991
  • This paper descibes the study of application of a modified Neocognitron model with backward path for the recognition of Hangul(Korean) syllabic characters. In this original report, Fukushima demonstrated that Neocognitron can recognize hand written numerical characters of $19{\times}19$ size. This version accepts $61{\times}61$ images of handwritten Hangul syllabic characters or a part thereof with a mouse or with a scanner. It consists of an input layer and 3 pairs of Uc layers. The last Uc layer of this version, recognition layer, consists of 24 planes of $5{\times}5$ cells which tell us the identity of a grapheme receiving attention at one time and its relative position in the input layer respectively. It has been trained 10 simple vowel graphemes and 14 simple consonant graphemes and their spatial features. Some patterns which are not easily trained have been trained more extrensively. The trained nerwork which can classify indivisual graphemes with possible deformation, noise, size variance, transformation or retation wre then used to recongnize Korean syllabic characters using its selective attention mechanism for image segmentation task within a syllabic characters. On initial sample tests on input characters our model could recognize correctly up to 79%of the various test patterns of handwritten Korean syllabic charactes. The results of this study indeed show Neocognitron as a powerful model to reconginze deformed handwritten charavters with big size characters set via segmenting its input images as recognizable parts. The same approach may be applied to the recogition of chinese characters, which are much complex both in its structures and its graphemes. But processing time appears to be the bottleneck before it can be implemented. Special hardware such as neural chip appear to be an essestial prerquisite for the practical use of the model. Further work is required before enabling the model to recognize Korean syllabic characters consisting of complex vowels and complex consonants. Correct recognition of the neighboring area between two simple graphemes would become more critical for this task.

Feasibility of Deep Learning Algorithms for Binary Classification Problems (이진 분류문제에서의 딥러닝 알고리즘의 활용 가능성 평가)

  • Kim, Kitae;Lee, Bomi;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.95-108
    • /
    • 2017
  • Recently, AlphaGo which is Bakuk (Go) artificial intelligence program by Google DeepMind, had a huge victory against Lee Sedol. Many people thought that machines would not be able to win a man in Go games because the number of paths to make a one move is more than the number of atoms in the universe unlike chess, but the result was the opposite to what people predicted. After the match, artificial intelligence technology was focused as a core technology of the fourth industrial revolution and attracted attentions from various application domains. Especially, deep learning technique have been attracted as a core artificial intelligence technology used in the AlphaGo algorithm. The deep learning technique is already being applied to many problems. Especially, it shows good performance in image recognition field. In addition, it shows good performance in high dimensional data area such as voice, image and natural language, which was difficult to get good performance using existing machine learning techniques. However, in contrast, it is difficult to find deep leaning researches on traditional business data and structured data analysis. In this study, we tried to find out whether the deep learning techniques have been studied so far can be used not only for the recognition of high dimensional data but also for the binary classification problem of traditional business data analysis such as customer churn analysis, marketing response prediction, and default prediction. And we compare the performance of the deep learning techniques with that of traditional artificial neural network models. The experimental data in the paper is the telemarketing response data of a bank in Portugal. It has input variables such as age, occupation, loan status, and the number of previous telemarketing and has a binary target variable that records whether the customer intends to open an account or not. In this study, to evaluate the possibility of utilization of deep learning algorithms and techniques in binary classification problem, we compared the performance of various models using CNN, LSTM algorithm and dropout, which are widely used algorithms and techniques in deep learning, with that of MLP models which is a traditional artificial neural network model. However, since all the network design alternatives can not be tested due to the nature of the artificial neural network, the experiment was conducted based on restricted settings on the number of hidden layers, the number of neurons in the hidden layer, the number of output data (filters), and the application conditions of the dropout technique. The F1 Score was used to evaluate the performance of models to show how well the models work to classify the interesting class instead of the overall accuracy. The detail methods for applying each deep learning technique in the experiment is as follows. The CNN algorithm is a method that reads adjacent values from a specific value and recognizes the features, but it does not matter how close the distance of each business data field is because each field is usually independent. In this experiment, we set the filter size of the CNN algorithm as the number of fields to learn the whole characteristics of the data at once, and added a hidden layer to make decision based on the additional features. For the model having two LSTM layers, the input direction of the second layer is put in reversed position with first layer in order to reduce the influence from the position of each field. In the case of the dropout technique, we set the neurons to disappear with a probability of 0.5 for each hidden layer. The experimental results show that the predicted model with the highest F1 score was the CNN model using the dropout technique, and the next best model was the MLP model with two hidden layers using the dropout technique. In this study, we were able to get some findings as the experiment had proceeded. First, models using dropout techniques have a slightly more conservative prediction than those without dropout techniques, and it generally shows better performance in classification. Second, CNN models show better classification performance than MLP models. This is interesting because it has shown good performance in binary classification problems which it rarely have been applied to, as well as in the fields where it's effectiveness has been proven. Third, the LSTM algorithm seems to be unsuitable for binary classification problems because the training time is too long compared to the performance improvement. From these results, we can confirm that some of the deep learning algorithms can be applied to solve business binary classification problems.

Rare Earth Elements (REE)-bearing Coal Deposits: Potential of Coal Beds as an Unconventional REE Source (함희토류 탄층: 비전통적 희토류 광체로서의 가능성에 대한 고찰)

  • Choi, Woohyun;Park, Changyun
    • Economic and Environmental Geology
    • /
    • v.55 no.3
    • /
    • pp.241-259
    • /
    • 2022
  • In general, the REE were produced by mining conventional deposits, such as the carbonatite or the clay-hosted REE deposits. However, because of the recent demand increase for REE in modern industries, unconventional REE deposits emerged as a necessary research topic. Among the unconventional REE recovery methods, the REE-bearing coal deposits are recently receiving attentions. R-types generally have detrital originations from the bauxite deposits, and show LREE enriched REE patterns. Tuffaceous-types are formed by syngenetic volcanic activities and following input of volcanic ash into the basin. This type shows specific occurrence of the detrital volcanic ash-driven minerals and the authigenic phosphorous minerals focused at narrow horizon between coal seams and tonstein layers. REE patterns of tuffaceous-types show flat shape in general. Hydrothermal-types can be formed by epigenetic inflow of REE originated from granitic intrusions. Occurrence of the authigenic halogen-bearing phosphorous minerals and the water-bearing minerals are the specific characteristics of this type. They generally show HREE enriched REE patterns. Each type of REE-bearing coal deposits may occur by independent genesis, but most of REE-bearing coal deposits with high REE concentrations have multiple genesis. For the case of the US, the rare earth oxides (REO) with high purity has been produced from REE-bearing coals and their byproducts in pilot plants from 2018. Their goal is to supply about 7% of national REE demand. For the coal deposits in Korea, lignite layers found in Gyungju-Yeongil coal fields shows coexistence of tuff layers and coal seams. They are also based in Tertiary basins, and low affection from compaction and coalification might resulted into high-REE tuffaceous-type coal deposits. Thus, detailed geologic researches and explorations for domestic coal deposits are required.

Study of Geological Log Database for Public Wells, Jeju Island (제주도 공공 관정 지질주상도 DB 구축 소개)

  • Pak, Song-Hyon;Koh, Giwon;Park, Junbeom;Moon, Dukchul;Yoon, Woo Seok
    • Economic and Environmental Geology
    • /
    • v.48 no.6
    • /
    • pp.509-523
    • /
    • 2015
  • This study introduces newly implemented geological well logs database for Jeju public water wells, built for a research project focusing on integrated hydrogeology database of Jeju Island. A detailed analysis of the existing 1,200 Jeju Island geological logs for the public wells developed since 1970 revealed six major indications to be improved for their use in Jeju geological logs DB construction: (1) lack of uniformity in rock name classification, (2) poor definitions of pyroclastic deposits and sand and gravel layers, (3) lack of well borehole aquifer information, (4) lack of information on well screen installation in many water wells, (5) differences by person in geological logging descriptions. A new Jeju geological logs DB enabling standardized input and output formats has been implemented to overcome the above indications by reestablishing the names of Jeju volcanic and sedimentary rocks and utilizing a commercial, database-based input structured, geological log program. The newly designed database structure in geological log program enables users to store a large number of geology, well drilling, and test data at the standardized DB input structure. Also, well borehole groundwater and aquifer test data can be easily added without modifying the existing database structure. Thus, the newly implemented geological logs DB could be a standardized DB for a large number of Jeju existing public wells and new wells to be developed in the future at Jeju Island. Also, the new geological logs DB will be a basis for ongoing project 'Developing GIS-based integrated interpretation system for Jeju Island hydrogeology'.

Korean Sentence Generation Using Phoneme-Level LSTM Language Model (한국어 음소 단위 LSTM 언어모델을 이용한 문장 생성)

  • Ahn, SungMahn;Chung, Yeojin;Lee, Jaejoon;Yang, Jiheon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.71-88
    • /
    • 2017
  • Language models were originally developed for speech recognition and language processing. Using a set of example sentences, a language model predicts the next word or character based on sequential input data. N-gram models have been widely used but this model cannot model the correlation between the input units efficiently since it is a probabilistic model which are based on the frequency of each unit in the training set. Recently, as the deep learning algorithm has been developed, a recurrent neural network (RNN) model and a long short-term memory (LSTM) model have been widely used for the neural language model (Ahn, 2016; Kim et al., 2016; Lee et al., 2016). These models can reflect dependency between the objects that are entered sequentially into the model (Gers and Schmidhuber, 2001; Mikolov et al., 2010; Sundermeyer et al., 2012). In order to learning the neural language model, texts need to be decomposed into words or morphemes. Since, however, a training set of sentences includes a huge number of words or morphemes in general, the size of dictionary is very large and so it increases model complexity. In addition, word-level or morpheme-level models are able to generate vocabularies only which are contained in the training set. Furthermore, with highly morphological languages such as Turkish, Hungarian, Russian, Finnish or Korean, morpheme analyzers have more chance to cause errors in decomposition process (Lankinen et al., 2016). Therefore, this paper proposes a phoneme-level language model for Korean language based on LSTM models. A phoneme such as a vowel or a consonant is the smallest unit that comprises Korean texts. We construct the language model using three or four LSTM layers. Each model was trained using Stochastic Gradient Algorithm and more advanced optimization algorithms such as Adagrad, RMSprop, Adadelta, Adam, Adamax, and Nadam. Simulation study was done with Old Testament texts using a deep learning package Keras based the Theano. After pre-processing the texts, the dataset included 74 of unique characters including vowels, consonants, and punctuation marks. Then we constructed an input vector with 20 consecutive characters and an output with a following 21st character. Finally, total 1,023,411 sets of input-output vectors were included in the dataset and we divided them into training, validation, testsets with proportion 70:15:15. All the simulation were conducted on a system equipped with an Intel Xeon CPU (16 cores) and a NVIDIA GeForce GTX 1080 GPU. We compared the loss function evaluated for the validation set, the perplexity evaluated for the test set, and the time to be taken for training each model. As a result, all the optimization algorithms but the stochastic gradient algorithm showed similar validation loss and perplexity, which are clearly superior to those of the stochastic gradient algorithm. The stochastic gradient algorithm took the longest time to be trained for both 3- and 4-LSTM models. On average, the 4-LSTM layer model took 69% longer training time than the 3-LSTM layer model. However, the validation loss and perplexity were not improved significantly or became even worse for specific conditions. On the other hand, when comparing the automatically generated sentences, the 4-LSTM layer model tended to generate the sentences which are closer to the natural language than the 3-LSTM model. Although there were slight differences in the completeness of the generated sentences between the models, the sentence generation performance was quite satisfactory in any simulation conditions: they generated only legitimate Korean letters and the use of postposition and the conjugation of verbs were almost perfect in the sense of grammar. The results of this study are expected to be widely used for the processing of Korean language in the field of language processing and speech recognition, which are the basis of artificial intelligence systems.

A Deep Learning Based Approach to Recognizing Accompanying Status of Smartphone Users Using Multimodal Data (스마트폰 다종 데이터를 활용한 딥러닝 기반의 사용자 동행 상태 인식)

  • Kim, Kilho;Choi, Sangwoo;Chae, Moon-jung;Park, Heewoong;Lee, Jaehong;Park, Jonghun
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.163-177
    • /
    • 2019
  • As smartphones are getting widely used, human activity recognition (HAR) tasks for recognizing personal activities of smartphone users with multimodal data have been actively studied recently. The research area is expanding from the recognition of the simple body movement of an individual user to the recognition of low-level behavior and high-level behavior. However, HAR tasks for recognizing interaction behavior with other people, such as whether the user is accompanying or communicating with someone else, have gotten less attention so far. And previous research for recognizing interaction behavior has usually depended on audio, Bluetooth, and Wi-Fi sensors, which are vulnerable to privacy issues and require much time to collect enough data. Whereas physical sensors including accelerometer, magnetic field and gyroscope sensors are less vulnerable to privacy issues and can collect a large amount of data within a short time. In this paper, a method for detecting accompanying status based on deep learning model by only using multimodal physical sensor data, such as an accelerometer, magnetic field and gyroscope, was proposed. The accompanying status was defined as a redefinition of a part of the user interaction behavior, including whether the user is accompanying with an acquaintance at a close distance and the user is actively communicating with the acquaintance. A framework based on convolutional neural networks (CNN) and long short-term memory (LSTM) recurrent networks for classifying accompanying and conversation was proposed. First, a data preprocessing method which consists of time synchronization of multimodal data from different physical sensors, data normalization and sequence data generation was introduced. We applied the nearest interpolation to synchronize the time of collected data from different sensors. Normalization was performed for each x, y, z axis value of the sensor data, and the sequence data was generated according to the sliding window method. Then, the sequence data became the input for CNN, where feature maps representing local dependencies of the original sequence are extracted. The CNN consisted of 3 convolutional layers and did not have a pooling layer to maintain the temporal information of the sequence data. Next, LSTM recurrent networks received the feature maps, learned long-term dependencies from them and extracted features. The LSTM recurrent networks consisted of two layers, each with 128 cells. Finally, the extracted features were used for classification by softmax classifier. The loss function of the model was cross entropy function and the weights of the model were randomly initialized on a normal distribution with an average of 0 and a standard deviation of 0.1. The model was trained using adaptive moment estimation (ADAM) optimization algorithm and the mini batch size was set to 128. We applied dropout to input values of the LSTM recurrent networks to prevent overfitting. The initial learning rate was set to 0.001, and it decreased exponentially by 0.99 at the end of each epoch training. An Android smartphone application was developed and released to collect data. We collected smartphone data for a total of 18 subjects. Using the data, the model classified accompanying and conversation by 98.74% and 98.83% accuracy each. Both the F1 score and accuracy of the model were higher than the F1 score and accuracy of the majority vote classifier, support vector machine, and deep recurrent neural network. In the future research, we will focus on more rigorous multimodal sensor data synchronization methods that minimize the time stamp differences. In addition, we will further study transfer learning method that enables transfer of trained models tailored to the training data to the evaluation data that follows a different distribution. It is expected that a model capable of exhibiting robust recognition performance against changes in data that is not considered in the model learning stage will be obtained.

Measurements of Ultrasound Attenuation Coefficient at Various Suspended Sediment Concentrations (부유물 농도 변화에 따른 초음파 신호의 감쇠계수 측정)

  • Lee, Changil;Choi, Jee Woong
    • The Journal of the Acoustical Society of Korea
    • /
    • v.33 no.1
    • /
    • pp.1-9
    • /
    • 2014
  • Coastal water including estuaries has distinctive environmental characteristics where sediments are transported and deposited by flowing river water, providing an environment in which fluid mud layers can be formed. Acoustic method is mostly used to detect or monitor the fluid mud layer. However, since sound propagating in this layer suffers severe attenuation, it is important to estimate the accurate attenuation coefficient for various concentrations of fluid mud layer for the successful use of the acoustic method. In this paper, measurement results of attenuation coefficient for 3.5, 5, and 7.5 MHz ultrasounds were presented. The measurements were made in a small-size water tank in which suspended sediment samples with various sediment concentrations were formed using kaolinite powder. The results were compared to the model predictions obtained by attenuation coefficient model in which the mean grain size (called as Mass-median-diameter, D50) was used as input parameter. There were reasonable agreements between measured attenuation coefficients and model outputs predicted using the particle range of D50 ${\pm}20%$. The comparison results imply that although the suspended sediments consist of various-sized particles, sound attenuation might be greatly influenced by amount of particle with a size which has a larger attenuation than that of any particle in the suspended sediments for the frequency used.

A Study on Development of a GIS based Post-processing System of the EFDC Model for Supporting Water Quality Management (수질관리 지원을 위한 GIS기반의 EFDC 모델 후처리 시스템 개발 연구)

  • Lee, Geon Hwi;Kim, Kye Hyun;Park, Yong Gil;Lee, Sung Joo
    • Spatial Information Research
    • /
    • v.22 no.4
    • /
    • pp.39-47
    • /
    • 2014
  • The Yeongsan river estuary has a serious water quality problem due to the water stagnation and it is imperative to predict the changes of water quality for mitigating water pollution. EFDC(Environmental Fluid Dynamics Code) model was mainly utilized to predict the changes of water quality for the estuary. The EFDC modeling normally accompanies the large volume of modeling output. For checking the spatial distribution of the modeling results, post-processing for converting of the output is prerequisite and mainly post-processing program is EFDC_Explorer. However, EFDC_Explorer only shows the spatial distribution of the time series and this doesn't support overlay function with other thematic maps. This means the impossible to the connection analysis with a various GIS data and high dimensional analysis. Therefore, this study aims to develop a post-processing system of a EFDC output to use them as GIS layers. For achieving this purpose, a editing module for main input files, and a module for converting binary format into an ASCII format, and a module for converting it into a layer format to use in a GIS based environment, and a module for visualizing the reconfigured model result efficiently were developed. Using the developed system, result file is possible to automatically convert the GIS based layer and it is possible to utilize for water quality management.

Automatic Parameter Acquisition of 12 leads ECG Using Continuous Data Processing Deep Neural Network (연속적 데이터 처리 심층신경망을 이용한 12 lead 심전도 파라미터의 자동 획득)

  • Kim, Ji Woon;Park, Sung Min;Choi, Seong Wook
    • Journal of Biomedical Engineering Research
    • /
    • v.41 no.2
    • /
    • pp.107-119
    • /
    • 2020
  • The deep neural networks (DNN) that can replicate the behavior of the human expert who recognizes the characteristics of ECG waveform have been developed and studied to analyze ECG. However, although the existing DNNs can not provide the explanations for their decisions, those trials have attempted to determine whether patients have certain diseases or not and those decisions could not be accepted because of the absence of relating theoretical basis. In addition, these DNNs required a lot of training data to obtain sufficient accuracy in spite of the difficulty in the acquisition of relating clinical data. In this study, a small-sized continuous data processing DNN (C-DNN) was suggested to determine the simple characteristics of ECG wave that were not required additional explanations about its decisions and the C-DNN can be easily trained with small training data. Although it can analyze small input data that was selected in narrow region on whole ECG, it can continuously scan all ECG data and find important points such as start and end points of P, QRS and T waves within a short time. The star and end points of ECG waves determined by the C-DNNs were compared with the results performed by human experts to estimate the accuracies of the C-DNNs. The C-DNN has 150 inputs, 51 outputs, two hidden layers and one output layer. To find the start and end points, two C-DNNs were trained through deep learning technology and applied to a parameter acquisition algorithms. 12 lead ECG data measured in four patients and obtained through PhysioNet was processed to make training data by human experts. The accuracy of the C-DNNs were evaluated with extra data that were not used at deep learning by comparing the results between C-DNNs and human experts. The averages of the time differences between the C-DNNs and experts were 0.1 msec and 13.5 msec respectively and those standard deviations were 17.6 msec and 15.7 msec. The final step combining the results of C-DNN through the waveforms of 12 leads was successfully determined all 33 waves without error that the time differences of human experts decision were over 20 msec. The reliable decision of the ECG wave's start and end points benefits the acquisition of accurate ECG parameters such as the wave lengths, amplitudes and intervals of P, QRS and T waves.

Construction of Three Dimensional Soil Cadmium Pollution Map Using Geotechnical Information DB System (국토지반정보시스템을 이용한 3차원 토양오염지도 구축)

  • Hwang, Dae Young;Kang, In Joon;Jang, Yong Gu;Kim, Soo Kyum
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.24 no.4
    • /
    • pp.13-19
    • /
    • 2016
  • This study presented the build-up of three-dimensional soil pollution map for precise analysis. To do this, survey on the existing pollutant region on Dongnae-gu, Busan that is the study subject, showed that it tended to produce 0.72 clusters. So, this study suggested to investigate center of $1km{\times}1km $ grid and, as the results of comparing the pollution map that input pollution figure values based on the actually investigation point showed precise results. And, it divided the standard of pollution into 5 levels in surface and underground space and the map was built up using IDW interpolation against the amount of polluted substance. The pollution of ground surface, flow of polluted substance, coefficient of permeability and ground water level that are 504 geotechnical informations were selected as the influential parameters in pollution analysis of underground space, and it calculated that to 0~20 points by dividing the characteristics. It enables the build-up of pollution map of ground surface-underground with depth that considers the characteristics of soil layers and it is considered that it is possible to analyze the general infiltration. And, it was considered that it enables more accurate forecast about influential analysis per depth and pollution of underground water.