• Title/Summary/Keyword: hidden-nodes

Search Result 201, Processing Time 0.173 seconds

Function Approximation for Refrigerant Using the Neural Networks (신경회로망을 사용한 냉매의 함수근사)

  • Park, Jin-Hyun;Lee, Tae-Hwan
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • v.9 no.2
    • /
    • pp.677-680
    • /
    • 2005
  • In numerical analysis on the thermal performance of the heat exchanger with phase change fluids, the numerical values of thermodynamic properties are needed. But the steam table should be modeled properly as the direct use of thermodynamic properties of the steam table is impossible. In this study the function approximation characteristics of neural networks was used in modeling the saturated vapor region of refrigerant R12. The neural network consists of one input layer with one node, two hidden layers with 10 and 20 nodes each and one output layer with 7 nodes. Input can be both saturation temperature and saturation pressure and two cases were examined. The proposed model gives percentage error of ${\pm}$0.005% for enthalpy and entropy, ${\pm}$0.02% for specific volume and ${\pm}$0.02% for saturation pressure and saturation temperature except several points. From this results neural network could be a powerful method in function approximation of saturated vapor region of R12.

  • PDF

Modelling the wide temperature range of steam table using the neural networks (신경회로망을 사용한 넓은 온도 범위의 증기표 모델링)

  • Lee, Tae-Hwan
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.10 no.11
    • /
    • pp.2008-2013
    • /
    • 2006
  • In numerical analysis on evaluating the thermal performance of the thermal equipment, numerical values of thermodynamic properties such as temperature, pressure, specific volume, enthalpy and entropy are required. But the steam table itself cannot be used without modelling. In this study applicability of neural networks in modelling the wide temperature range of wet saturated vapor region was examined. the multi-layer neural network consists of a input layer with 1 node, two hidden layers with 10 and 20 nodes respectively and a output layer with 6 nodes. Quadratic and cubic spline interpoations methods were also applied for comparison. Neural network model revealed similar percentage error to spline interpolation. From these results, it is confirmed that the neural networks could be powerful method in modelling the wide range of the steam table.

Document classification using a deep neural network in text mining (텍스트 마이닝에서 심층 신경망을 이용한 문서 분류)

  • Lee, Bo-Hui;Lee, Su-Jin;Choi, Yong-Seok
    • The Korean Journal of Applied Statistics
    • /
    • v.33 no.5
    • /
    • pp.615-625
    • /
    • 2020
  • The document-term frequency matrix is a term extracted from documents in which the group information exists in text mining. In this study, we generated the document-term frequency matrix for document classification according to research field. We applied the traditional term weighting function term frequency-inverse document frequency (TF-IDF) to the generated document-term frequency matrix. In addition, we applied term frequency-inverse gravity moment (TF-IGM). We also generated a document-keyword weighted matrix by extracting keywords to improve the document classification accuracy. Based on the keywords matrix extracted, we classify documents using a deep neural network. In order to find the optimal model in the deep neural network, the accuracy of document classification was verified by changing the number of hidden layers and hidden nodes. Consequently, the model with eight hidden layers showed the highest accuracy and all TF-IGM document classification accuracy (according to parameter changes) were higher than TF-IDF. In addition, the deep neural network was confirmed to have better accuracy than the support vector machine. Therefore, we propose a method to apply TF-IGM and a deep neural network in the document classification.

A Deep Learning Performance Comparison of R and Tensorflow (R과 텐서플로우 딥러닝 성능 비교)

  • Sung-Bong Jang
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.4
    • /
    • pp.487-494
    • /
    • 2023
  • In this study, performance comparison was performed on R and TensorFlow, which are free deep learning tools. In the experiment, six types of deep neural networks were built using each tool, and the neural networks were trained using the 10-year Korean temperature dataset. The number of nodes in the input layer of the constructed neural network was set to 10, the number of output layers was set to 5, and the hidden layer was set to 5, 10, and 20 to conduct experiments. The dataset includes 3600 temperature data collected from Gangnam-gu, Seoul from March 1, 2013 to March 29, 2023. For performance comparison, the future temperature was predicted for 5 days using the trained neural network, and the root mean square error (RMSE) value was measured using the predicted value and the actual value. Experiment results shows that when there was one hidden layer, the learning error of R was 0.04731176, and TensorFlow was measured at 0.06677193, and when there were two hidden layers, R was measured at 0.04782134 and TensorFlow was measured at 0.05799060. Overall, R was measured to have better performance. We tried to solve the difficulties in tool selection by providing quantitative performance information on the two tools to users who are new to machine learning.

Application of Support Vector Regression for Improving the Performance of the Emotion Prediction Model (감정예측모형의 성과개선을 위한 Support Vector Regression 응용)

  • Kim, Seongjin;Ryoo, Eunchung;Jung, Min Kyu;Kim, Jae Kyeong;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.3
    • /
    • pp.185-202
    • /
    • 2012
  • .Since the value of information has been realized in the information society, the usage and collection of information has become important. A facial expression that contains thousands of information as an artistic painting can be described in thousands of words. Followed by the idea, there has recently been a number of attempts to provide customers and companies with an intelligent service, which enables the perception of human emotions through one's facial expressions. For example, MIT Media Lab, the leading organization in this research area, has developed the human emotion prediction model, and has applied their studies to the commercial business. In the academic area, a number of the conventional methods such as Multiple Regression Analysis (MRA) or Artificial Neural Networks (ANN) have been applied to predict human emotion in prior studies. However, MRA is generally criticized because of its low prediction accuracy. This is inevitable since MRA can only explain the linear relationship between the dependent variables and the independent variable. To mitigate the limitations of MRA, some studies like Jung and Kim (2012) have used ANN as the alternative, and they reported that ANN generated more accurate prediction than the statistical methods like MRA. However, it has also been criticized due to over fitting and the difficulty of the network design (e.g. setting the number of the layers and the number of the nodes in the hidden layers). Under this background, we propose a novel model using Support Vector Regression (SVR) in order to increase the prediction accuracy. SVR is an extensive version of Support Vector Machine (SVM) designated to solve the regression problems. The model produced by SVR only depends on a subset of the training data, because the cost function for building the model ignores any training data that is close (within a threshold ${\varepsilon}$) to the model prediction. Using SVR, we tried to build a model that can measure the level of arousal and valence from the facial features. To validate the usefulness of the proposed model, we collected the data of facial reactions when providing appropriate visual stimulating contents, and extracted the features from the data. Next, the steps of the preprocessing were taken to choose statistically significant variables. In total, 297 cases were used for the experiment. As the comparative models, we also applied MRA and ANN to the same data set. For SVR, we adopted '${\varepsilon}$-insensitive loss function', and 'grid search' technique to find the optimal values of the parameters like C, d, ${\sigma}^2$, and ${\varepsilon}$. In the case of ANN, we adopted a standard three-layer backpropagation network, which has a single hidden layer. The learning rate and momentum rate of ANN were set to 10%, and we used sigmoid function as the transfer function of hidden and output nodes. We performed the experiments repeatedly by varying the number of nodes in the hidden layer to n/2, n, 3n/2, and 2n, where n is the number of the input variables. The stopping condition for ANN was set to 50,000 learning events. And, we used MAE (Mean Absolute Error) as the measure for performance comparison. From the experiment, we found that SVR achieved the highest prediction accuracy for the hold-out data set compared to MRA and ANN. Regardless of the target variables (the level of arousal, or the level of positive / negative valence), SVR showed the best performance for the hold-out data set. ANN also outperformed MRA, however, it showed the considerably lower prediction accuracy than SVR for both target variables. The findings of our research are expected to be useful to the researchers or practitioners who are willing to build the models for recognizing human emotions.

Efficient MAC Protocol for Achieving Fairness in Wireless Ad-hoc Networks under the DCF (Wireless Ad-hoc 망의 DCF환경에서 Fairness 구현을 위한 효과적인 MAC 프로토콜)

  • 임희선;박승권
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.8A
    • /
    • pp.928-936
    • /
    • 2004
  • This paper proposes simple and effective Contention Window (CW) adjusting algorithm to solve the fairness problem of the IEEE 802.11 under the Distributed Coordination Function (DCF). This adjusting algorithm can support variable packet length as well as both RTS/CTS access method and hidden nodes. Simulation results show that fairness problem can be very severe in the original MAC protocol of the wireless network. In case of implement our algorithm, fair sharing was enlarged prominently and the fairness sharing was improved about maximum 30%.

Classification of Premature Ventricular Contraction using Error Back-Propagation

  • Jeon, Eunkwang;Jung, Bong-Keun;Nam, Yunyoung;Lee, HwaMin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.2
    • /
    • pp.988-1001
    • /
    • 2018
  • Arrhythmia has recently emerged as one of the major causes of death in Koreans. Premature Ventricular Contraction (PVC) is the most common arrhythmia that can be found in clinical practice, and it may be a precursor to dangerous arrhythmias, such as paroxysmal insomnia, ventricular fibrillation, and coronary artery disease. Therefore, we need for a method that can detect an abnormal heart beat and diagnose arrhythmia early. We extracted the features corresponding to the QRS pattern from the subject's ECG signal and classify the premature ventricular contraction waveform using the features. We modified the weighting and bias values based on the error back-propagation algorithm through learning data. We classify the normal signal and the premature ventricular contraction signal through the modified weights and deflection values. MIT-BIH arrhythmia data sets were used for performance tests. We used RR interval, QS interval, QR amplitude and RS amplitude features. And the hidden layer with two nodes is composed of two layers to form a total three layers (input layer 0, output layer 3).

Robust Multi-Layer Hierarchical Model for Digit Character Recognition

  • Yang, Jie;Sun, Yadong;Zhang, Liangjun;Zhang, Qingnian
    • Journal of Electrical Engineering and Technology
    • /
    • v.10 no.2
    • /
    • pp.699-707
    • /
    • 2015
  • Although digit character recognition has got a significant improvement in recent years, it is still challenging to achieve satisfied result if the data contains an amount of distracting factors. This paper proposes a novel digit character recognition approach using a multi-layer hierarchical model, Hybrid Restricted Boltzmann Machines (HRBMs), which allows the learning architecture to be robust to background distracting factors. The insight behind the proposed model is that useful high-level features appear more frequently than distracting factors during learning, thus the high-level features can be decompose into hybrid hierarchical structures by using only small label information. In order to extract robust and compact features, a stochastic 0-1 layer is employed, which enables the model's hidden nodes to independently capture the useful character features during training. Experiments on the variations of Mixed National Institute of Standards and Technology (MNIST) dataset show that improvements of the multi-layer hierarchical model can be achieved by the proposed method. Finally, the paper shows the proposed technique which is used in a real-world application, where it is able to identify digit characters under various complex background images.

Tuning the Architecture of Neural Networks for Multi-Class Classification (다집단 분류 인공신경망 모형의 아키텍쳐 튜닝)

  • Jeong, Chulwoo;Min, Jae H.
    • Journal of the Korean Operations Research and Management Science Society
    • /
    • v.38 no.1
    • /
    • pp.139-152
    • /
    • 2013
  • The purpose of this study is to claim the validity of tuning the architecture of neural network models for multi-class classification. A neural network model for multi-class classification is basically constructed by building a series of neural network models for binary classification. Building a neural network model, we are required to set the values of parameters such as number of hidden nodes and weight decay parameter in advance, which draws special attention as the performance of the model can be quite different by the values of the parameters. For better performance of the model, it is absolutely necessary to have a prior process of tuning the parameters every time the neural network model is built. Nonetheless, previous studies have not mentioned the necessity of the tuning process or proved its validity. In this study, we claim that we should tune the parameters every time we build the neural network model for multi-class classification. Through empirical analysis using wine data, we show that the performance of the model with the tuned parameters is superior to those of untuned models.

Optimal Design of Radial Basis Function Network Using Time-Frequency Localization (시간-주파수 지역화를 이용한 방사 기준 함수 구조의 최적 설계)

  • Kim, Yong-Taek;Kim, Seong-Joo;Seo, Jae-Yong;Jeon, Hong-Tae
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.38 no.5
    • /
    • pp.1-6
    • /
    • 2001
  • In this paper, we propose the initial optimized structure of the Radial Basis Function Network(RBFN) which is more simple in the part of the structure and converges more faster than Neural Network. For this, we use the analysis method using time frequency localization and we can decide the initial structure of the RBFN suitable for the given problem. When we compose the hidden nodes with the radial basis functions whose localization are similar with the target function in the plane of the time and frequency, we can make a good decision of the initial structure having an ability of approximation.

  • PDF