• Title/Summary/Keyword: Hyper Parameter Optimization

Search Result 34, Processing Time 0.031 seconds

A Study on the Prediction of Mass and Length of Injection-molded Product Using Artificial Neural Network (인공신경망을 활용한 사출성형품의 질량과 치수 예측에 관한 연구)

  • Yang, Dong-Cheol;Lee, Jun-Han;Kim, Jong-Sun
    • Design & Manufacturing
    • /
    • v.14 no.3
    • /
    • pp.1-7
    • /
    • 2020
  • This paper predicts the mass and the length of injection-molded products through the Artificial Neural Network (ANN) method. The ANN was implemented with 5 input parameters and 2 output parameters(mass, length). The input parameters, such as injection time, melt temperature, mold temperature, packing pressure and packing time were selected. 44 experiments that are based on the mixed sampling method were performed to generate training data for the ANN model. The generated training data were normalized to eliminate scale differences between factors to improve the prediction performance of the ANN model. A random search method was used to find the optimized hyper-parameter of the ANN model. After the ANN completed the training, the ANN model predicted the mass and the length of the injection-molded product. According to the result, average error of the ANN for mass was 0.3 %. In the case of length, the average deviation of ANN was 0.043 mm.

Enhanced CNN Model for Brain Tumor Classification

  • Kasukurthi, Aravinda;Paleti, Lakshmikanth;Brahmaiah, Madamanchi;Sree, Ch.Sudha
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.5
    • /
    • pp.143-148
    • /
    • 2022
  • Brain tumor classification is an important process that allows doctors to plan treatment for patients based on the stages of the tumor. To improve classification performance, various CNN-based architectures are used for brain tumor classification. Existing methods for brain tumor segmentation suffer from overfitting and poor efficiency when dealing with large datasets. The enhanced CNN architecture proposed in this study is based on U-Net for brain tumor segmentation, RefineNet for pattern analysis, and SegNet architecture for brain tumor classification. The brain tumor benchmark dataset was used to evaluate the enhanced CNN model's efficiency. Based on the local and context information of the MRI image, the U-Net provides good segmentation. SegNet selects the most important features for classification while also reducing the trainable parameters. In the classification of brain tumors, the enhanced CNN method outperforms the existing methods. The enhanced CNN model has an accuracy of 96.85 percent, while the existing CNN with transfer learning has an accuracy of 94.82 percent.

Development of AI oxygen temperature measurement technology using hyperspectral optical visualization technology (초분광 광학가시화 기술을 활용한 인공지능 산소온도 측정기술 개발)

  • Jeong Hun Lee;Bo Ra Kim;Seung Hun Lee;Joon Sik Kim;Min Yoon;Gyeong Rae Cho
    • Journal of the Korean Society of Visualization
    • /
    • v.21 no.1
    • /
    • pp.103-109
    • /
    • 2023
  • This research developed a measurement technique that can measure the oxygen temperature inside a high temperature furnace. Instead of measuring only changes in frequency components within a small range used in the existing variable laser absorption spectroscopy, laser spectroscopy technology was used to spread out wavelength of the light source passing through the gas Based on a total of 20,000 image data, research was conducted to predict the temperature of a high-temperature furnace using CNN with black and white images in the form of spectral bands by temperature of 25 to 800 degrees. The optimal model was found through Hyper parameter optimization, R2 score is 0.89, and the accuracy of the test data is 88.73%. Based on this research, it is expected that concentration measurement and air-fuel ratio control technology can be applied.

Pragmatic Assessment of Optimizers in Deep Learning

  • Ajeet K. Jain;PVRD Prasad Rao ;K. Venkatesh Sharma
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.10
    • /
    • pp.115-128
    • /
    • 2023
  • Deep learning has been incorporating various optimization techniques motivated by new pragmatic optimizing algorithm advancements and their usage has a central role in Machine learning. In recent past, new avatars of various optimizers are being put into practice and their suitability and applicability has been reported on various domains. The resurgence of novelty starts from Stochastic Gradient Descent to convex and non-convex and derivative-free approaches. In the contemporary of these horizons of optimizers, choosing a best-fit or appropriate optimizer is an important consideration in deep learning theme as these working-horse engines determines the final performance predicted by the model. Moreover with increasing number of deep layers tantamount higher complexity with hyper-parameter tuning and consequently need to delve for a befitting optimizer. We empirically examine most popular and widely used optimizers on various data sets and networks-like MNIST and GAN plus others. The pragmatic comparison focuses on their similarities, differences and possibilities of their suitability for a given application. Additionally, the recent optimizer variants are highlighted with their subtlety. The article emphasizes on their critical role and pinpoints buttress options while choosing among them.

Analysis of Accuracy and Loss Performance According to Hyperparameter in RNN Model (RNN모델에서 하이퍼파라미터 변화에 따른 정확도와 손실 성능 분석)

  • Kim, Joon-Yong;Park, Koo-Rack
    • Journal of Convergence for Information Technology
    • /
    • v.11 no.7
    • /
    • pp.31-38
    • /
    • 2021
  • In this paper, in order to obtain the optimization of the RNN model used for sentiment analysis, the correlation of each model was studied by observing the trend of loss and accuracy according to hyperparameter tuning. As a research method, after configuring the hidden layer with LSTM and the embedding layer that are most optimized to process sequential data, the loss and accuracy of each model were measured by tuning the unit, batch-size, and embedding size of the LSTM. As a result of the measurement, the loss was 41.9% and the accuracy was 11.4%, and the trend of the optimization model showed a consistently stable graph, confirming that the tuning of the hyperparameter had a profound effect on the model. In addition, it was confirmed that the decision of the embedding size among the three hyperparameters had the greatest influence on the model. In the future, this research will be continued, and research on an algorithm that allows the model to directly find the optimal hyperparameter will continue.

A study on the 3-step classification algorithm for the diagnosis and classification of refrigeration system failures and their types (냉동시스템 고장 진단 및 고장유형 분석을 위한 3단계 분류 알고리즘에 관한 연구)

  • Lee, Kangbae;Park, Sungho;Lee, Hui-Won;Lee, Seung-Jae;Lee, Seung-hyun
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.8
    • /
    • pp.31-37
    • /
    • 2021
  • As the size of buildings increases due to urbanization due to the development of industry, the need to purify the air and maintain a comfortable indoor environment is also increasing. With the development of monitoring technology for refrigeration systems, it has become possible to manage the amount of electricity consumed in buildings. In particular, refrigeration systems account for about 40% of power consumption in commercial buildings. Therefore, in order to develop the refrigeration system failure diagnosis algorithm in this study, the purpose of this study was to understand the structure of the refrigeration system, collect and analyze data generated during the operation of the refrigeration system, and quickly detect and classify failure situations with various types and severity . In particular, in order to improve the classification accuracy of failure types that are difficult to classify, a three-step diagnosis and classification algorithm was developed and proposed. A model based on SVM and LGBM was presented as a classification model suitable for each stage after a number of experiments and hyper-parameter optimization process. In this study, the characteristics affecting failure were preserved as much as possible, and all failure types, including refrigerant-related failures, which had been difficult in previous studies, were derived with excellent results.

Evaporative demand drought index forecasting in Busan-Ulsan-Gyeongnam region using machine learning methods (기계학습기법을 이용한 부산-울산-경남 지역의 증발수요 가뭄지수 예측)

  • Lee, Okjeong;Won, Jeongeun;Seo, Jiyu;Kim, Sangdan
    • Journal of Korea Water Resources Association
    • /
    • v.54 no.8
    • /
    • pp.617-628
    • /
    • 2021
  • Drought is a major natural disaster that causes serious social and economic losses. Local drought forecasts can provide important information for drought preparedness. In this study, we propose a new machine learning model that predicts drought by using historical drought indices and meteorological data from 10 sites from 1981 to 2020 in the southeastern part of the Korean Peninsula, Busan-Ulsan-Gyeongnam. Using Bayesian optimization techniques, a hyper-parameter-tuned Random Forest, XGBoost, and Light GBM model were constructed to predict the evaporative demand drought index on a 6-month time scale after 1-month. The model performance was compared by constructing a single site model and a regional model, respectively. In addition, the possibility of improving the model performance was examined by constructing a fine-tuned model using data from a individual site based on the regional model.

Forecasting Baltic Dry Index by Implementing Time-Series Decomposition and Data Augmentation Techniques (시계열 분해 및 데이터 증강 기법 활용 건화물운임지수 예측)

  • Han, Min Soo;Yu, Song Jin
    • Journal of Korean Society for Quality Management
    • /
    • v.50 no.4
    • /
    • pp.701-716
    • /
    • 2022
  • Purpose: This study aims to predict the dry cargo transportation market economy. The subject of this study is the BDI (Baltic Dry Index) time-series, an index representing the dry cargo transport market. Methods: In order to increase the accuracy of the BDI time-series, we have pre-processed the original time-series via time-series decomposition and data augmentation techniques and have used them for ANN learning. The ANN algorithms used are Multi-Layer Perceptron (MLP), Recurrent Neural Network (RNN), and Long Short-Term Memory (LSTM) to compare and analyze the case of learning and predicting by applying time-series decomposition and data augmentation techniques. The forecast period aims to make short-term predictions at the time of t+1. The period to be studied is from '22. 01. 07 to '22. 08. 26. Results: Only for the case of the MAPE (Mean Absolute Percentage Error) indicator, all ANN models used in the research has resulted in higher accuracy (1.422% on average) in multivariate prediction. Although it is not a remarkable improvement in prediction accuracy compared to uni-variate prediction results, it can be said that the improvement in ANN prediction performance has been achieved by utilizing time-series decomposition and data augmentation techniques that were significant and targeted throughout this study. Conclusion: Nevertheless, due to the nature of ANN, additional performance improvements can be expected according to the adjustment of the hyper-parameter. Therefore, it is necessary to try various applications of multiple learning algorithms and ANN optimization techniques. Such an approach would help solve problems with a small number of available data, such as the rapidly changing business environment or the current shipping market.

AutoFe-Sel: A Meta-learning based methodology for Recommending Feature Subset Selection Algorithms

  • Irfan Khan;Xianchao Zhang;Ramesh Kumar Ayyasam;Rahman Ali
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.7
    • /
    • pp.1773-1793
    • /
    • 2023
  • Automated machine learning, often referred to as "AutoML," is the process of automating the time-consuming and iterative procedures that are associated with the building of machine learning models. There have been significant contributions in this area across a number of different stages of accomplishing a data-mining task, including model selection, hyper-parameter optimization, and preprocessing method selection. Among them, preprocessing method selection is a relatively new and fast growing research area. The current work is focused on the recommendation of preprocessing methods, i.e., feature subset selection (FSS) algorithms. One limitation in the existing studies regarding FSS algorithm recommendation is the use of a single learner for meta-modeling, which restricts its capabilities in the metamodeling. Moreover, the meta-modeling in the existing studies is typically based on a single group of data characterization measures (DCMs). Nonetheless, there are a number of complementary DCM groups, and their combination will allow them to leverage their diversity, resulting in improved meta-modeling. This study aims to address these limitations by proposing an architecture for preprocess method selection that uses ensemble learning for meta-modeling, namely AutoFE-Sel. To evaluate the proposed method, we performed an extensive experimental evaluation involving 8 FSS algorithms, 3 groups of DCMs, and 125 datasets. Results show that the proposed method achieves better performance compared to three baseline methods. The proposed architecture can also be easily extended to other preprocessing method selections, e.g., noise-filter selection and imbalance handling method selection.

An Extended Function Point Model for Estimating the Implementing Cost of Machine Learning Applications (머신러닝 애플리케이션 구현 비용 평가를 위한 확장형 기능 포인트 모델)

  • Seokjin Im
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.2
    • /
    • pp.475-481
    • /
    • 2023
  • Softwares, especially like machine learning applications, affect human's life style tremendously. Accordingly, the importance of the cost model for softwares increases rapidly. As cost models, LOC(Line of Code) and M/M(Man-Month) estimates the quantitative aspects of the software. Differently from them, FP(Function Point) focuses on estimating the functional characteristics of software. FP is efficient in the aspect that it estimates qualitative characteristics. FP, however, has a limit for evaluating machine learning softwares because FP does not evaluate the critical factors of machine learning software. In this paper, we propose an extended function point(ExFP) that extends FP to adopt hyper parameter and the complexity of its optimization as the characteristics of the machine learning applications. In the evaluation reflecting the characteristics of machine learning applications. we reveals the effectiveness of the proposed ExFP.