• Title/Summary/Keyword: 하이퍼파라미터 최적화

Search Result 35, Processing Time 0.028 seconds

Bayesian Optimization Framework for Improved Cross-Version Defect Prediction (향상된 교차 버전 결함 예측을 위한 베이지안 최적화 프레임워크)

  • Choi, Jeongwhan;Ryu, Duksan
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.10 no.9
    • /
    • pp.339-348
    • /
    • 2021
  • In recent software defect prediction research, defect prediction between cross projects and cross-version projects are actively studied. Cross-version defect prediction studies assume WP(Within-Project) so far. However, in the CV(Cross-Version) environment, the previous work does not consider the distribution difference between project versions is important. In this study, we propose an automated Bayesian optimization framework that considers distribution differences between different versions. Through this, it automatically selects whether to perform transfer learning according to the difference in distribution. This framework is a technique that optimizes the distribution difference between versions, transfer learning, and hyper-parameters of the classifier. We confirmed that the method of automatically selecting whether to perform transfer learning based on the distribution difference is effective through experiments. Moreover, we can see that using our optimization framework is effective in improving performance and, as a result, can reduce software inspection effort. This is expected to support practical quality assurance activities for new version projects in a cross-version project environment.

A Performance Analysis by Adjusting Learning Methods in Stock Price Prediction Model Using LSTM (LSTM을 이용한 주가예측 모델의 학습방법에 따른 성능분석)

  • Jung, Jongjin;Kim, Jiyeon
    • Journal of Digital Convergence
    • /
    • v.18 no.11
    • /
    • pp.259-266
    • /
    • 2020
  • Many developments have been steadily carried out by researchers with applying knowledge-based expert system or machine learning algorithms to the financial field. In particular, it is now common to perform knowledge based system trading in using stock prices. Recently, deep learning technologies have been applied to real fields of stock trading marketplace as GPU performance and large scaled data have been supported enough. Especially, LSTM has been tried to apply to stock price prediction because of its compatibility for time series data. In this paper, we implement stock price prediction using LSTM. In modeling of LSTM, we propose a fitness combination of model parameters and activation functions for best performance. Specifically, we propose suitable selection methods of initializers of weights and bias, regularizers to avoid over-fitting, activation functions and optimization methods. We also compare model performances according to the different selections of the above important modeling considering factors on the real-world stock price data of global major companies. Finally, our experimental work brings a fitness method of applying LSTM model to stock price prediction.

Performance Comparison of Automatic Classification Using Word Embeddings of Book Titles (단행본 서명의 단어 임베딩에 따른 자동분류의 성능 비교)

  • Yong-Gu Lee
    • Journal of the Korean Society for information Management
    • /
    • v.40 no.4
    • /
    • pp.307-327
    • /
    • 2023
  • To analyze the impact of word embedding on book titles, this study utilized word embedding models (Word2vec, GloVe, fastText) to generate embedding vectors from book titles. These vectors were then used as classification features for automatic classification. The classifier utilized the k-nearest neighbors (kNN) algorithm, with the categories for automatic classification based on the DDC (Dewey Decimal Classification) main class 300 assigned by libraries to books. In the automatic classification experiment applying word embeddings to book titles, the Skip-gram architectures of Word2vec and fastText showed better results in the automatic classification performance of the kNN classifier compared to the TF-IDF features. In the optimization of various hyperparameters across the three models, the Skip-gram architecture of the fastText model demonstrated overall good performance. Specifically, better performance was observed when using hierarchical softmax and larger embedding dimensions as hyperparameters in this model. From a performance perspective, fastText can generate embeddings for substrings or subwords using the n-gram method, which has been shown to increase recall. The Skip-gram architecture of the Word2vec model generally showed good performance at low dimensions(size 300) and with small sizes of negative sampling (3 or 5).

A Study on the Drug Classification Using Machine Learning Techniques (머신러닝 기법을 이용한 약물 분류 방법 연구)

  • Anmol Kumar Singh;Ayush Kumar;Adya Singh;Akashika Anshum;Pradeep Kumar Mallick
    • Advanced Industrial SCIence
    • /
    • v.3 no.2
    • /
    • pp.8-16
    • /
    • 2024
  • This paper shows the system of drug classification, the goal of this is to foretell the apt drug for the patients based on their demographic and physiological traits. The dataset consists of various attributes like Age, Sex, BP (Blood Pressure), Cholesterol Level, and Na_to_K (Sodium to Potassium ratio), with the objective to determine the kind of drug being given. The models used in this paper are K-Nearest Neighbors (KNN), Logistic Regression and Random Forest. Further to fine-tune hyper parameters using 5-fold cross-validation, GridSearchCV was used and each model was trained and tested on the dataset. To assess the performance of each model both with and without hyper parameter tuning evaluation metrics like accuracy, confusion matrices, and classification reports were used and the accuracy of the models without GridSearchCV was 0.7, 0.875, 0.975 and with GridSearchCV was 0.75, 1.0, 0.975. According to GridSearchCV Logistic Regression is the most suitable model for drug classification among the three-model used followed by the K-Nearest Neighbors. Also, Na_to_K is an essential feature in predicting the outcome.

Sound Event Detection based on Deep Neural Networks (딥 뉴럴네트워크 기반의 소리 이벤트 검출)

  • Chung, Suk-Hwan;Chung, Yong-Joo
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.14 no.2
    • /
    • pp.389-396
    • /
    • 2019
  • In this paper, various architectures of deep neural networks were applied for sound event detection and their performances were compared using a common audio database. The FNN, CNN, RNN and CRNN were implemented using hyper-parameters optimized for the database as well as the architecture of each neural network. Among the implemented deep neural networks, CRNN performed best at all testing conditions and CNN followed CRNN in performance. Although RNN has a merit in tracking the time-correlations in audio signals, it showed poor performance compared with CNN and CRNN.

The Prediction Model Development for Water Supply Monitoring System based on Machine Learning (머신러닝을 고려한 상수도 모니터링 시스템 예측 모델 개발)

  • Shim, Kyu Dae;Choung, Joon Yeon;Kim, Chang Ryong;Kim, Dong Kyun
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2022.05a
    • /
    • pp.395-395
    • /
    • 2022
  • 본 연구는 머신러닝 기반의 상수도 모니터링 시스템의 예측 모델을 개발하고, 예측 모델의 적용이 가능성을 검토하였다. 상수도모니터링 시스템은 상수관망에 설치된 센서에서 수집된 자료를 모니터링 할 수 있어 운영자의 상수도 시설물의 관리 편의성을 높일 수 있다. 특히 수리학적 모델을 적용하여 계산된 값과 측정된 값을 비교해 이상치가 발생하면 운영자에게 이를 알려주므로 시스템내의 문제점을 빠르게 확인할 수 있다. 그러나 수리학적 모델은 입력자료가 증가됨에 따라 계산시간이 많이 소요되는 문제가 있고, 계산된 값의 정확도가 낮아지므로. 이러한 문제를 보완하기 위해 머신러닝 기반의 예측 모델을 개발하여 이를 해결하고자 하였다. 예측 모델은 GS 이니마 브라질(GS Inima Brazil)에서 운영중인 아라사투바(Aracatuba) 지역 주사라(Jussara) DMA(District Metered Area)의 2018년 1월에서 7월까지의 운영자료를 이용하였으며, 상수도 모니터링 시스템에서 상수관로 수압에 영향을 미치는 영향 인자들을 분석하고, 하이퍼파라미터 최적화를 통한 수압 예측 모델을 개선하였다. 금회 연구는 머신러닝 기반의 모델을 통하여 상수관망의 시간변화에 따른 장래 예측 수압을 검토할 수 있었다는데 큰 의의가 있다.

  • PDF

Hyperparameter optimization for Lightweight and Resource-Efficient Deep Learning Model in Human Activity Recognition using Short-range mmWave Radar (mmWave 레이더 기반 사람 행동 인식 딥러닝 모델의 경량화와 자원 효율성을 위한 하이퍼파라미터 최적화 기법)

  • Jiheon Kang
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.6
    • /
    • pp.319-325
    • /
    • 2023
  • In this study, we proposed a method for hyperparameter optimization in the building and training of a deep learning model designed to process point cloud data collected by a millimeter-wave radar system. The primary aim of this study is to facilitate the deployment of a baseline model in resource-constrained IoT devices. We evaluated a RadHAR baseline deep learning model trained on a public dataset composed of point clouds representing five distinct human activities. Additionally, we introduced a coarse-to-fine hyperparameter optimization procedure, showing substantial potential to enhance model efficiency without compromising predictive performance. Experimental results show the feasibility of significantly reducing model size without adversely impacting performance. Specifically, the optimized model demonstrated a 3.3% improvement in classification accuracy despite a 16.8% reduction in number of parameters compared th the baseline model. In conclusion, this research offers valuable insights for the development of deep learning models for resource-constrained IoT devices, underscoring the potential of hyperparameter optimization and model size reduction strategies. This work contributes to enhancing the practicality and usability of deep learning models in real-world environments, where high levels of accuracy and efficiency in data processing and classification tasks are required.

Performance Evaluation of Recurrent Neural Network Algorithms for Recommendation System in E-commerce (전자상거래 추천시스템을 위한 순환신경망 알고리즘들의 성능평가)

  • Seo, Jihye;Yong, Hwan-Seung
    • KIISE Transactions on Computing Practices
    • /
    • v.23 no.7
    • /
    • pp.440-445
    • /
    • 2017
  • Due to the advance of e-commerce systems, the number of people using online shopping and products has significantly increased. Therefore, the need for an accurate recommendation system is becoming increasingly more important. Recurrent neural network is a deep-learning algorithm that utilizes sequential information in training. In this paper, an evaluation is performed on the application of recurrent neural networks to recommendation systems. We evaluated three recurrent algorithms (RNN, LSTM and GRU) and three optimal algorithms(Adagrad, RMSProp and Adam) which are commonly used. In the experiments, we used the TensorFlow open source library produced by Google and e-commerce session data from RecSys Challenge 2015. The results using the optimal hyperparameters found in this study are compared with those of RecSys Challenge 2015 participants.

Proposal of a Step-by-Step Optimized Campus Power Forecast Model using CNN-LSTM Deep Learning (CNN-LSTM 딥러닝 기반 캠퍼스 전력 예측 모델 최적화 단계 제시)

  • Kim, Yein;Lee, Seeun;Kwon, Youngsung
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.10
    • /
    • pp.8-15
    • /
    • 2020
  • A forecasting method using deep learning does not have consistent results due to the differences in the characteristics of the dataset, even though they have the same forecasting models and parameters. For example, the forecasting model X optimized with dataset A would not produce the optimized result with another dataset B. The forecasting model with the characteristics of the dataset needs to be optimized to increase the accuracy of the forecasting model. Therefore, this paper proposes novel optimization steps for outlier removal, dataset classification, and a CNN-LSTM-based hyperparameter tuning process to forecast the daily power usage of a university campus based on the hourly interval. The proposing model produces high forecasting accuracy with a 2% of MAPE with a single power input variable. The proposing model can be used in EMS to suggest improved strategies to users and consequently to improve the power efficiency.

Prediction model for electric power consumption of seawater desalination based on machine learning by seawater quality change in future (장래 해수수질 변화에 따른 머신러닝 기반 해수담수 전력비 예측 모형 개발)

  • Shim, Kyudae;Ko, Young-Hee
    • Journal of Korea Water Resources Association
    • /
    • v.54 no.spc1
    • /
    • pp.1023-1035
    • /
    • 2021
  • The electricity cost of a desalination facility was also predicted and reviewed, which allowed the proposed model to be incorporated into the future design of such facilities. Input data from 2003 to 2014 of the Korea Hydrographic and Oceanographic Agency (KHOA) were used, and the structure of the model was determined using the trial and error method to analyze as well as hyperparameters such as salinity and seawater temperature. The future seawater quality was estimated by optimizing the prediction model based on machine learning. Results indicated that the seawater temperature would be similar to the existing pattern, and salinity showed a gradual decrease in the maximum value from the past measurement data. Therefore, it was reviewed that the electricity cost for seawater desalination decreased by approximately 0.80% and a process configuration was determined to be necessary. This study aimed at establishing a machine-learning-based prediction model to predict future water quality changes, reviewed the impact on the scale of seawater desalination facilities, and suggested alternatives.