• Title/Summary/Keyword: HyperParameter

Search Result 111, Processing Time 0.031 seconds

Hyper-Parameter in Hidden Markov Random Field

  • Lim, Jo-Han;Yu, Dong-Hyeon;Pyu, Kyung-Suk
    • The Korean Journal of Applied Statistics
    • /
    • v.24 no.1
    • /
    • pp.177-183
    • /
    • 2011
  • Hidden Markov random eld(HMRF) is one of the most common model for image segmentation which is an important preprocessing in many imaging devices. The HMRF has unknown hyper-parameters on Markov random field to be estimated in segmenting testing images. However, in practice, due to computational complexity, it is often assumed to be a fixed constant. In this paper, we numerically show that the segmentation results very depending on the fixed hyper-parameter, and, if the parameter is misspecified, they further depend on the choice of the class-labelling algorithm. In contrast, the HMRF with estimated hyper-parameter provides consistent segmentation results regardless of the choice of class labelling and the estimation method. Thus, we recommend practitioners estimate the hyper-parameter even though it is computationally complex.

A Study on the Hyper-parameter Optimization of Bitcoin Price Prediction LSTM Model (비트코인 가격 예측을 위한 LSTM 모델의 Hyper-parameter 최적화 연구)

  • Kim, Jun-Ho;Sung, Hanul
    • Journal of the Korea Convergence Society
    • /
    • v.13 no.4
    • /
    • pp.17-24
    • /
    • 2022
  • Bitcoin is a peer-to-peer cryptocurrency designed for electronic transactions that do not depend on the government or financial institutions. Since Bitcoin was first issued, a huge blockchain financial market has been created, and as a result, research to predict Bitcoin price data using machine learning has been increasing. However, the inefficient Hyper-parameter optimization process of machine learning research is interrupting the progress of the research. In this paper, we analyzes and presents the direction of Hyper-parameter optimization through experiments that compose the entire combination of the Timesteps, the number of LSTM units, and the Dropout ratio among the most representative Hyper-parameter and measure the predictive performance for each combination based on Bitcoin price prediction model using LSTM layer.

Effects of Hyper-parameters and Dataset on CNN Training

  • Nguyen, Huu Nhan;Lee, Chanho
    • Journal of IKEEE
    • /
    • v.22 no.1
    • /
    • pp.14-20
    • /
    • 2018
  • The purpose of training a convolutional neural network (CNN) is to obtain weight factors that give high classification accuracies. The initial values of hyper-parameters affect the training results, and it is important to train a CNN with a suitable hyper-parameter set of a learning rate, a batch size, the initialization of weight factors, and an optimizer. We investigate the effects of a single hyper-parameter while others are fixed in order to obtain a hyper-parameter set that gives higher classification accuracies and requires shorter training time using a proposed VGG-like CNN for training since the VGG is widely used. The CNN is trained for four datasets of CIFAR10, CIFAR100, GTSRB and DSDL-DB. The effects of the normalization and the data transformation for datasets are also investigated, and a training scheme using merged datasets is proposed.

Hyper-parameter Optimization for Monte Carlo Tree Search using Self-play

  • Lee, Jin-Seon;Oh, Il-Seok
    • Smart Media Journal
    • /
    • v.9 no.4
    • /
    • pp.36-43
    • /
    • 2020
  • The Monte Carlo tree search (MCTS) is a popular method for implementing an intelligent game program. It has several hyper-parameters that require an optimization for showing the best performance. Due to the stochastic nature of the MCTS, the hyper-parameter optimization is difficult to solve. This paper uses the self-playing capability of the MCTS-based game program for optimizing the hyper-parameters. It seeks a winner path over the hyper-parameter space while performing the self-play. The top-q longest winners in the winner path compete for the final winner. The experiment using the 15-15-5 game (Omok in Korean name) showed a promising result.

Optimal Hyper Parameter for Korean Face Data Generation with BEGAN (BEGAN을 통해 한국인 얼굴 데이터 생성을 하는데 최적의 HyperParameter)

  • Cho, Kyu Cheol;Kim, San
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2021.07a
    • /
    • pp.459-460
    • /
    • 2021
  • 본 논문에서는 BEGAN을 활용한 한국인 얼굴 데이터 생성을 위한 최적의 Hyper Parameter를 제안한다. 연구에서는 GAN의 발전된 모델인 BEGAN을 이용한다. 위의 모델을 작성하기 위하여 본 논문에서는 Anaconda 기반의 Jupyter Notebook에서 Python Tensorflow 모델을 작성하여 테스트하고, 만들어진 모델을 FID를 통해 모델의 성능을 비교한다. 본 연구에서는 제안하는 방법들을 통해서 만들어진 모델을 이용해 한국인 얼굴 데이터를 구하고, 생성된 이미지에 대한 정량적인 평가를 진행한다.

  • PDF

Comparison of Hyper-Parameter Optimization Methods for Deep Neural Networks

  • Kim, Ho-Chan;Kang, Min-Jae
    • Journal of IKEEE
    • /
    • v.24 no.4
    • /
    • pp.969-974
    • /
    • 2020
  • Research into hyper parameter optimization (HPO) has recently revived with interest in models containing many hyper parameters, such as deep neural networks. In this paper, we introduce the most widely used HPO methods, such as grid search, random search, and Bayesian optimization, and investigate their characteristics through experiments. The MNIST data set is used to compare results in experiments to find the best method that can be used to achieve higher accuracy in a relatively short time simulation. The learning rate and weight decay have been chosen for this experiment because these are the commonly used parameters in this kind of experiment.

On the actual coverage probability of hypergeometric parameter (초기하분포의 모수에 대한 신뢰구간추정)

  • Kim, Dae-Hak
    • Journal of the Korean Data and Information Science Society
    • /
    • v.21 no.6
    • /
    • pp.1109-1115
    • /
    • 2010
  • In this paper, exact confidence interval of hyper-geometric parameter, that is the probability of success p in the population is discussed. Usually, binomial distribution is a well known discrete distribution with abundant usage. Hypergeometric distribution frequently replaces a binomial distribution when it is desirable to make allowance for the finiteness of the population size. For example, an application of the hypergeometric distribution arises in describing a probability model for the number of children attacked by an infectious disease, when a fixed number of them are exposed to it. Exact confidence interval estimation of hypergeometric parameter is reviewed. We consider the performance of exact confidence interval estimates of hypergeometric parameter in terms of actual coverage probability by small sample Monte Carlo simulation.

Generative AI parameter tuning for online self-directed learning

  • Jin-Young Jun;Youn-A Min
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.4
    • /
    • pp.31-38
    • /
    • 2024
  • This study proposes hyper-parameter settings for developing a generative AI-based learning support tool to facilitate programming education in online distance learning. We implemented an experimental tool that can set research hyper-parameters according to three different learning contexts, and evaluated the quality of responses from the generative AI using the tool. The experiment with the default hyper-parameter settings of the generative AI was used as the control group, and the experiment with the research hyper-parameters was used as the experimental group. The experiment results showed no significant difference between the two groups in the "Learning Support" context. However, in other two contexts ("Code Generation" and "Comment Generation"), it showed the average evaluation scores of the experimental group were found to be 11.6% points and 23% points higher than those of the control group respectively. Lastly, this study also observed that when the expected influence of response on learning motivation was presented in the 'system content', responses containing emotional support considering learning emotions were generated.

Multi-Class Classification Framework for Brain Tumor MR Image Classification by Using Deep CNN with Grid-Search Hyper Parameter Optimization Algorithm

  • Mukkapati, Naveen;Anbarasi, MS
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.4
    • /
    • pp.101-110
    • /
    • 2022
  • Histopathological analysis of biopsy specimens is still used for diagnosis and classifying the brain tumors today. The available procedures are intrusive, time consuming, and inclined to human error. To overcome these disadvantages, need of implementing a fully automated deep learning-based model to classify brain tumor into multiple classes. The proposed CNN model with an accuracy of 92.98 % for categorizing tumors into five classes such as normal tumor, glioma tumor, meningioma tumor, pituitary tumor, and metastatic tumor. Using the grid search optimization approach, all of the critical hyper parameters of suggested CNN framework were instantly assigned. Alex Net, Inception v3, Res Net -50, VGG -16, and Google - Net are all examples of cutting-edge CNN models that are compared to the suggested CNN model. Using huge, publicly available clinical datasets, satisfactory classification results were produced. Physicians and radiologists can use the suggested CNN model to confirm their first screening for brain tumor Multi-classification.

Hyper Parameter Tuning Method based on Sampling for Optimal LSTM Model

  • Kim, Hyemee;Jeong, Ryeji;Bae, Hyerim
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.1
    • /
    • pp.137-143
    • /
    • 2019
  • As the performance of computers increases, the use of deep learning, which has faced technical limitations in the past, is becoming more diverse. In many fields, deep learning has contributed to the creation of added value and used on the bases of more data as the application become more divers. The process for obtaining a better performance model will require a longer time than before, and therefore it will be necessary to find an optimal model that shows the best performance more quickly. In the artificial neural network modeling a tuning process that changes various elements of the neural network model is used to improve the model performance. Except Gride Search and Manual Search, which are widely used as tuning methods, most methodologies have been developed focusing on heuristic algorithms. The heuristic algorithm can get the results in a short time, but the results are likely to be the local optimal solution. Obtaining a global optimal solution eliminates the possibility of a local optimal solution. Although the Brute Force Method is commonly used to find the global optimal solution, it is not applicable because of an infinite number of hyper parameter combinations. In this paper, we use a statistical technique to reduce the number of possible cases, so that we can find the global optimal solution.