• Title/Summary/Keyword: hyper-parameters

Search Result 188, Processing Time 0.031 seconds

Effects of Hyper-parameters and Dataset on CNN Training

  • Nguyen, Huu Nhan;Lee, Chanho
    • Journal of IKEEE
    • /
    • v.22 no.1
    • /
    • pp.14-20
    • /
    • 2018
  • The purpose of training a convolutional neural network (CNN) is to obtain weight factors that give high classification accuracies. The initial values of hyper-parameters affect the training results, and it is important to train a CNN with a suitable hyper-parameter set of a learning rate, a batch size, the initialization of weight factors, and an optimizer. We investigate the effects of a single hyper-parameter while others are fixed in order to obtain a hyper-parameter set that gives higher classification accuracies and requires shorter training time using a proposed VGG-like CNN for training since the VGG is widely used. The CNN is trained for four datasets of CIFAR10, CIFAR100, GTSRB and DSDL-DB. The effects of the normalization and the data transformation for datasets are also investigated, and a training scheme using merged datasets is proposed.

Hyper-parameter Optimization for Monte Carlo Tree Search using Self-play

  • Lee, Jin-Seon;Oh, Il-Seok
    • Smart Media Journal
    • /
    • v.9 no.4
    • /
    • pp.36-43
    • /
    • 2020
  • The Monte Carlo tree search (MCTS) is a popular method for implementing an intelligent game program. It has several hyper-parameters that require an optimization for showing the best performance. Due to the stochastic nature of the MCTS, the hyper-parameter optimization is difficult to solve. This paper uses the self-playing capability of the MCTS-based game program for optimizing the hyper-parameters. It seeks a winner path over the hyper-parameter space while performing the self-play. The top-q longest winners in the winner path compete for the final winner. The experiment using the 15-15-5 game (Omok in Korean name) showed a promising result.

BOUNDS ON THE HYPER-ZAGREB INDEX

  • FALAHATI-NEZHAD, FARZANEH;AZARI, MAHDIEH
    • Journal of applied mathematics & informatics
    • /
    • v.34 no.3_4
    • /
    • pp.319-330
    • /
    • 2016
  • The hyper-Zagreb index HM(G) of a simple graph G is defined as the sum of the terms (du+dv)2 over all edges uv of G, where du denotes the degree of the vertex u of G. In this paper, we present several upper and lower bounds on the hyper-Zagreb index in terms of some molecular structural parameters and relate this index to various well-known molecular descriptors.

Comparison of Hyper-Parameter Optimization Methods for Deep Neural Networks

  • Kim, Ho-Chan;Kang, Min-Jae
    • Journal of IKEEE
    • /
    • v.24 no.4
    • /
    • pp.969-974
    • /
    • 2020
  • Research into hyper parameter optimization (HPO) has recently revived with interest in models containing many hyper parameters, such as deep neural networks. In this paper, we introduce the most widely used HPO methods, such as grid search, random search, and Bayesian optimization, and investigate their characteristics through experiments. The MNIST data set is used to compare results in experiments to find the best method that can be used to achieve higher accuracy in a relatively short time simulation. The learning rate and weight decay have been chosen for this experiment because these are the commonly used parameters in this kind of experiment.

A Study on Abnormal Data Processing Process of LSTM AE - With applying Data based Intelligent Factory

  • Youn-A Min
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.15 no.2
    • /
    • pp.240-247
    • /
    • 2023
  • In this paper, effective data management in industrial sites such as intelligent factories using time series data was studied. For effective management of time series data, variables considering the significance of the data were used, and hyper parameters calculated through LSTM AE were applied. We propose an optimized modeling considering the importance of each data section, and through this, outlier data of time series data can be efficiently processed. In the case of applying data significance and applying hyper parameters to which the research in this paper was applied, it was confirmed that the error rate was measured at 5.4%/4.8%/3.3%, and the significance of each data section and the significance of applying hyper parameters to optimize modeling were confirmed.

Hyper-Parameter in Hidden Markov Random Field

  • Lim, Jo-Han;Yu, Dong-Hyeon;Pyu, Kyung-Suk
    • The Korean Journal of Applied Statistics
    • /
    • v.24 no.1
    • /
    • pp.177-183
    • /
    • 2011
  • Hidden Markov random eld(HMRF) is one of the most common model for image segmentation which is an important preprocessing in many imaging devices. The HMRF has unknown hyper-parameters on Markov random field to be estimated in segmenting testing images. However, in practice, due to computational complexity, it is often assumed to be a fixed constant. In this paper, we numerically show that the segmentation results very depending on the fixed hyper-parameter, and, if the parameter is misspecified, they further depend on the choice of the class-labelling algorithm. In contrast, the HMRF with estimated hyper-parameter provides consistent segmentation results regardless of the choice of class labelling and the estimation method. Thus, we recommend practitioners estimate the hyper-parameter even though it is computationally complex.

Kernel method for autoregressive data

  • Shim, Joo-Yong;Lee, Jang-Taek
    • Journal of the Korean Data and Information Science Society
    • /
    • v.20 no.5
    • /
    • pp.949-954
    • /
    • 2009
  • The autoregressive process is applied in this paper to kernel regression in order to infer nonlinear models for predicting responses. We propose a kernel method for the autoregressive data which estimates the mean function by kernel machines. We also present the model selection method which employs the cross validation techniques for choosing the hyper-parameters which affect the performance of kernel regression. Artificial and real examples are provided to indicate the usefulness of the proposed method for the estimation of mean function in the presence of autocorrelation between data.

  • PDF

Generative AI parameter tuning for online self-directed learning

  • Jin-Young Jun;Youn-A Min
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.4
    • /
    • pp.31-38
    • /
    • 2024
  • This study proposes hyper-parameter settings for developing a generative AI-based learning support tool to facilitate programming education in online distance learning. We implemented an experimental tool that can set research hyper-parameters according to three different learning contexts, and evaluated the quality of responses from the generative AI using the tool. The experiment with the default hyper-parameter settings of the generative AI was used as the control group, and the experiment with the research hyper-parameters was used as the experimental group. The experiment results showed no significant difference between the two groups in the "Learning Support" context. However, in other two contexts ("Code Generation" and "Comment Generation"), it showed the average evaluation scores of the experimental group were found to be 11.6% points and 23% points higher than those of the control group respectively. Lastly, this study also observed that when the expected influence of response on learning motivation was presented in the 'system content', responses containing emotional support considering learning emotions were generated.

Fuzzy neural network modeling using hyper elliptic gaussian membership functions (초타원 가우시안 소속함수를 사용한 퍼지신경망 모델링)

  • 권오국;주영훈;박진배
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1997.10a
    • /
    • pp.442-445
    • /
    • 1997
  • We present a hybrid self-tuning method of fuzzy inference systems with hyper elliptic Gaussian membership functions using genetic algorithm(GA) and back-propagation algorithm. The proposed self-tuning method has two phases : one is the coarse tuning process based on GA and the other is the fine tuning process based on back-propagation. But the parameters which is obtained by a GA are near optimal solutions. In order to solve the problem in GA applications, it uses a back-propagation algorithm, which is one of learning algorithms in neural networks, to finely tune the parameters obtained by a GA. We provide Box-Jenkins time series to evaluate the advantage and effectiveness of the proposed approach and compare with the conventional method.

  • PDF

Critical Parameters governing on the Fatigue Properties in the Hyper-eutectoid Steel Wires used for Automotive Tire (고강도 극 세선의 피로 특성 향상을 위한 특정 인자 제시)

  • Yang, Y.S.;Bae, J.G.;Park, C.G.
    • Proceedings of the Korean Society for Technology of Plasticity Conference
    • /
    • 2007.10a
    • /
    • pp.124-127
    • /
    • 2007
  • In this study, we focused on investigation of governing parameters affected on the fatigue properties in the hyper-eutectoid steel wires used for TBR tires. Steel wires are fabricated under different drawing strain from 3.36 to 3.80. Their diameters are 0.21 mm and 0.185mm, respectively. The fatigue properties was measured by hunter rotating beam tester, specially designed thin-sized steel wires. The results showed that the fatigue properites of steel wire, marked as A-1, were greater than the others, due to the low value of residual stress. In order to elucidate the variations of fatigue properties, the microstructure, surface defect and residual stress were observed and measured by useful analysis technique, such as TEM, 3D profiler and FIB.

  • PDF