• Title/Summary/Keyword: Regression Algorithms

Search Result 401, Processing Time 0.02 seconds

Performance Comparison of Logistic Regression Algorithms on RHadoop

  • Jung, Byung Ho;Lim, Dong Hoon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.22 no.4
    • /
    • pp.9-16
    • /
    • 2017
  • Machine learning has found widespread implementations and applications in many different domains in our life. Logistic regression is a type of classification in machine leaning, and is used widely in many fields, including medicine, economics, marketing and social sciences. In this paper, we present the MapReduce implementation of three existing algorithms, this is, Gradient Descent algorithm, Cost Minimization algorithm and Newton-Raphson algorithm, for logistic regression on RHadoop that integrates R and Hadoop environment applicable to large scale data. We compare the performance of these algorithms for estimation of logistic regression coefficients with real and simulated data sets. We also compare the performance of our RHadoop and RHIPE platforms. The performance experiments showed that our Newton-Raphson algorithm when compared to Gradient Descent and Cost Minimization algorithms appeared to be better to all data tested, also showed that our RHadoop was better than RHIPE in real data, and was opposite in simulated data.

Variable Selection with Regression Trees

  • Chang, Young-Jae
    • The Korean Journal of Applied Statistics
    • /
    • v.23 no.2
    • /
    • pp.357-366
    • /
    • 2010
  • Many tree algorithms have been developed for regression problems. Although they are regarded as good algorithms, most of them suffer from loss of prediction accuracy when there are many noise variables. To handle this problem, we propose the multi-step GUIDE, which is a regression tree algorithm with a variable selection process. The multi-step GUIDE performs better than some of the well-known algorithms such as Random Forest and MARS. The results based on simulation study shows that the multi-step GUIDE outperforms other algorithms in terms of variable selection and prediction accuracy. It generally selects the important variables correctly with relatively few noise variables and eventually gives good prediction accuracy.

Fast robust variable selection using VIF regression in large datasets (대형 데이터에서 VIF회귀를 이용한 신속 강건 변수선택법)

  • Seo, Han Son
    • The Korean Journal of Applied Statistics
    • /
    • v.31 no.4
    • /
    • pp.463-473
    • /
    • 2018
  • Variable selection algorithms for linear regression models of large data are considered. Many algorithms are proposed focusing on the speed and the robustness of algorithms. Among them variance inflation factor (VIF) regression is fast and accurate due to the use of a streamwise regression approach. But a VIF regression is susceptible to outliers because it estimates a model by a least-square method. A robust criterion using a weighted estimator has been proposed for the robustness of algorithm; in addition, a robust VIF regression has also been proposed for the same purpose. In this article a fast and robust variable selection method is suggested via a VIF regression with detecting and removing potential outliers. A simulation study and an analysis of a dataset are conducted to compare the suggested method with other methods.

A Fast Kernel Regression Framework for Video Super-Resolution

  • Yu, Wen-Sen;Wang, Ming-Hui;Chang, Hua-Wen;Chen, Shu-Qing
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.1
    • /
    • pp.232-248
    • /
    • 2014
  • A series of kernel regression (KR) algorithms, such as the classic kernel regression (CKR), the 2- and 3-D steering kernel regression (SKR), have been proposed for image and video super-resolution. In existing KR frameworks, a single algorithm is usually adopted and applied for a whole image/video, regardless of region characteristics. However, their performances and computational efficiencies can differ in regions of different characteristics. To take full advantage of the KR algorithms and avoid their disadvantage, this paper proposes a kernel regression framework for video super-resolution. In this framework, each video frame is first analyzed and divided into three types of regions: flat, non-flat-stationary, and non-flat-moving regions. Then different KR algorithm is selected according to the region type. The CKR and 2-D SKR algorithms are applied to flat and non-flat-stationary regions, respectively. For non-flat-moving regions, this paper proposes a similarity-assisted steering kernel regression (SASKR) algorithm, which can give better performance and higher computational efficiency than the 3-D SKR algorithm. Experimental results demonstrate that the computational efficiency of the proposed framework is greatly improved without apparent degradation in performance.

Predictiong long-term workers in the company using regression

  • SON, Ho Min;SEO, Jung Hwa
    • Korean Journal of Artificial Intelligence
    • /
    • v.10 no.1
    • /
    • pp.15-19
    • /
    • 2022
  • This study is to understand the relationship between turnover and various conditions. Turnover refers to workers moving from one company to another, which exists in various ways and forms. Currently, a large number of workers are considering many turnover rates to satisfy their income levels, distance between work and residence, and age. In addition, they consider changing jobs a lot depending on the type of work, the decision-making ability of workers, and the level of education. The company needs to accept the conditions required by workers so that competent workers can work for a long time and predict what measures should be taken to convert them into long-term workers. The study was conducted because it was necessary to predict what conditions workers must meet in order to become long-term workers by comparing various conditions and turnover using regression and decision trees. It used Microsoft Azure machines to produce results, and it found that among the various conditions, it looked for different items for long-term work. Various methods were attempted in conducting the research, and among them, suitable algorithms adopted algorithms that classify various kinds of algorithms and derive results, and among them, two decision tree algorithms were used to derive results.

An assessment of machine learning models for slump flow and examining redundant features

  • Unlu, Ramazan
    • Computers and Concrete
    • /
    • v.25 no.6
    • /
    • pp.565-574
    • /
    • 2020
  • Over the years, several machine learning approaches have been proposed and utilized to create a prediction model for the high-performance concrete (HPC) slump flow. Despite HPC is a highly complex material, predicting its pattern is a rather ambitious process. Hence, choosing and applying the correct method remain a crucial task. Like some other problems, prediction of HPC slump flow suffers from abnormal attributes which might both have an influence on prediction accuracy and increases variance. In recent years, different studies are proposed to optimize the prediction accuracy for HPC slump flow. However, more state-of-the-art regression algorithms can be implemented to create a better model. This study focuses on several methods with different mathematical backgrounds to get the best possible results. Four well-known algorithms Support Vector Regression, M5P Trees, Random Forest, and MLPReg are implemented with optimum parameters as base learners. Also, redundant features are examined to better understand both how ingredients influence on prediction models and whether possible to achieve acceptable results with a few components. Based on the findings, the MLPReg algorithm with optimum parameters gives better results than others in terms of commonly used statistical error evaluation metrics. Besides, chosen algorithms can give rather accurate results using just a few attributes of a slump flow dataset.

Least absolute deviation estimator based consistent model selection in regression

  • Shende, K.S.;Kashid, D.N.
    • Communications for Statistical Applications and Methods
    • /
    • v.26 no.3
    • /
    • pp.273-293
    • /
    • 2019
  • We consider the problem of model selection in multiple linear regression with outliers and non-normal error distributions. In this article, the robust model selection criterion is proposed based on the robust estimation method with the least absolute deviation (LAD). The proposed criterion is shown to be consistent. We suggest proposed criterion based algorithms that are suitable for a large number of predictors in the model. These algorithms select only relevant predictor variables with probability one for large sample sizes. An exhaustive simulation study shows that the criterion performs well. However, the proposed criterion is applied to a real data set to examine its applicability. The simulation results show the proficiency of algorithms in the presence of outliers, non-normal distribution, and multicollinearity.

Panel data analysis with regression trees (회귀나무 모형을 이용한 패널데이터 분석)

  • Chang, Youngjae
    • Journal of the Korean Data and Information Science Society
    • /
    • v.25 no.6
    • /
    • pp.1253-1262
    • /
    • 2014
  • Regression tree is a tree-structured solution in which a simple regression model is fitted to the data in each node made by recursive partitioning of predictor space. There have been many efforts to apply tree algorithms to various regression problems like logistic regression and quantile regression. Recently, algorithms have been expanded to the panel data analysis such as RE-EM algorithm by Sela and Simonoff (2012), and extension of GUIDE by Loh and Zheng (2013). The algorithms are briefly introduced and prediction accuracy of three methods are compared in this paper. In general, RE-EM shows good prediction accuracy with least MSE's in the simulation study. A RE-EM tree fitted to business survey index (BSI) panel data shows that sales BSI is the main factor which affects business entrepreneurs' economic sentiment. The economic sentiment BSI of non-manufacturing industries is higher than that of manufacturing ones among the relatively high sales group.

Response prediction of laced steel-concrete composite beams using machine learning algorithms

  • Thirumalaiselvi, A.;Verma, Mohit;Anandavalli, N.;Rajasankar, J.
    • Structural Engineering and Mechanics
    • /
    • v.66 no.3
    • /
    • pp.399-409
    • /
    • 2018
  • This paper demonstrates the potential application of machine learning algorithms for approximate prediction of the load and deflection capacities of the novel type of Laced Steel Concrete-Composite (LSCC) beams proposed by Anandavalli et al. (Engineering Structures 2012). Initially, global and local responses measured on LSCC beam specimen in an experiment are used to validate nonlinear FE model of the LSCC beams. The data for the machine learning algorithms is then generated using validated FE model for a range of values of the identified sensitive parameters. The performance of four well-known machine learning algorithms, viz., Support Vector Regression (SVR), Minimax Probability Machine Regression (MPMR), Relevance Vector Machine (RVM) and Multigene Genetic Programing (MGGP) for the approximate estimation of the load and deflection capacities are compared in terms of well-defined error indices. Through relative comparison of the estimated values, it is demonstrated that the algorithms explored in the present study provide a good alternative to expensive experimental testing and sophisticated numerical simulation of the response of LSCC beams. The load carrying and displacement capacity of the LSCC was predicted well by MGGP and MPMR, respectively.

Data Mining for Knowledge Management in a Health Insurance Domain

  • Chae, Young-Moon;Ho, Seung-Hee;Cho, Kyoung-Won;Lee, Dong-Ha;Ji, Sun-Ha
    • Journal of Intelligence and Information Systems
    • /
    • v.6 no.1
    • /
    • pp.73-82
    • /
    • 2000
  • This study examined the characteristicso f the knowledge discovery and data mining algorithms to demonstrate how they can be used to predict health outcomes and provide policy information for hypertension management using the Korea Medical Insurance Corporation database. Specifically this study validated the predictive power of data mining algorithms by comparing the performance of logistic regression and two decision tree algorithms CHAID (Chi-squared Automatic Interaction Detection) and C5.0 (a variant of C4.5) since logistic regression has assumed a major position in the healthcare field as a method for predicting or classifying health outcomes based on the specific characteristics of each individual case. This comparison was performed using the test set of 4,588 beneficiaries and the training set of 13,689 beneficiaries that were used to develop the models. On the contrary to the previous study CHAID algorithm performed better than logistic regression in predicting hypertension but C5.0 had the lowest predictive power. In addition CHAID algorithm and association rule also provided the segment characteristics for the risk factors that may be used in developing hypertension management programs. This showed that data mining approach can be a useful analytic tool for predicting and classifying health outcomes data.

  • PDF