• Title/Summary/Keyword: Shrinkage Estimator

Search Result 31, Processing Time 0.032 seconds

A Small Area Estimation for Monthly Wage Using Mean Squared Percentage Error (MSPE를 이용한 임금총액 소지역 추정)

  • Hwang, Hee-Jin;Shin, Key-Il
    • The Korean Journal of Applied Statistics
    • /
    • v.22 no.2
    • /
    • pp.403-414
    • /
    • 2009
  • Many researches have been devoted to the small area estimation related with the area level statistics. Almost all of the small area estimation methods are derived based on minimization of mean squared error(MSE). Recently Hwang and Shin (2008) suggested an alternative small area estimation method by minimizing mean squared percentage error. In this paper we apply this small area estimation method to the labor statistics, especially monthly wages by a branch area of labor department. The Monthly Labor Survey data (2007) is used for analysis and comparison of these methods.

A Comparative Study of Covariance Matrix Estimators in High-Dimensional Data (고차원 데이터에서 공분산행렬의 추정에 대한 비교연구)

  • Lee, DongHyuk;Lee, Jae Won
    • The Korean Journal of Applied Statistics
    • /
    • v.26 no.5
    • /
    • pp.747-758
    • /
    • 2013
  • The covariance matrix is important in multivariate statistical analysis and a sample covariance matrix is used as an estimator of the covariance matrix. High dimensional data has a larger dimension than the sample size; therefore, the sample covariance matrix may not be suitable since it is known to perform poorly and event not invertible. A number of covariance matrix estimators have been recently proposed with three different approaches of shrinkage, thresholding, and modified Cholesky decomposition. We compare the performance of these newly proposed estimators in various situations.

Estimators Shrinking towards Projection Vector for Multivariate Normal Mean Vector under the Norm with a Known Interval

  • Baek, Hoh Yoo
    • Journal of Integrative Natural Science
    • /
    • v.11 no.3
    • /
    • pp.154-160
    • /
    • 2018
  • Consider the problem of estimating a $p{\times}1$ mean vector ${\theta}(p-r{\geq}3)$, r = rank(K) with a projection matrix K under the quadratic loss, based on a sample $Y_1$, $Y_2$, ${\cdots}$, $Y_n$. In this paper a James-Stein type estimator with shrinkage form is given when it's variance distribution is specified and when the norm ${\parallel}{\theta}-K{\theta}{\parallel}$ is constrain, where K is an idempotent and symmetric matrix and rank(K) = r. It is characterized a minimal complete class of James-Stein type estimators in this case. And the subclass of James-Stein type estimators that dominate the sample mean is derived.

Bias-reduced ℓ1-trend filtering

  • Donghyeon Yu;Johan Lim;Won Son
    • Communications for Statistical Applications and Methods
    • /
    • v.30 no.2
    • /
    • pp.149-162
    • /
    • 2023
  • The ℓ1-trend filtering method is one of the most widely used methods for extracting underlying trends from noisy observations. Contrary to the Hodrick-Prescott filtering, the ℓ1-trend filtering gives piecewise linear trends. One of the advantages of the ℓ1-trend filtering is that it can be used for identifying change points in piecewise linear trends. However, since the ℓ1-trend filtering employs total variation as a penalty term, estimated piecewise linear trends tend to be biased. In this study, we demonstrate the biasedness of the ℓ1-trend filtering in trend level estimation and propose a two-stage bias-reduction procedure. The newly suggested estimator is based on the estimated change points of the ℓ1-trend filtering. Numerical examples illustrate that the proposed method yields less biased estimates for piecewise linear trends.

SMOOTH SINGULAR VALUE THRESHOLDING ALGORITHM FOR LOW-RANK MATRIX COMPLETION PROBLEM

  • Geunseop Lee
    • Journal of the Korean Mathematical Society
    • /
    • v.61 no.3
    • /
    • pp.427-444
    • /
    • 2024
  • The matrix completion problem is to predict missing entries of a data matrix using the low-rank approximation of the observed entries. Typical approaches to matrix completion problem often rely on thresholding the singular values of the data matrix. However, these approaches have some limitations. In particular, a discontinuity is present near the thresholding value, and the thresholding value must be manually selected. To overcome these difficulties, we propose a shrinkage and thresholding function that smoothly thresholds the singular values to obtain more accurate and robust estimation of the data matrix. Furthermore, the proposed function is differentiable so that the thresholding values can be adaptively calculated during the iterations using Stein unbiased risk estimate. The experimental results demonstrate that the proposed algorithm yields a more accurate estimation with a faster execution than other matrix completion algorithms in image inpainting problems.

On the Use of Adaptive Weights for the F-Norm Support Vector Machine

  • Bang, Sung-Wan;Jhun, Myoung-Shic
    • The Korean Journal of Applied Statistics
    • /
    • v.25 no.5
    • /
    • pp.829-835
    • /
    • 2012
  • When the input features are generated by factors in a classification problem, it is more meaningful to identify important factors, rather than individual features. The $F_{\infty}$-norm support vector machine(SVM) has been developed to perform automatic factor selection in classification. However, the $F_{\infty}$-norm SVM may suffer from estimation inefficiency and model selection inconsistency because it applies the same amount of shrinkage to each factor without assessing its relative importance. To overcome such a limitation, we propose the adaptive $F_{\infty}$-norm ($AF_{\infty}$-norm) SVM, which penalizes the empirical hinge loss by the sum of the adaptively weighted factor-wise $L_{\infty}$-norm penalty. The $AF_{\infty}$-norm SVM computes the weights by the 2-norm SVM estimator and can be formulated as a linear programming(LP) problem which is similar to the one of the $F_{\infty}$-norm SVM. The simulation studies show that the proposed $AF_{\infty}$-norm SVM improves upon the $F_{\infty}$-norm SVM in terms of classification accuracy and factor selection performance.

A Study for the Drivers of Movie Box-office Performance (영화흥행 영향요인 선택에 관한 연구)

  • Kim, Yon Hyong;Hong, Jeong Han
    • The Korean Journal of Applied Statistics
    • /
    • v.26 no.3
    • /
    • pp.441-452
    • /
    • 2013
  • This study analyzed the relationship between key film and a box office record success factors based on movies released in the first quarter of 2013 in Korea. An over-fitting problem can happen if there are too many explanatory variables inserted to regression model; in addition, there is a risk that the estimator is instable when there is multi-collinearity among the explanatory variables. For this reason, optimal variable selection based on high explanatory variables in box-office performance is of importance. Among the numerous ways to select variables, LASSO estimation applied by a generalized linear model has the smallest prediction error that can efficiently and quickly find variables with the highest explanatory power to box-office performance in order.

The Improved BAMS Filter for Image Denoising (영상 잡음제거를 위한 개선된 BAMS 필터)

  • Woo, Chang-Yong;Park, Nam-Chun
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.11 no.4
    • /
    • pp.270-277
    • /
    • 2010
  • The BAMS filter is a kind of wavelet shrinkage filter based on the Bayes estimators with no simulation, therefore it can be used for a real time filter. The denoising efficiency of BAMS filter is seriously affected by the estimated noise variance in each wavelet band. To remove noise in signals in existing BAMS filter, the noise variance is estimated by using the quartile of the finest level of details in the wavelet decomposition, and with this variance, the noise of the level is removed. In this paper, to remove the image noise includingodified quartile of the level of detail is proposed. And by these techniques, the image noises of mid and high frequency bands are removed, and the results showed that the increased PSNR of ab the midband noise, the noise variance estimation method using the monotonic transform and the mout 2[dB] and the effectiveness in denosing of low noise deviation images.

Intelligent System for the Prediction of Heart Diseases Using Machine Learning Algorithms with Anew Mixed Feature Creation (MFC) technique

  • Rawia Elarabi;Abdelrahman Elsharif Karrar;Murtada El-mukashfi El-taher
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.5
    • /
    • pp.148-162
    • /
    • 2023
  • Classification systems can significantly assist the medical sector by allowing for the precise and quick diagnosis of diseases. As a result, both doctors and patients will save time. A possible way for identifying risk variables is to use machine learning algorithms. Non-surgical technologies, such as machine learning, are trustworthy and effective in categorizing healthy and heart-disease patients, and they save time and effort. The goal of this study is to create a medical intelligent decision support system based on machine learning for the diagnosis of heart disease. We have used a mixed feature creation (MFC) technique to generate new features from the UCI Cleveland Cardiology dataset. We select the most suitable features by using Least Absolute Shrinkage and Selection Operator (LASSO), Recursive Feature Elimination with Random Forest feature selection (RFE-RF) and the best features of both LASSO RFE-RF (BLR) techniques. Cross-validated and grid-search methods are used to optimize the parameters of the estimator used in applying these algorithms. and classifier performance assessment metrics including classification accuracy, specificity, sensitivity, precision, and F1-Score, of each classification model, along with execution time and RMSE the results are presented independently for comparison. Our proposed work finds the best potential outcome across all available prediction models and improves the system's performance, allowing physicians to diagnose heart patients more accurately.

Wavelet Thresholding Techniques to Support Multi-Scale Decomposition for Financial Forecasting Systems

  • Shin, Taeksoo;Han, Ingoo
    • Proceedings of the Korea Database Society Conference
    • /
    • 1999.06a
    • /
    • pp.175-186
    • /
    • 1999
  • Detecting the features of significant patterns from their own historical data is so much crucial to good performance specially in time-series forecasting. Recently, a new data filtering method (or multi-scale decomposition) such as wavelet analysis is considered more useful for handling the time-series that contain strong quasi-cyclical components than other methods. The reason is that wavelet analysis theoretically makes much better local information according to different time intervals from the filtered data. Wavelets can process information effectively at different scales. This implies inherent support fer multiresolution analysis, which correlates with time series that exhibit self-similar behavior across different time scales. The specific local properties of wavelets can for example be particularly useful to describe signals with sharp spiky, discontinuous or fractal structure in financial markets based on chaos theory and also allows the removal of noise-dependent high frequencies, while conserving the signal bearing high frequency terms of the signal. To date, the existing studies related to wavelet analysis are increasingly being applied to many different fields. In this study, we focus on several wavelet thresholding criteria or techniques to support multi-signal decomposition methods for financial time series forecasting and apply to forecast Korean Won / U.S. Dollar currency market as a case study. One of the most important problems that has to be solved with the application of the filtering is the correct choice of the filter types and the filter parameters. If the threshold is too small or too large then the wavelet shrinkage estimator will tend to overfit or underfit the data. It is often selected arbitrarily or by adopting a certain theoretical or statistical criteria. Recently, new and versatile techniques have been introduced related to that problem. Our study is to analyze thresholding or filtering methods based on wavelet analysis that use multi-signal decomposition algorithms within the neural network architectures specially in complex financial markets. Secondly, through the comparison with different filtering techniques' results we introduce the present different filtering criteria of wavelet analysis to support the neural network learning optimization and analyze the critical issues related to the optimal filter design problems in wavelet analysis. That is, those issues include finding the optimal filter parameter to extract significant input features for the forecasting model. Finally, from existing theory or experimental viewpoint concerning the criteria of wavelets thresholding parameters we propose the design of the optimal wavelet for representing a given signal useful in forecasting models, specially a well known neural network models.

  • PDF