• Title/Summary/Keyword: forecasting the market size

Search Result 44, Processing Time 0.026 seconds

An Estimation on the Market Size of Aqua-cultured Flatfish in Korea (양식 넙치 중장기 시장 규모 추정)

  • Kim, Bae-Sung;Kim, Chung-Hyeon;Cho, Jae-Hwan;Lee, Nam-Su
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.16 no.11
    • /
    • pp.7781-7787
    • /
    • 2015
  • The purpose of This paper is to address the development on supply-demand outlook model of aqua-cultured korean Flatfish and introduces a projection of supply-demand and market prices during 2015-2017 using developed model. The supply-demand outlook model is composed as a partial equilibrium model of Korean Flat fish. Each equation in the model is estimated by the econometric techniques. A reviews of the demand-outlook model stability is also carried out by the references based on RMSPE. MAPE, and Theil's inequality coefficients. According to the reference of RMSPE, the error rates of the forecasting values of the aqua culture area, culturing quantity, production quantity, market price show less than 4%, The production quantity and farm price are predicted respectively to be 42,561MT and 10,191KW per kg in 2017.

Correlation Analyses of the Temperature Time Series Data from the Heat Box for Energy Modeling in the Automobile Drying Process (자동차 건조 공정 에너지 예측 모형을 위한 공조기 온도 시계열 데이터의 상관관계 분석)

  • Lee, Chang-Yong;Song, Gensoo;Kim, Jinho
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.37 no.2
    • /
    • pp.27-34
    • /
    • 2014
  • In this paper, we investigate the statistical correlation of the time series for temperature measured at the heat box in the automobile drying process. We show, in terms of the sample variance, that a significant non-linear correlation exists in the time series that consist of absolute temperature changes. To investigate further the non-linear correlation, we utilize the volatility, an important concept in the financial market, and induce volatility time series from absolute temperature changes. We analyze the time series of volatilities in terms of the de-trended fluctuation analysis (DFA), a method especially suitable for testing the long-range correlation of non-stationary data, from the correlation perspective. We uncover that the volatility exhibits a long-range correlation regardless of the window size. We also analyze the cross correlation between two (inlet and outlet) volatility time series to characterize any correlation between the two, and disclose the dependence of the correlation strength on the time lag. These results can contribute as important factors to the modeling of forecasting and management of the heat box's temperature.

Analysis of Productivity by Environmental Factors in Regional Base Public Hospitals (지역거점 공공병원의 환경적 요인에 따른 생산성 분석)

  • Lee, Jinwoo
    • Korea Journal of Hospital Management
    • /
    • v.22 no.3
    • /
    • pp.46-60
    • /
    • 2017
  • The purpose of this study is to analyze the difference of productivity according to environmental factors among 25 Regional base public hospitals. Also this study is to propose a method to improve the productivity of Regional base public hospitals in the future by improving the public performance and stable management performance by studying the productivity variables affecting profitability. The survey period was based on the last three years, and 25 Regional base public hospitals were selected for the survey. The dependent variable is the total capital medical marginal profitability and the medical profit marginal profitability which are the indicators of profitability. The independent variable, productivity, is classified into three indicators: capital productivity, labor productivity, and value added productivity. The ANOVA analysis method was used to analyze the productivity difference according to the frequency factor and the environmental factors of the Regional base public hospitals. Finally, we conducted a hierarchical regression analysis to examine the productivity variables affecting profitability. The results of this study showed that there were differences in productivity due to environmental factors such as hospital size, competition in the local medical market, and differences in management performance. The difference in productivity and profitability depending on the environmental factors suggests that it is difficult for Regional base public hospitals in each regional base to perform a balanced public service. In order to overcome this, it is necessary to provide balanced medical services such as government financial support expansion, regional medical demand forecasting and facility infrastructure construction.

Long-Term Projection of Demand for Reverse Mortgage Using the Bass Diffusion Model in Korea (Bass 확산모형을 활용한 국내 주택연금의 중·장기 수요예측)

  • Yang, Jin-Ah;Min, Daiki;Choi, Hyung-Suk
    • Journal of the Korean Operations Research and Management Science Society
    • /
    • v.42 no.1
    • /
    • pp.29-41
    • /
    • 2017
  • Korea is expected to become a super-aged society by 2050. Given an aging population and the increasing pressure for the early retirement, a sufficient social safety net for elderly population becomes important. The Korean government introduced public reverse mortgage program in 2007, which is a product for aging seniors and the elderly, The number of reverse mortgage subscribers has also steadily grown. The demand continues to grow, but the reverse mortgage over a long period of time is a highly uncertain and risky product in the position of guarantee or lending institution. Thus, suitable demand prediction of the reverse mortgage subscribers is necessary for stable and sustainable operation. This study uses a Bass diffusion model to forecast the long-term demand for reverse mortgage and provides insight into reverse mortgage by forecasting demand for stability and substantiality of the loan product. We represent the projections of new subscribers on the basis of the data obtained from Korea Housing Finance Corporation. Results show that potential market size of Korean reverse mortgage reaches approximately 760,000-1,160,000 households by 2020. We validate the results by comparing the estimate of the cumulative number of subscribers with that found in literature.

A sign test for random walk hypothesis based on slopes (기울기를 이용한 랜덤워크 부호검정)

  • Kim, Tae Yoon;Park, Cheolyong;Kim, Seul Gee;Kim, Chan Jin;Kim, Hyun;Yu, Ju Hyung;Jang, Kyung Min;Jang, Young Seok
    • Journal of the Korean Data and Information Science Society
    • /
    • v.25 no.2
    • /
    • pp.385-392
    • /
    • 2014
  • Random walk hypothesis is a hypothesis that explains theoretically the difficulty in forecasting in financial market. Various tests for the hypothesis have been developed so far but it is known that those tests suffer from low power and size distortion. In this article, a sign test based on slopes are suggested to overcome these difficulties. A simulation study is conducted to compare this test to the often used Dickey and Fuller (1979) test.

Developing an Entropic Drawdown-at-Risk (EDaR) Fluctuation Forecasting Model for Commodity Futures Market Using Entropy-Based Dependency and Causality Network Modularity (엔트로피 기반 인과관계 네트워크의 모듈성을 활용한 상품 선물 시장의 EDaR 변동 예측 모형 개발)

  • Choi, Insu;Kim, Woo Chang
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2022.05a
    • /
    • pp.370-373
    • /
    • 2022
  • 본 연구에서는 전이 엔트로피 개념을 활용하여 주요 상품 선물의 하방 리스크 지수의 정보 흐름을 바탕으로 한 인과관계 네트워크를 구성하였다. 그리고 구성된 네트워크를 활용하여 금융 시장을 분석하였으며, 또한 정보 흐름의 존재 여부를 바탕으로 상품 선물의 하방 리스크 지수의 예측력이 개선될 수 있는지 확인하고자 하였다. 이를 위하여 정보 불확실성의 감소량을 측정하는 전이 엔트로피를 인과관계의 측정 지표로 상정하였으며, 전이 엔트로피 측정 시 발생할 수 있는 유한크기효과(finite size effect)를 조정하는 데 있어서 효과적인 지표인 효율적 전이 엔트로피를 활용하여 정보 흐름 네트워크를 구성하였으며 이를 이용하여 금융 지수 간의 인과관계를 분석하고 EDaR 의 등락 예측에 활용하였다. 그 결과, 금융 시장 지수를 효율적 전이 엔트로피를 이용한 인과관계 네트워크를 활용하여 금융 시장의 복잡계 네트워크 분석이 가능함을 확인하였고, 구성된 네트워크를 활용하여 국내 금융 시장 등락 예측에 있어 더 적은 데이터 열을 활용하여 거의 유사한 예측 결과를 냄으로써 상품 선물 시장 관련 예측의 데이터 열 선택에 활용할 수 있음을 확인하였다.

An Accurate Cryptocurrency Price Forecasting using Reverse Walk-Forward Validation (역순 워크 포워드 검증을 이용한 암호화폐 가격 예측)

  • Ahn, Hyun;Jang, Baekcheol
    • Journal of Internet Computing and Services
    • /
    • v.23 no.4
    • /
    • pp.45-55
    • /
    • 2022
  • The size of the cryptocurrency market is growing. For example, market capitalization of bitcoin exceeded 500 trillion won. Accordingly, many studies have been conducted to predict the price of cryptocurrency, and most of them have similar methodology of predicting stock prices. However, unlike stock price predictions, machine learning become best model in cryptocurrency price predictions, conceptually cryptocurrency has no passive income from ownership, and statistically, cryptocurrency has at least three times higher liquidity than stocks. Thats why we argue that a methodology different from stock price prediction should be applied to cryptocurrency price prediction studies. We propose Reverse Walk-forward Validation (RWFV), which modifies Walk-forward Validation (WFV). Unlike WFV, RWFV measures accuracy for Validation by pinning the Validation dataset directly in front of the Test dataset in time series, and gradually increasing the size of the Training dataset in front of it in time series. Train data were cut according to the size of the Train dataset with the highest accuracy among all measured Validation accuracy, and then combined with Validation data to measure the accuracy of the Test data. Logistic regression analysis and Support Vector Machine (SVM) were used as the analysis model, and various algorithms and parameters such as L1, L2, rbf, and poly were applied for the reliability of our proposed RWFV. As a result, it was confirmed that all analysis models showed improved accuracy compared to existing studies, and on average, the accuracy increased by 1.23%p. This is a significant improvement in accuracy, given that most of the accuracy of cryptocurrency price prediction remains between 50% and 60% through previous studies.

Bankruptcy Forecasting Model using AdaBoost: A Focus on Construction Companies (적응형 부스팅을 이용한 파산 예측 모형: 건설업을 중심으로)

  • Heo, Junyoung;Yang, Jin Yong
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.1
    • /
    • pp.35-48
    • /
    • 2014
  • According to the 2013 construction market outlook report, the liquidation of construction companies is expected to continue due to the ongoing residential construction recession. Bankruptcies of construction companies have a greater social impact compared to other industries. However, due to the different nature of the capital structure and debt-to-equity ratio, it is more difficult to forecast construction companies' bankruptcies than that of companies in other industries. The construction industry operates on greater leverage, with high debt-to-equity ratios, and project cash flow focused on the second half. The economic cycle greatly influences construction companies. Therefore, downturns tend to rapidly increase the bankruptcy rates of construction companies. High leverage, coupled with increased bankruptcy rates, could lead to greater burdens on banks providing loans to construction companies. Nevertheless, the bankruptcy prediction model concentrated mainly on financial institutions, with rare construction-specific studies. The bankruptcy prediction model based on corporate finance data has been studied for some time in various ways. However, the model is intended for all companies in general, and it may not be appropriate for forecasting bankruptcies of construction companies, who typically have high liquidity risks. The construction industry is capital-intensive, operates on long timelines with large-scale investment projects, and has comparatively longer payback periods than in other industries. With its unique capital structure, it can be difficult to apply a model used to judge the financial risk of companies in general to those in the construction industry. Diverse studies of bankruptcy forecasting models based on a company's financial statements have been conducted for many years. The subjects of the model, however, were general firms, and the models may not be proper for accurately forecasting companies with disproportionately large liquidity risks, such as construction companies. The construction industry is capital-intensive, requiring significant investments in long-term projects, therefore to realize returns from the investment. The unique capital structure means that the same criteria used for other industries cannot be applied to effectively evaluate financial risk for construction firms. Altman Z-score was first published in 1968, and is commonly used as a bankruptcy forecasting model. It forecasts the likelihood of a company going bankrupt by using a simple formula, classifying the results into three categories, and evaluating the corporate status as dangerous, moderate, or safe. When a company falls into the "dangerous" category, it has a high likelihood of bankruptcy within two years, while those in the "safe" category have a low likelihood of bankruptcy. For companies in the "moderate" category, it is difficult to forecast the risk. Many of the construction firm cases in this study fell in the "moderate" category, which made it difficult to forecast their risk. Along with the development of machine learning using computers, recent studies of corporate bankruptcy forecasting have used this technology. Pattern recognition, a representative application area in machine learning, is applied to forecasting corporate bankruptcy, with patterns analyzed based on a company's financial information, and then judged as to whether the pattern belongs to the bankruptcy risk group or the safe group. The representative machine learning models previously used in bankruptcy forecasting are Artificial Neural Networks, Adaptive Boosting (AdaBoost) and, the Support Vector Machine (SVM). There are also many hybrid studies combining these models. Existing studies using the traditional Z-Score technique or bankruptcy prediction using machine learning focus on companies in non-specific industries. Therefore, the industry-specific characteristics of companies are not considered. In this paper, we confirm that adaptive boosting (AdaBoost) is the most appropriate forecasting model for construction companies by based on company size. We classified construction companies into three groups - large, medium, and small based on the company's capital. We analyzed the predictive ability of AdaBoost for each group of companies. The experimental results showed that AdaBoost has more predictive ability than the other models, especially for the group of large companies with capital of more than 50 billion won.

Optimization of Support Vector Machines for Financial Forecasting (재무예측을 위한 Support Vector Machine의 최적화)

  • Kim, Kyoung-Jae;Ahn, Hyun-Chul
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.4
    • /
    • pp.241-254
    • /
    • 2011
  • Financial time-series forecasting is one of the most important issues because it is essential for the risk management of financial institutions. Therefore, researchers have tried to forecast financial time-series using various data mining techniques such as regression, artificial neural networks, decision trees, k-nearest neighbor etc. Recently, support vector machines (SVMs) are popularly applied to this research area because they have advantages that they don't require huge training data and have low possibility of overfitting. However, a user must determine several design factors by heuristics in order to use SVM. For example, the selection of appropriate kernel function and its parameters and proper feature subset selection are major design factors of SVM. Other than these factors, the proper selection of instance subset may also improve the forecasting performance of SVM by eliminating irrelevant and distorting training instances. Nonetheless, there have been few studies that have applied instance selection to SVM, especially in the domain of stock market prediction. Instance selection tries to choose proper instance subsets from original training data. It may be considered as a method of knowledge refinement and it maintains the instance-base. This study proposes the novel instance selection algorithm for SVMs. The proposed technique in this study uses genetic algorithm (GA) to optimize instance selection process with parameter optimization simultaneously. We call the model as ISVM (SVM with Instance selection) in this study. Experiments on stock market data are implemented using ISVM. In this study, the GA searches for optimal or near-optimal values of kernel parameters and relevant instances for SVMs. This study needs two sets of parameters in chromosomes in GA setting : The codes for kernel parameters and for instance selection. For the controlling parameters of the GA search, the population size is set at 50 organisms and the value of the crossover rate is set at 0.7 while the mutation rate is 0.1. As the stopping condition, 50 generations are permitted. The application data used in this study consists of technical indicators and the direction of change in the daily Korea stock price index (KOSPI). The total number of samples is 2218 trading days. We separate the whole data into three subsets as training, test, hold-out data set. The number of data in each subset is 1056, 581, 581 respectively. This study compares ISVM to several comparative models including logistic regression (logit), backpropagation neural networks (ANN), nearest neighbor (1-NN), conventional SVM (SVM) and SVM with the optimized parameters (PSVM). In especial, PSVM uses optimized kernel parameters by the genetic algorithm. The experimental results show that ISVM outperforms 1-NN by 15.32%, ANN by 6.89%, Logit and SVM by 5.34%, and PSVM by 4.82% for the holdout data. For ISVM, only 556 data from 1056 original training data are used to produce the result. In addition, the two-sample test for proportions is used to examine whether ISVM significantly outperforms other comparative models. The results indicate that ISVM outperforms ANN and 1-NN at the 1% statistical significance level. In addition, ISVM performs better than Logit, SVM and PSVM at the 5% statistical significance level.

The Effect of Data Size on the k-NN Predictability: Application to Samsung Electronics Stock Market Prediction (데이터 크기에 따른 k-NN의 예측력 연구: 삼성전자주가를 사례로)

  • Chun, Se-Hak
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.239-251
    • /
    • 2019
  • Statistical methods such as moving averages, Kalman filtering, exponential smoothing, regression analysis, and ARIMA (autoregressive integrated moving average) have been used for stock market predictions. However, these statistical methods have not produced superior performances. In recent years, machine learning techniques have been widely used in stock market predictions, including artificial neural network, SVM, and genetic algorithm. In particular, a case-based reasoning method, known as k-nearest neighbor is also widely used for stock price prediction. Case based reasoning retrieves several similar cases from previous cases when a new problem occurs, and combines the class labels of similar cases to create a classification for the new problem. However, case based reasoning has some problems. First, case based reasoning has a tendency to search for a fixed number of neighbors in the observation space and always selects the same number of neighbors rather than the best similar neighbors for the target case. So, case based reasoning may have to take into account more cases even when there are fewer cases applicable depending on the subject. Second, case based reasoning may select neighbors that are far away from the target case. Thus, case based reasoning does not guarantee an optimal pseudo-neighborhood for various target cases, and the predictability can be degraded due to a deviation from the desired similar neighbor. This paper examines how the size of learning data affects stock price predictability through k-nearest neighbor and compares the predictability of k-nearest neighbor with the random walk model according to the size of the learning data and the number of neighbors. In this study, Samsung electronics stock prices were predicted by dividing the learning dataset into two types. For the prediction of next day's closing price, we used four variables: opening value, daily high, daily low, and daily close. In the first experiment, data from January 1, 2000 to December 31, 2017 were used for the learning process. In the second experiment, data from January 1, 2015 to December 31, 2017 were used for the learning process. The test data is from January 1, 2018 to August 31, 2018 for both experiments. We compared the performance of k-NN with the random walk model using the two learning dataset. The mean absolute percentage error (MAPE) was 1.3497 for the random walk model and 1.3570 for the k-NN for the first experiment when the learning data was small. However, the mean absolute percentage error (MAPE) for the random walk model was 1.3497 and the k-NN was 1.2928 for the second experiment when the learning data was large. These results show that the prediction power when more learning data are used is higher than when less learning data are used. Also, this paper shows that k-NN generally produces a better predictive power than random walk model for larger learning datasets and does not when the learning dataset is relatively small. Future studies need to consider macroeconomic variables related to stock price forecasting including opening price, low price, high price, and closing price. Also, to produce better results, it is recommended that the k-nearest neighbor needs to find nearest neighbors using the second step filtering method considering fundamental economic variables as well as a sufficient amount of learning data.