• Title/Summary/Keyword: Stock Forecasting

Search Result 159, Processing Time 0.025 seconds

An Empirical Study on The Relationship between Stock Index Futures Return and Trading Volume (주가지수 선물 수익률과 거래량간 관계에 관한 실증연구)

  • Hwang Sung Soo;Yoo Young Joong
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.5 no.6
    • /
    • pp.580-587
    • /
    • 2004
  • The purpose of this study is to examine if the trading volume can apply to the short-term forecasting of the futures price change by verificating the casuality between trading volume and futures price in the KOSPI 200 futures market. The outcome of the research is summarized as follows. In the analysis of subordinate periods, based on the yearly time segments, trading volume were found to lead futures price. As for trading volume, it was under comparably greater influence of its self of the past than the return rate of futures. In the analysis of subordinate periods, based on the trend of the futures market, trading volume lead return rate of futures feebly in a bull market. But return rate of futures lead trading volume significantly in a bearish market.

  • PDF

Generating Firm's Performance Indicators by Applying PCA (PCA를 활용한 기업실적 예측변수 생성)

  • Lee, Joonhyuck;Kim, Gabjo;Park, Sangsung;Jang, Dongsik
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.25 no.2
    • /
    • pp.191-196
    • /
    • 2015
  • There have been many studies on statistical forecasting on firm's performance and stock price by applying various financial indicators such as debt ratio and sales growth rate. Selecting predictors for constructing a prediction model among the various financial indicators is very important for precise prediction. Most of the previous studies applied variable selection algorithms for selecting predictors. However, the variable selection algorithm is considered to be at risk of eliminating certain amount of information from the indicators that were excluded from model construction. Therefore, we propose a firm's performance prediction model which principal component analysis is applied instead of the variable selection algorithm, in order to reduce dimensionality of input variables of the prediction model. In this study, we constructed the proposed prediction model by using financial data of American IT companies to empirically analyze prediction performance of the model.

Threshold heterogeneous autoregressive modeling for realized volatility (임계 HAR 모형을 이용한 실현 변동성 분석)

  • Sein Moon;Minsu Park;Changryong Baek
    • The Korean Journal of Applied Statistics
    • /
    • v.36 no.4
    • /
    • pp.295-307
    • /
    • 2023
  • The heterogeneous autoregressive (HAR) model is a simple linear model that is commonly used to explain long memory in the realized volatility. However, as realized volatility has more complicated features such as conditional heteroscedasticity, leverage effect, and volatility clustering, it is necessary to extend the simple HAR model. Therefore, to better incorporate the stylized facts, we propose a threshold HAR model with GARCH errors, namely the THAR-GARCH model. That is, the THAR-GARCH model is a nonlinear model whose coefficients vary according to a threshold value, and the conditional heteroscedasticity is explained through the GARCH errors. Model parameters are estimated using an iterative weighted least squares estimation method. Our simulation study supports the consistency of the iterative estimation method. In addition, we show that the proposed THAR-GARCH model has better forecasting power by applying to the realized volatility of major 21 stock indices around the world.

Robo-Advisor Algorithm with Intelligent View Model (지능형 전망모형을 결합한 로보어드바이저 알고리즘)

  • Kim, Sunwoong
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.39-55
    • /
    • 2019
  • Recently banks and large financial institutions have introduced lots of Robo-Advisor products. Robo-Advisor is a Robot to produce the optimal asset allocation portfolio for investors by using the financial engineering algorithms without any human intervention. Since the first introduction in Wall Street in 2008, the market size has grown to 60 billion dollars and is expected to expand to 2,000 billion dollars by 2020. Since Robo-Advisor algorithms suggest asset allocation output to investors, mathematical or statistical asset allocation strategies are applied. Mean variance optimization model developed by Markowitz is the typical asset allocation model. The model is a simple but quite intuitive portfolio strategy. For example, assets are allocated in order to minimize the risk on the portfolio while maximizing the expected return on the portfolio using optimization techniques. Despite its theoretical background, both academics and practitioners find that the standard mean variance optimization portfolio is very sensitive to the expected returns calculated by past price data. Corner solutions are often found to be allocated only to a few assets. The Black-Litterman Optimization model overcomes these problems by choosing a neutral Capital Asset Pricing Model equilibrium point. Implied equilibrium returns of each asset are derived from equilibrium market portfolio through reverse optimization. The Black-Litterman model uses a Bayesian approach to combine the subjective views on the price forecast of one or more assets with implied equilibrium returns, resulting a new estimates of risk and expected returns. These new estimates can produce optimal portfolio by the well-known Markowitz mean-variance optimization algorithm. If the investor does not have any views on his asset classes, the Black-Litterman optimization model produce the same portfolio as the market portfolio. What if the subjective views are incorrect? A survey on reports of stocks performance recommended by securities analysts show very poor results. Therefore the incorrect views combined with implied equilibrium returns may produce very poor portfolio output to the Black-Litterman model users. This paper suggests an objective investor views model based on Support Vector Machines(SVM), which have showed good performance results in stock price forecasting. SVM is a discriminative classifier defined by a separating hyper plane. The linear, radial basis and polynomial kernel functions are used to learn the hyper planes. Input variables for the SVM are returns, standard deviations, Stochastics %K and price parity degree for each asset class. SVM output returns expected stock price movements and their probabilities, which are used as input variables in the intelligent views model. The stock price movements are categorized by three phases; down, neutral and up. The expected stock returns make P matrix and their probability results are used in Q matrix. Implied equilibrium returns vector is combined with the intelligent views matrix, resulting the Black-Litterman optimal portfolio. For comparisons, Markowitz mean-variance optimization model and risk parity model are used. The value weighted market portfolio and equal weighted market portfolio are used as benchmark indexes. We collect the 8 KOSPI 200 sector indexes from January 2008 to December 2018 including 132 monthly index values. Training period is from 2008 to 2015 and testing period is from 2016 to 2018. Our suggested intelligent view model combined with implied equilibrium returns produced the optimal Black-Litterman portfolio. The out of sample period portfolio showed better performance compared with the well-known Markowitz mean-variance optimization portfolio, risk parity portfolio and market portfolio. The total return from 3 year-period Black-Litterman portfolio records 6.4%, which is the highest value. The maximum draw down is -20.8%, which is also the lowest value. Sharpe Ratio shows the highest value, 0.17. It measures the return to risk ratio. Overall, our suggested view model shows the possibility of replacing subjective analysts's views with objective view model for practitioners to apply the Robo-Advisor asset allocation algorithms in the real trading fields.

Verification and Estimation of the Contributed Concentration of CH4 Emissions Using the WRF-CMAQ Model in Korea (WRF-CMAQ 모델을 이용한 한반도 CH4 배출의 기여농도 추정 및 검증)

  • Moon, Yun-Seob;Lim, Yun-Kyu;Hong, Sungwook;Chang, Eunmi
    • Journal of the Korean earth science society
    • /
    • v.34 no.3
    • /
    • pp.209-223
    • /
    • 2013
  • The purpose of this study was to estimate the contributed concentration of each emission source to $CH_4$ by verifying the simulated concentration of $CH_4$ in the Korean peninsula, and then to compare the $CH_4$ emission used to the $CH_4$ simulation with that of a box model. We simulated the Weather Research Forecasting-Community Multiscale Air Quality (WRF-CMAQ) model to estimate the mean concentration of $CH_4$ during the period of April 1 to 22 August 2010 in the Korean peninsula. The $CH_4$ emissions within the model were adopted by the anthropogenic emission inventory of both the EDGAR of the global emissions and the GHG-CAPSS of the green house gases in Korea, and by the global biogenic emission inventory of the MEGAN. These $CH_4$ emission data were validated by comparing the $CH_4$ modeling data with the concentration data measured at two different location, Ulnungdo and Anmyeondo in Korea. The contributed concentration of $CH_4$ estimated from the domestic emission sources in verification of the $CH_4$ modeling at Ulnungdo was represented in about 20%, which originated from $CH_4$ sources such as stock farm products (8%), energy contribution and industrial processes (6%), wastes (5%), and biogenesis and landuse (1%) in the Korean peninsula. In addition, one that transported from China was about 9%, and the background concentration of $CH_4$ was shown in about 70%. Furthermore, the $CH_4$ emission estimated from a box model was similar to that of the WRF-CMAQ model.

The Economic Growth of Korea Since 1990 : Contributing Factors from Demand and Supply Sides (1990년대 이후 한국경제의 성장: 수요 및 공급 측 요인의 문제)

  • Hur, Seok-Kyun
    • KDI Journal of Economic Policy
    • /
    • v.31 no.1
    • /
    • pp.169-206
    • /
    • 2009
  • This study stems from a question, "How should we understand the pattern of the Korean economy after the 1990s?" Among various analytic methods applicable, this study chooses a Structural Vector Autoregression (SVAR) with long-run restrictions, identifies diverse impacts that gave rise to the current status of the Korean economy, and differentiates relative contributions of those impacts. To that end, SVAR is applied to four economic models; Blanchard and Quah (1989)'s 2-variable model, its 3-variable extensions, and the two other New Keynesian type linear models modified from Stock and Watson (2002). Especially, the latter two models are devised to reflect the recent transitions in the determination of foreign exchange rate (from a fixed rate regime to a flexible rate one) as well as the monetary policy rule (from aggregate targeting to inflation targeting). When organizing the assumed results in the form of impulse response and forecasting error variance decomposition, two common denominators are found as follows. First, changes in the rate of economic growth are mainly attributable to the impact on productivity, and such trend has grown strong since the 2000s, which indicates that Korea's economic growth since the 2000s has been closely associated with its potential growth rate. Second, the magnitude or consistency of impact responses tends to have subsided since the 2000s. Given Korea's high dependence on trade, it is possible that low interest rates, low inflation, steady growth, and the economic emergence of China as a world player have helped secure capital and demand for export and import, which therefore might reduced the impact of each sector on overall economic status. Despite the fact that a diverse mixture of models and impacts has been used for analysis, always two common findings are observed in the result. Therefore, it can be concluded that the decreased rate of economic growth of Korea since 2000 appears to be on the same track as the decrease in Korea's potential growth rate. The contents of this paper are constructed as follows: The second section observes the recent trend of the economic development of Korea and related Korean articles, which might help in clearly defining the scope and analytic methodology of this study. The third section provides an analysis model to be used in this study, which is Structural VAR as mentioned above. Variables used, estimation equations, and identification conditions of impacts are explained. The fourth section reports estimation results derived by the previously introduced model, and the fifth section concludes.

  • PDF

Unbilled Revenue and Analysts' Earnings Forecasts (진행기준 수익인식 방법과 재무분석가 이익예측 - 미청구공사 계정을 중심으로 -)

  • Lee, Bo-Mi;Park, Bo-Young
    • Management & Information Systems Review
    • /
    • v.36 no.3
    • /
    • pp.151-165
    • /
    • 2017
  • This study investigates the effect of revenue recognition by percentage of completion method on financial analysts' earnings forecasting information in order industry. Specifically, we examines how the analysts' earnings forecast errors and biases differ according to whether or not to report the unbilled revenue account balance and the level of unbilled revenue account balance. The sample consists of 453 firm-years listed in Korea Stock Exchange during the period from 2010 to 2014 since the information on unbilled revenue accounts can be obtained after the adoption of K-IFRS. The results are as follows. First, we find that the firms with unbilled revenue account balances have lower analysts' earnings forecast accuracy than the firms who do not report unbilled revue account balances. In addition, we find that the accuracy of analysts' earnings forecasts decreases as the amount of unbilled revenue increases. Unbilled revenue account balances occur when the revenue recognition of the contractor is faster than the client. There is a possibility that managerial discretionary judgment and estimation may intervene when the contractor calculates the progress rate. The difference between the actual progress of the construction and the progress recognized by the company lowers the predictive value of financial statements. Our results suggest that the analysts' earnings forecasts may be more difficult for the firms that report unbilled revenue balances as applying the revenue recognition method based on the progress criteria. Second, we find that the firms reporting unbilled revenue account balances tend to have higher the optimistic biases in analysts' earnings forecast than the firms who do not report unbilled revenue account balances. And we find that the analysts' earnings forecast biases are increases as the amount of unbilled revenue increases. This study suggests an effort to reduce the arbitrary adjustment and estimation in the measurement of the progress as well as the introduction of the progress measurement method which can reflect the actual progress. Investors are encouraged to invest and analyze the characteristics of the order-based industry accounting standards. In addition, the results of this study empower the accounting transparency enhancement plan for order industry proposed by the policy authorities.

  • PDF

Dynamic forecasts of bankruptcy with Recurrent Neural Network model (RNN(Recurrent Neural Network)을 이용한 기업부도예측모형에서 회계정보의 동적 변화 연구)

  • Kwon, Hyukkun;Lee, Dongkyu;Shin, Minsoo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.3
    • /
    • pp.139-153
    • /
    • 2017
  • Corporate bankruptcy can cause great losses not only to stakeholders but also to many related sectors in society. Through the economic crises, bankruptcy have increased and bankruptcy prediction models have become more and more important. Therefore, corporate bankruptcy has been regarded as one of the major topics of research in business management. Also, many studies in the industry are in progress and important. Previous studies attempted to utilize various methodologies to improve the bankruptcy prediction accuracy and to resolve the overfitting problem, such as Multivariate Discriminant Analysis (MDA), Generalized Linear Model (GLM). These methods are based on statistics. Recently, researchers have used machine learning methodologies such as Support Vector Machine (SVM), Artificial Neural Network (ANN). Furthermore, fuzzy theory and genetic algorithms were used. Because of this change, many of bankruptcy models are developed. Also, performance has been improved. In general, the company's financial and accounting information will change over time. Likewise, the market situation also changes, so there are many difficulties in predicting bankruptcy only with information at a certain point in time. However, even though traditional research has problems that don't take into account the time effect, dynamic model has not been studied much. When we ignore the time effect, we get the biased results. So the static model may not be suitable for predicting bankruptcy. Thus, using the dynamic model, there is a possibility that bankruptcy prediction model is improved. In this paper, we propose RNN (Recurrent Neural Network) which is one of the deep learning methodologies. The RNN learns time series data and the performance is known to be good. Prior to experiment, we selected non-financial firms listed on the KOSPI, KOSDAQ and KONEX markets from 2010 to 2016 for the estimation of the bankruptcy prediction model and the comparison of forecasting performance. In order to prevent a mistake of predicting bankruptcy by using the financial information already reflected in the deterioration of the financial condition of the company, the financial information was collected with a lag of two years, and the default period was defined from January to December of the year. Then we defined the bankruptcy. The bankruptcy we defined is the abolition of the listing due to sluggish earnings. We confirmed abolition of the list at KIND that is corporate stock information website. Then we selected variables at previous papers. The first set of variables are Z-score variables. These variables have become traditional variables in predicting bankruptcy. The second set of variables are dynamic variable set. Finally we selected 240 normal companies and 226 bankrupt companies at the first variable set. Likewise, we selected 229 normal companies and 226 bankrupt companies at the second variable set. We created a model that reflects dynamic changes in time-series financial data and by comparing the suggested model with the analysis of existing bankruptcy predictive models, we found that the suggested model could help to improve the accuracy of bankruptcy predictions. We used financial data in KIS Value (Financial database) and selected Multivariate Discriminant Analysis (MDA), Generalized Linear Model called logistic regression (GLM), Support Vector Machine (SVM), Artificial Neural Network (ANN) model as benchmark. The result of the experiment proved that RNN's performance was better than comparative model. The accuracy of RNN was high in both sets of variables and the Area Under the Curve (AUC) value was also high. Also when we saw the hit-ratio table, the ratio of RNNs that predicted a poor company to be bankrupt was higher than that of other comparative models. However the limitation of this paper is that an overfitting problem occurs during RNN learning. But we expect to be able to solve the overfitting problem by selecting more learning data and appropriate variables. From these result, it is expected that this research will contribute to the development of a bankruptcy prediction by proposing a new dynamic model.

Visualizing the Results of Opinion Mining from Social Media Contents: Case Study of a Noodle Company (소셜미디어 콘텐츠의 오피니언 마이닝결과 시각화: N라면 사례 분석 연구)

  • Kim, Yoosin;Kwon, Do Young;Jeong, Seung Ryul
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.4
    • /
    • pp.89-105
    • /
    • 2014
  • After emergence of Internet, social media with highly interactive Web 2.0 applications has provided very user friendly means for consumers and companies to communicate with each other. Users have routinely published contents involving their opinions and interests in social media such as blogs, forums, chatting rooms, and discussion boards, and the contents are released real-time in the Internet. For that reason, many researchers and marketers regard social media contents as the source of information for business analytics to develop business insights, and many studies have reported results on mining business intelligence from Social media content. In particular, opinion mining and sentiment analysis, as a technique to extract, classify, understand, and assess the opinions implicit in text contents, are frequently applied into social media content analysis because it emphasizes determining sentiment polarity and extracting authors' opinions. A number of frameworks, methods, techniques and tools have been presented by these researchers. However, we have found some weaknesses from their methods which are often technically complicated and are not sufficiently user-friendly for helping business decisions and planning. In this study, we attempted to formulate a more comprehensive and practical approach to conduct opinion mining with visual deliverables. First, we described the entire cycle of practical opinion mining using Social media content from the initial data gathering stage to the final presentation session. Our proposed approach to opinion mining consists of four phases: collecting, qualifying, analyzing, and visualizing. In the first phase, analysts have to choose target social media. Each target media requires different ways for analysts to gain access. There are open-API, searching tools, DB2DB interface, purchasing contents, and so son. Second phase is pre-processing to generate useful materials for meaningful analysis. If we do not remove garbage data, results of social media analysis will not provide meaningful and useful business insights. To clean social media data, natural language processing techniques should be applied. The next step is the opinion mining phase where the cleansed social media content set is to be analyzed. The qualified data set includes not only user-generated contents but also content identification information such as creation date, author name, user id, content id, hit counts, review or reply, favorite, etc. Depending on the purpose of the analysis, researchers or data analysts can select a suitable mining tool. Topic extraction and buzz analysis are usually related to market trends analysis, while sentiment analysis is utilized to conduct reputation analysis. There are also various applications, such as stock prediction, product recommendation, sales forecasting, and so on. The last phase is visualization and presentation of analysis results. The major focus and purpose of this phase are to explain results of analysis and help users to comprehend its meaning. Therefore, to the extent possible, deliverables from this phase should be made simple, clear and easy to understand, rather than complex and flashy. To illustrate our approach, we conducted a case study on a leading Korean instant noodle company. We targeted the leading company, NS Food, with 66.5% of market share; the firm has kept No. 1 position in the Korean "Ramen" business for several decades. We collected a total of 11,869 pieces of contents including blogs, forum contents and news articles. After collecting social media content data, we generated instant noodle business specific language resources for data manipulation and analysis using natural language processing. In addition, we tried to classify contents in more detail categories such as marketing features, environment, reputation, etc. In those phase, we used free ware software programs such as TM, KoNLP, ggplot2 and plyr packages in R project. As the result, we presented several useful visualization outputs like domain specific lexicons, volume and sentiment graphs, topic word cloud, heat maps, valence tree map, and other visualized images to provide vivid, full-colored examples using open library software packages of the R project. Business actors can quickly detect areas by a swift glance that are weak, strong, positive, negative, quiet or loud. Heat map is able to explain movement of sentiment or volume in categories and time matrix which shows density of color on time periods. Valence tree map, one of the most comprehensive and holistic visualization models, should be very helpful for analysts and decision makers to quickly understand the "big picture" business situation with a hierarchical structure since tree-map can present buzz volume and sentiment with a visualized result in a certain period. This case study offers real-world business insights from market sensing which would demonstrate to practical-minded business users how they can use these types of results for timely decision making in response to on-going changes in the market. We believe our approach can provide practical and reliable guide to opinion mining with visualized results that are immediately useful, not just in food industry but in other industries as well.