• Title/Summary/Keyword: Forecasting error

Search Result 539, Processing Time 0.024 seconds

Impacts of OSTIA Sea Surface Temperature in Regional Ocean Data Assimilation System (지역 해양순환예측시스템에 대한 OSTIA 해수면온도 자료동화 효과에 관한 연구)

  • Kim, Ji Hye;Eom, Hyun-Min;Choi, Jong-Kuk;Lee, Sang-Min;Kim, Young-Ho;Chang, Pil-Hun
    • The Sea:JOURNAL OF THE KOREAN SOCIETY OF OCEANOGRAPHY
    • /
    • v.20 no.1
    • /
    • pp.1-15
    • /
    • 2015
  • Impacts of Sea Surface Temperature (SST) assimilation to the prediction of upper ocean temperature is investigated by using a regional ocean forecasting system, in which 3-dimensional optimal interpolation is applied. In the present study, Sea Surface Temperature and Sea Ice Analysis (OSTIA) dataset is adopted for the daily SST assimilation. This study mainly compares two experimental results with (Exp. DA) and without data assimilation (Exp. NoDA). When comparing both results with OSTIA SST data during Sept. 2011, Exp. NoDA shows Root Mean Square Error (RMSE) of about $1.5^{\circ}C$ at 24, 48, 72 forecast hour. On the other hand, Exp. DA yields the relatively lower RMSE of below $0.8^{\circ}C$ at all forecast hour. In particular, RMSE from Exp. DA reaches $0.57^{\circ}C$ at 24 forecast hour, indicating that the assimilation of daily SST (i.e., OSTIA) improves the performance in the early SST prediction. Furthermore, reduction ratio of RMSE in the Exp. DA reaches over 60% in the Yellow and East seas. In order to examine impacts in the shallow costal region, the SST measured by eight moored buoys around Korean peninsula is compared with both experiments. Exp. DA reveals reduction ratio of RMSE over 70% in all season except for summer, showing the contribution of OSTIA assimilation to the short-range prediction in the coastal region. In addition, the effect of SST assimilation in the upper ocean temperature is examined by the comparison with Argo data in the East Sea. The comparison shows that RMSE from Exp. DA is reduced by $1.5^{\circ}C$ up to 100 m depth in winter where vertical mixing is strong. Thus, SST assimilation is found to be efficient also in the upper ocean prediction. However, the temperature below the mixed layer in winter reveals larger difference in Exp. DA, implying that SST assimilation has still a limitation to the prediction of ocean interior.

The Effect of Data Size on the k-NN Predictability: Application to Samsung Electronics Stock Market Prediction (데이터 크기에 따른 k-NN의 예측력 연구: 삼성전자주가를 사례로)

  • Chun, Se-Hak
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.239-251
    • /
    • 2019
  • Statistical methods such as moving averages, Kalman filtering, exponential smoothing, regression analysis, and ARIMA (autoregressive integrated moving average) have been used for stock market predictions. However, these statistical methods have not produced superior performances. In recent years, machine learning techniques have been widely used in stock market predictions, including artificial neural network, SVM, and genetic algorithm. In particular, a case-based reasoning method, known as k-nearest neighbor is also widely used for stock price prediction. Case based reasoning retrieves several similar cases from previous cases when a new problem occurs, and combines the class labels of similar cases to create a classification for the new problem. However, case based reasoning has some problems. First, case based reasoning has a tendency to search for a fixed number of neighbors in the observation space and always selects the same number of neighbors rather than the best similar neighbors for the target case. So, case based reasoning may have to take into account more cases even when there are fewer cases applicable depending on the subject. Second, case based reasoning may select neighbors that are far away from the target case. Thus, case based reasoning does not guarantee an optimal pseudo-neighborhood for various target cases, and the predictability can be degraded due to a deviation from the desired similar neighbor. This paper examines how the size of learning data affects stock price predictability through k-nearest neighbor and compares the predictability of k-nearest neighbor with the random walk model according to the size of the learning data and the number of neighbors. In this study, Samsung electronics stock prices were predicted by dividing the learning dataset into two types. For the prediction of next day's closing price, we used four variables: opening value, daily high, daily low, and daily close. In the first experiment, data from January 1, 2000 to December 31, 2017 were used for the learning process. In the second experiment, data from January 1, 2015 to December 31, 2017 were used for the learning process. The test data is from January 1, 2018 to August 31, 2018 for both experiments. We compared the performance of k-NN with the random walk model using the two learning dataset. The mean absolute percentage error (MAPE) was 1.3497 for the random walk model and 1.3570 for the k-NN for the first experiment when the learning data was small. However, the mean absolute percentage error (MAPE) for the random walk model was 1.3497 and the k-NN was 1.2928 for the second experiment when the learning data was large. These results show that the prediction power when more learning data are used is higher than when less learning data are used. Also, this paper shows that k-NN generally produces a better predictive power than random walk model for larger learning datasets and does not when the learning dataset is relatively small. Future studies need to consider macroeconomic variables related to stock price forecasting including opening price, low price, high price, and closing price. Also, to produce better results, it is recommended that the k-nearest neighbor needs to find nearest neighbors using the second step filtering method considering fundamental economic variables as well as a sufficient amount of learning data.

Analysis of the Effect of Objective Functions on Hydrologic Model Calibration and Simulation (목적함수에 따른 매개변수 추정 및 수문모형 정확도 비교·분석)

  • Lee, Gi Ha;Yeon, Min Ho;Kim, Young Hun;Jung, Sung Ho
    • Journal of Korean Society of Disaster and Security
    • /
    • v.15 no.1
    • /
    • pp.1-12
    • /
    • 2022
  • An automatic optimization technique is used to estimate the optimal parameters of the hydrologic model, and different hydrologic response results can be provided depending on objective functions. In this study, the parameters of the event-based rainfall-runoff model were estimated using various objective functions, the reproducibility of the hydrograph according to the objective functions was evaluated, and appropriate objective functions were proposed. As the rainfall-runoff model, the storage function model(SFM), which is a lumped hydrologic model used for runoff simulation in the current Korean flood forecasting system, was selected. In order to evaluate the reproducibility of the hydrograph for each objective function, 9 rainfall events were selected for the Cheoncheon basin, which is the upstream basin of Yongdam Dam, and widely-used 7 objective functions were selected for parameter estimation of the SFM for each rainfall event. Then, the reproducibility of the simulated hydrograph using the optimal parameter sets based on the different objective functions was analyzed. As a result, RMSE, NSE, and RSR, which include the error square term in the objective function, showed the highest accuracy for all rainfall events except for Event 7. In addition, in the case of PBIAS and VE, which include an error term compared to the observed flow, it also showed relatively stable reproducibility of the hydrograph. However, in the case of MIA, which adjusts parameters sensitive to high flow and low flow simultaneously, the hydrograph reproducibility performance was found to be very low.

High-Resolution Numerical Simulations with WRF/Noah-MP in Cheongmicheon Farmland in Korea During the 2014 Special Observation Period (2014년 특별관측 기간 동안 청미천 농경지에서의 WRF/Noah-MP 고해상도 수치모의)

  • Song, Jiae;Lee, Seung-Jae;Kang, Minseok;Moon, Minkyu;Lee, Jung-Hoon;Kim, Joon
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.17 no.4
    • /
    • pp.384-398
    • /
    • 2015
  • In this paper, the high-resolution Weather Research and Forecasting/Noah-MultiParameterization (WRF/Noah-MP) modeling system is configured for the Cheongmicheon Farmland site in Korea (CFK), and its performance in land and atmospheric simulation is evaluated using the observed data at CFK during the 2014 special observation period (21 August-10 September). In order to explore the usefulness of turning on Noah-MP dynamic vegetation in midterm simulations of surface and atmospheric variables, two numerical experiments are conducted without dynamic vegetation and with dynamic vegetation (referred to as CTL and DVG experiments, respectively). The main results are as following. 1) CTL showed a tendency of overestimating daytime net shortwave radiation, thereby surface heat fluxes and Bowen ratio. The CTL experiment showed reasonable magnitudes and timing of air temperature at 2 m and 10 m; especially the small error in simulating minimum air temperature showed high potential for predicting frost and leaf wetness duration. The CTL experiment overestimated 10-m wind and precipitation, but the beginning and ending time of precipitation were well captured. 2) When the dynamic vegetation was turned on, the WRF/Noah-MP system showed more realistic values of leaf area index (LAI), net shortwave radiation, surface heat fluxes, Bowen ratio, air temperature, wind and precipitation. The DVG experiment, where LAI is a prognostic variable, produced larger LAI than CTL, and the larger LAI showed better agreement with the observed. The simulated Bowen ratio got closer to the observed ratio, indicating reasonable surface energy partition. The DVG experiment showed patterns similar to CTL, with differences for maximum air temperature. Both experiments showed faster rising of 10-m air temperature during the morning growth hours, presumably due to the rapid growth of daytime mixed layers in the Yonsei University (YSU) boundary layer scheme. The DVG experiment decreased errors in simulating 10-m wind and precipitation. 3) As horizontal resolution increases, the models did not show practical improvement in simulation performance for surface fluxes, air temperature, wind and precipitation, and required three-dimensional observation for more agricultural land spots as well as consistency in model topography and land cover data.

The Economic Growth of Korea Since 1990 : Contributing Factors from Demand and Supply Sides (1990년대 이후 한국경제의 성장: 수요 및 공급 측 요인의 문제)

  • Hur, Seok-Kyun
    • KDI Journal of Economic Policy
    • /
    • v.31 no.1
    • /
    • pp.169-206
    • /
    • 2009
  • This study stems from a question, "How should we understand the pattern of the Korean economy after the 1990s?" Among various analytic methods applicable, this study chooses a Structural Vector Autoregression (SVAR) with long-run restrictions, identifies diverse impacts that gave rise to the current status of the Korean economy, and differentiates relative contributions of those impacts. To that end, SVAR is applied to four economic models; Blanchard and Quah (1989)'s 2-variable model, its 3-variable extensions, and the two other New Keynesian type linear models modified from Stock and Watson (2002). Especially, the latter two models are devised to reflect the recent transitions in the determination of foreign exchange rate (from a fixed rate regime to a flexible rate one) as well as the monetary policy rule (from aggregate targeting to inflation targeting). When organizing the assumed results in the form of impulse response and forecasting error variance decomposition, two common denominators are found as follows. First, changes in the rate of economic growth are mainly attributable to the impact on productivity, and such trend has grown strong since the 2000s, which indicates that Korea's economic growth since the 2000s has been closely associated with its potential growth rate. Second, the magnitude or consistency of impact responses tends to have subsided since the 2000s. Given Korea's high dependence on trade, it is possible that low interest rates, low inflation, steady growth, and the economic emergence of China as a world player have helped secure capital and demand for export and import, which therefore might reduced the impact of each sector on overall economic status. Despite the fact that a diverse mixture of models and impacts has been used for analysis, always two common findings are observed in the result. Therefore, it can be concluded that the decreased rate of economic growth of Korea since 2000 appears to be on the same track as the decrease in Korea's potential growth rate. The contents of this paper are constructed as follows: The second section observes the recent trend of the economic development of Korea and related Korean articles, which might help in clearly defining the scope and analytic methodology of this study. The third section provides an analysis model to be used in this study, which is Structural VAR as mentioned above. Variables used, estimation equations, and identification conditions of impacts are explained. The fourth section reports estimation results derived by the previously introduced model, and the fifth section concludes.

  • PDF

The NCAM Land-Atmosphere Modeling Package (LAMP) Version 1: Implementation and Evaluation (국가농림기상센터 지면대기모델링패키지(NCAM-LAMP) 버전 1: 구축 및 평가)

  • Lee, Seung-Jae;Song, Jiae;Kim, Yu-Jung
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.18 no.4
    • /
    • pp.307-319
    • /
    • 2016
  • A Land-Atmosphere Modeling Package (LAMP) for supporting agricultural and forest management was developed at the National Center for AgroMeteorology (NCAM). The package is comprised of two components; one is the Weather Research and Forecasting modeling system (WRF) coupled with Noah-Multiparameterization options (Noah-MP) Land Surface Model (LSM) and the other is an offline one-dimensional LSM. The objective of this paper is to briefly describe the two components of the NCAM-LAMP and to evaluate their initial performance. The coupled WRF/Noah-MP system is configured with a parent domain over East Asia and three nested domains with a finest horizontal grid size of 810 m. The innermost domain covers two Gwangneung deciduous and coniferous KoFlux sites (GDK and GCK). The model is integrated for about 8 days with the initial and boundary conditions taken from the National Centers for Environmental Prediction (NCEP) Final Analysis (FNL) data. The verification variables are 2-m air temperature, 10-m wind, 2-m humidity, and surface precipitation for the WRF/Noah-MP coupled system. Skill scores are calculated for each domain and two dynamic vegetation options using the difference between the observed data from the Korea Meteorological Administration (KMA) and the simulated data from the WRF/Noah-MP coupled system. The accuracy of precipitation simulation is examined using a contingency table that is made up of the Probability of Detection (POD) and the Equitable Threat Score (ETS). The standalone LSM simulation is conducted for one year with the original settings and is compared with the KoFlux site observation for net radiation, sensible heat flux, latent heat flux, and soil moisture variables. According to results, the innermost domain (810 m resolution) among all domains showed the minimum root mean square error for 2-m air temperature, 10-m wind, and 2-m humidity. Turning on the dynamic vegetation had a tendency of reducing 10-m wind simulation errors in all domains. The first nested domain (7,290 m resolution) showed the highest precipitation score, but showed little advantage compared with using the dynamic vegetation. On the other hand, the offline one-dimensional Noah-MP LSM simulation captured the site observed pattern and magnitude of radiative fluxes and soil moisture, and it left room for further improvement through supplementing the model input of leaf area index and finding a proper combination of model physics.

A Case Study on Forecasting Inbound Calls of Motor Insurance Company Using Interactive Data Mining Technique (대화식 데이터 마이닝 기법을 활용한 자동차 보험사의 인입 콜량 예측 사례)

  • Baek, Woong;Kim, Nam-Gyu
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.3
    • /
    • pp.99-120
    • /
    • 2010
  • Due to the wide spread of customers' frequent access of non face-to-face services, there have been many attempts to improve customer satisfaction using huge amounts of data accumulated throughnon face-to-face channels. Usually, a call center is regarded to be one of the most representative non-faced channels. Therefore, it is important that a call center has enough agents to offer high level customer satisfaction. However, managing too many agents would increase the operational costs of a call center by increasing labor costs. Therefore, predicting and calculating the appropriate size of human resources of a call center is one of the most critical success factors of call center management. For this reason, most call centers are currently establishing a department of WFM(Work Force Management) to estimate the appropriate number of agents and to direct much effort to predict the volume of inbound calls. In real world applications, inbound call prediction is usually performed based on the intuition and experience of a domain expert. In other words, a domain expert usually predicts the volume of calls by calculating the average call of some periods and adjusting the average according tohis/her subjective estimation. However, this kind of approach has radical limitations in that the result of prediction might be strongly affected by the expert's personal experience and competence. It is often the case that a domain expert may predict inbound calls quite differently from anotherif the two experts have mutually different opinions on selecting influential variables and priorities among the variables. Moreover, it is almost impossible to logically clarify the process of expert's subjective prediction. Currently, to overcome the limitations of subjective call prediction, most call centers are adopting a WFMS(Workforce Management System) package in which expert's best practices are systemized. With WFMS, a user can predict the volume of calls by calculating the average call of each day of the week, excluding some eventful days. However, WFMS costs too much capital during the early stage of system establishment. Moreover, it is hard to reflect new information ontothe system when some factors affecting the amount of calls have been changed. In this paper, we attempt to devise a new model for predicting inbound calls that is not only based on theoretical background but also easily applicable to real world applications. Our model was mainly developed by the interactive decision tree technique, one of the most popular techniques in data mining. Therefore, we expect that our model can predict inbound calls automatically based on historical data, and it can utilize expert's domain knowledge during the process of tree construction. To analyze the accuracy of our model, we performed intensive experiments on a real case of one of the largest car insurance companies in Korea. In the case study, the prediction accuracy of the devised two models and traditional WFMS are analyzed with respect to the various error rates allowable. The experiments reveal that our data mining-based two models outperform WFMS in terms of predicting the amount of accident calls and fault calls in most experimental situations examined.

Global Ocean Data Assimilation and Prediction System in KMA: Description and Assessment (기상청 전지구 해양자료동화시스템(GODAPS): 개요 및 검증)

  • Chang, Pil-Hun;Hwang, Seung-On;Choo, Sung-Ho;Lee, Johan;Lee, Sang-Min;Boo, Kyung-On
    • Atmosphere
    • /
    • v.31 no.2
    • /
    • pp.229-240
    • /
    • 2021
  • The Global Ocean Data Assimilation and Prediction System (GODAPS) in operation at the KMA (Korea Meteorological Administration) is introduced. GODAPS consists of ocean model, ice model, and 3-d variational ocean data assimilation system. GODAPS assimilates conventional and satellite observations for sea surface temperature and height, observations of sea-ice concentration, as well as temperature and salinity profiles for the ocean using a 24-hour data assimilation window. It finally produces ocean analysis fields with a resolution of 0.25 ORCA (tripolar) grid and 75-layer in depth. This analysis is used for providing a boundary condition for the atmospheric model of the KMA Global Seasonal Forecasting System version 5 (GloSea5) in addition to monitoring on the global ocean and ice. For the purpose of evaluating the quality of ocean analysis produced by GODAPS, a one-year data assimilation experiment was performed. Assimilation of global observing system in GODAPS results in producing improved analysis and forecast fields with reduced error in terms of RMSE of innovation and analysis increment. In addition, comparison with an unassimilated experiment shows a mostly positive impact, especially over the region with large oceanic variability.

A Study on Developing a VKOSPI Forecasting Model via GARCH Class Models for Intelligent Volatility Trading Systems (지능형 변동성트레이딩시스템개발을 위한 GARCH 모형을 통한 VKOSPI 예측모형 개발에 관한 연구)

  • Kim, Sun-Woong
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.2
    • /
    • pp.19-32
    • /
    • 2010
  • Volatility plays a central role in both academic and practical applications, especially in pricing financial derivative products and trading volatility strategies. This study presents a novel mechanism based on generalized autoregressive conditional heteroskedasticity (GARCH) models that is able to enhance the performance of intelligent volatility trading systems by predicting Korean stock market volatility more accurately. In particular, we embedded the concept of the volatility asymmetry documented widely in the literature into our model. The newly developed Korean stock market volatility index of KOSPI 200, VKOSPI, is used as a volatility proxy. It is the price of a linear portfolio of the KOSPI 200 index options and measures the effect of the expectations of dealers and option traders on stock market volatility for 30 calendar days. The KOSPI 200 index options market started in 1997 and has become the most actively traded market in the world. Its trading volume is more than 10 million contracts a day and records the highest of all the stock index option markets. Therefore, analyzing the VKOSPI has great importance in understanding volatility inherent in option prices and can afford some trading ideas for futures and option dealers. Use of the VKOSPI as volatility proxy avoids statistical estimation problems associated with other measures of volatility since the VKOSPI is model-free expected volatility of market participants calculated directly from the transacted option prices. This study estimates the symmetric and asymmetric GARCH models for the KOSPI 200 index from January 2003 to December 2006 by the maximum likelihood procedure. Asymmetric GARCH models include GJR-GARCH model of Glosten, Jagannathan and Runke, exponential GARCH model of Nelson and power autoregressive conditional heteroskedasticity (ARCH) of Ding, Granger and Engle. Symmetric GARCH model indicates basic GARCH (1, 1). Tomorrow's forecasted value and change direction of stock market volatility are obtained by recursive GARCH specifications from January 2007 to December 2009 and are compared with the VKOSPI. Empirical results indicate that negative unanticipated returns increase volatility more than positive return shocks of equal magnitude decrease volatility, indicating the existence of volatility asymmetry in the Korean stock market. The point value and change direction of tomorrow VKOSPI are estimated and forecasted by GARCH models. Volatility trading system is developed using the forecasted change direction of the VKOSPI, that is, if tomorrow VKOSPI is expected to rise, a long straddle or strangle position is established. A short straddle or strangle position is taken if VKOSPI is expected to fall tomorrow. Total profit is calculated as the cumulative sum of the VKOSPI percentage change. If forecasted direction is correct, the absolute value of the VKOSPI percentage changes is added to trading profit. It is subtracted from the trading profit if forecasted direction is not correct. For the in-sample period, the power ARCH model best fits in a statistical metric, Mean Squared Prediction Error (MSPE), and the exponential GARCH model shows the highest Mean Correct Prediction (MCP). The power ARCH model best fits also for the out-of-sample period and provides the highest probability for the VKOSPI change direction tomorrow. Generally, the power ARCH model shows the best fit for the VKOSPI. All the GARCH models provide trading profits for volatility trading system and the exponential GARCH model shows the best performance, annual profit of 197.56%, during the in-sample period. The GARCH models present trading profits during the out-of-sample period except for the exponential GARCH model. During the out-of-sample period, the power ARCH model shows the largest annual trading profit of 38%. The volatility clustering and asymmetry found in this research are the reflection of volatility non-linearity. This further suggests that combining the asymmetric GARCH models and artificial neural networks can significantly enhance the performance of the suggested volatility trading system, since artificial neural networks have been shown to effectively model nonlinear relationships.