• Title/Summary/Keyword: ALL

Search Result 127,126, Processing Time 0.142 seconds

The Framework of Research Network and Performance Evaluation on Personal Information Security: Social Network Analysis Perspective (개인정보보호 분야의 연구자 네트워크와 성과 평가 프레임워크: 소셜 네트워크 분석을 중심으로)

  • Kim, Minsu;Choi, Jaewon;Kim, Hyun Jin
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.1
    • /
    • pp.177-193
    • /
    • 2014
  • Over the past decade, there has been a rapid diffusion of electronic commerce and a rising number of interconnected networks, resulting in an escalation of security threats and privacy concerns. Electronic commerce has a built-in trade-off between the necessity of providing at least some personal information to consummate an online transaction, and the risk of negative consequences from providing such information. More recently, the frequent disclosure of private information has raised concerns about privacy and its impacts. This has motivated researchers in various fields to explore information privacy issues to address these concerns. Accordingly, the necessity for information privacy policies and technologies for collecting and storing data, and information privacy research in various fields such as medicine, computer science, business, and statistics has increased. The occurrence of various information security accidents have made finding experts in the information security field an important issue. Objective measures for finding such experts are required, as it is currently rather subjective. Based on social network analysis, this paper focused on a framework to evaluate the process of finding experts in the information security field. We collected data from the National Discovery for Science Leaders (NDSL) database, initially collecting about 2000 papers covering the period between 2005 and 2013. Outliers and the data of irrelevant papers were dropped, leaving 784 papers to test the suggested hypotheses. The co-authorship network data for co-author relationship, publisher, affiliation, and so on were analyzed using social network measures including centrality and structural hole. The results of our model estimation are as follows. With the exception of Hypothesis 3, which deals with the relationship between eigenvector centrality and performance, all of our hypotheses were supported. In line with our hypothesis, degree centrality (H1) was supported with its positive influence on the researchers' publishing performance (p<0.001). This finding indicates that as the degree of cooperation increased, the more the publishing performance of researchers increased. In addition, closeness centrality (H2) was also positively associated with researchers' publishing performance (p<0.001), suggesting that, as the efficiency of information acquisition increased, the more the researchers' publishing performance increased. This paper identified the difference in publishing performance among researchers. The analysis can be used to identify core experts and evaluate their performance in the information privacy research field. The co-authorship network for information privacy can aid in understanding the deep relationships among researchers. In addition, extracting characteristics of publishers and affiliations, this paper suggested an understanding of the social network measures and their potential for finding experts in the information privacy field. Social concerns about securing the objectivity of experts have increased, because experts in the information privacy field frequently participate in political consultation, and business education support and evaluation. In terms of practical implications, this research suggests an objective framework for experts in the information privacy field, and is useful for people who are in charge of managing research human resources. This study has some limitations, providing opportunities and suggestions for future research. Presenting the difference in information diffusion according to media and proximity presents difficulties for the generalization of the theory due to the small sample size. Therefore, further studies could consider an increased sample size and media diversity, the difference in information diffusion according to the media type, and information proximity could be explored in more detail. Moreover, previous network research has commonly observed a causal relationship between the independent and dependent variable (Kadushin, 2012). In this study, degree centrality as an independent variable might have causal relationship with performance as a dependent variable. However, in the case of network analysis research, network indices could be computed after the network relationship is created. An annual analysis could help mitigate this limitation.

Bankruptcy Forecasting Model using AdaBoost: A Focus on Construction Companies (적응형 부스팅을 이용한 파산 예측 모형: 건설업을 중심으로)

  • Heo, Junyoung;Yang, Jin Yong
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.1
    • /
    • pp.35-48
    • /
    • 2014
  • According to the 2013 construction market outlook report, the liquidation of construction companies is expected to continue due to the ongoing residential construction recession. Bankruptcies of construction companies have a greater social impact compared to other industries. However, due to the different nature of the capital structure and debt-to-equity ratio, it is more difficult to forecast construction companies' bankruptcies than that of companies in other industries. The construction industry operates on greater leverage, with high debt-to-equity ratios, and project cash flow focused on the second half. The economic cycle greatly influences construction companies. Therefore, downturns tend to rapidly increase the bankruptcy rates of construction companies. High leverage, coupled with increased bankruptcy rates, could lead to greater burdens on banks providing loans to construction companies. Nevertheless, the bankruptcy prediction model concentrated mainly on financial institutions, with rare construction-specific studies. The bankruptcy prediction model based on corporate finance data has been studied for some time in various ways. However, the model is intended for all companies in general, and it may not be appropriate for forecasting bankruptcies of construction companies, who typically have high liquidity risks. The construction industry is capital-intensive, operates on long timelines with large-scale investment projects, and has comparatively longer payback periods than in other industries. With its unique capital structure, it can be difficult to apply a model used to judge the financial risk of companies in general to those in the construction industry. Diverse studies of bankruptcy forecasting models based on a company's financial statements have been conducted for many years. The subjects of the model, however, were general firms, and the models may not be proper for accurately forecasting companies with disproportionately large liquidity risks, such as construction companies. The construction industry is capital-intensive, requiring significant investments in long-term projects, therefore to realize returns from the investment. The unique capital structure means that the same criteria used for other industries cannot be applied to effectively evaluate financial risk for construction firms. Altman Z-score was first published in 1968, and is commonly used as a bankruptcy forecasting model. It forecasts the likelihood of a company going bankrupt by using a simple formula, classifying the results into three categories, and evaluating the corporate status as dangerous, moderate, or safe. When a company falls into the "dangerous" category, it has a high likelihood of bankruptcy within two years, while those in the "safe" category have a low likelihood of bankruptcy. For companies in the "moderate" category, it is difficult to forecast the risk. Many of the construction firm cases in this study fell in the "moderate" category, which made it difficult to forecast their risk. Along with the development of machine learning using computers, recent studies of corporate bankruptcy forecasting have used this technology. Pattern recognition, a representative application area in machine learning, is applied to forecasting corporate bankruptcy, with patterns analyzed based on a company's financial information, and then judged as to whether the pattern belongs to the bankruptcy risk group or the safe group. The representative machine learning models previously used in bankruptcy forecasting are Artificial Neural Networks, Adaptive Boosting (AdaBoost) and, the Support Vector Machine (SVM). There are also many hybrid studies combining these models. Existing studies using the traditional Z-Score technique or bankruptcy prediction using machine learning focus on companies in non-specific industries. Therefore, the industry-specific characteristics of companies are not considered. In this paper, we confirm that adaptive boosting (AdaBoost) is the most appropriate forecasting model for construction companies by based on company size. We classified construction companies into three groups - large, medium, and small based on the company's capital. We analyzed the predictive ability of AdaBoost for each group of companies. The experimental results showed that AdaBoost has more predictive ability than the other models, especially for the group of large companies with capital of more than 50 billion won.

Comparison of CT based-CTV plan and CT based-ICRU38 plan in brachytherapy planning of uterine cervix cancer (자궁경부암 강내조사 시 CT를 이용한 CTV에 근거한 치료계획과 ICRU 38에 근거할 치료계획의 비교)

  • Shim JinSup;Jo JungKun;Si ChangKeun;Lee KiHo;Lee DuHyun;Choi KyeSuk
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.16 no.2
    • /
    • pp.9-17
    • /
    • 2004
  • Purpose : Although Improve of CT, MRI Radio-diagnosis and Radiation Therapy Planing, but we still use ICRU38 Planning system(2D film-based) broadly. 3-Dimensional ICR plan(CT image based) is not only offer tumor and normal tissue dose but also support DVH information. On this study, we plan irradiation-goal dose on CTV(CTV plan) and irradiation-goal dose on ICRU 38 point(ICRU38 plan) by use CT image. And compare with tumor-dose, rectal-dose, bladder-dose on both planning, and analysis DVH Method and Material : Sample 11 patients who treated by Ir-192 HDR. After 40Gy external radiation therapy, ICR plan established. All the patients carry out CT-image scanned by CT-simulator. And we use PLATO(Nucletron) v.14.2 planing system. We draw CTV, rectum, bladder on the CT image. And establish plan irradiation-$100\%$ dose on CTV(CTV plan) and irradiation-$100\%$ dose on A-point(ICRU38 plan) Result : CTV volume($average{\pm}SD$) is $21.8{\pm}26.6cm^3$, rectum volume($average{\pm}SD$) is $60.9{\pm}25.0cm^3$, bladder volume($average{\pm}SD$) is $116.1{\pm}40.1cm^3$ sampled 11 patients. The volume including $100\%$ dose is $126.7{\pm}18.9cm^3$ on ICRU plan and $98.2{\pm}74.5cm^3$ on CTV plan. On ICRU planning, the other one's $22.0cm^3$ CTV volume who residual tumor size excess 4cm is not including $100\%$ isodose. 8 patient's $12.9{\pm}5.9cm^3$ tumor volume who residual tumor size belows 4cm irradiated $100\%$ dose. Bladder dose(recommended by ICRU 38) is $90.1{\pm}21.3\%$ on ICRU plan, $68.7{\pm}26.6\%$ on CTV plan, and rectal dose is $86.4{\pm}18.3\%,\;76.9{\pm}15.6\%$. Bladder and Rectum maximum dose is $137.2{\pm}50.1\%,\;101.1{\pm}41.8\%$ on ICRU plan, $107.6{\pm}47.9\%,\;86.9{\pm}30.8\%$ on CTV plan. Therefore CTV plan more less normal issue-irradiated dose than ICRU plan. But one patient case who residual tumor size excess 4cm, Normal tissue dose more higher than critical dose remarkably on CTV plan. $80\%$over-Irradiated rectal dose(V80rec) is $1.8{\pm}2.4cm^3$ on ICRU plan, $0.7{\pm}1.0cm^3$ on CTV plan. $80\%$over-Irradiated bladder dose(V80bla) is $12.2{\pm}8.9cm^3$ on ICRU plan, $3.5{\pm}4.1cm^3$ on CTV plan. Likewise, CTV plan more less irradiated normal tissue than ICRU38 plan. Conclusion : Although, prove effect and stability about previous ICRU plan, if we use CTV plan by CT image, we will reduce normal tissue dose and irradiated goal-dose at residual tumor on small residual tumor case. But bigger residual tumor case, we need more research about effective 3D-planning.

  • PDF

Bone mineral density and nutritional state according to milk consumption in Korean postmenopausal women who drink coffee: Using the 2008~2009 Korea National Health and Nutrition Examination Survey (한국 폐경 후 여성 커피소비자에서 우유섭취여부에 따른 골밀도와 영양상태 비교 : 2008~2009년 국민건강영양조사 자료 이용)

  • Ryu, Sun-Hyoung;Suh, Yoon Suk
    • Journal of Nutrition and Health
    • /
    • v.49 no.5
    • /
    • pp.347-357
    • /
    • 2016
  • Purpose: This study investigated bone mineral density and nutritional state according to consumption of milk in Korean postmenopausal women who drink coffee. Methods: Using the 2008~2009 Korean National Health & Nutrition Examination Survey data, a total of 1,373 postmenopausal females aged 50 yrs and over were analyzed after excluding those with diseases related to bone health. According to coffee and/or milk consumption, subjects were divided into four groups: coffee only, both coffee & milk, milk only, and none of the above. All data were processed after application of weighted values and adjustment of age, body mass index, physical activity, drinking, and smoking using a general linear model. For analysis of nutrient intake and bone density, data were additionally adjusted by total energy and calcium intake. Results: The coffee & milk group had more subjects younger than 65 yrs and higher education, urban residents, and higher income than any other group. The coffee only group showed somewhat similar characteristics as the none of the above group, which showed the highest percentage of subjects older than 65 and in a lower education and socio-economic state. Body weight, height, body mass index, and lean mass were the highest in coffee & milk group and lowest in the none of the above group. On the other hand, the milk only group showed the lowest values for body mass index and waist circumference, whereas percent body fat did not show any difference among the groups. The coffee and milk group showed the highest bone mineral density in the total femur and lumbar spine as well as the highest nutritional state and most food group intakes, followed by the milk only group, coffee only group, and none of the above group. In the assessment of osteoporosis based on T-score of bone mineral density, although not significant, the coffee and milk group and milk only group, which showed a better nutritional state, included more subjects with a normal bone density, whereas the none of the above group included more subjects with osteoporosis than any other group. Conclusion: Bone mineral density in postmenopausal women might not be affected by coffee drinking if their diets are accompanied by balanced food and nutrient intake including milk.

Batch Scale Storage of Sprouting Foods by Irradiation Combined with Natural Low Temperature - III. Storage of Onions - (방사선조사(放射線照射)와 자연저온(自然低溫)에 의한 발아식품(發芽食品)의 Batch Scale 저장(貯藏)에 관한 연구(硏究) - 제3보(第三報) 양파의 저장(貯藏) -)

  • Cho, Han-Ok;Kwon, Joong-Ho;Byun, Myung-Woo;Yang, Ho-Sook
    • Applied Biological Chemistry
    • /
    • v.26 no.2
    • /
    • pp.82-89
    • /
    • 1983
  • In order to develop a commercial storage method of onions by irradiation combined with natural low temperature, two local varieties of onions, precocious species and late ripening, were stored at natural low temperature storage room ($450{\times}650{\times}250cmH.$; year-round temperature change, $2{\sim}17^{\circ}C$; R.H., $80{\sim}85%$) on batch scale following irradiation with optimum dose level. Precocious and late varieties were all sprouted after five to seven months storage, whereas $10{\sim}15$ Krad irradiated precocious variety was $2{\sim}4%$ sprouted after nine months storage, but sprouting was completly inhibited at the same dose for late variety. The extent of loss due to rot attack after ten months storage were $23{\sim}49%$ in both control and irradiated group of precocious variety but those of late variety were only $4{\sim}10%$. The weight loss of irradiated precocious variety after ten months storage was $13{\sim}16$, while that of late variety was $5.3{\sim}5.9%$ after nine months storage. The moisture content, during whole storage period, of two varieties were $90{\sim}93$ with negligible changes. The total sugar content differed little with varieties and doses immediatly after irradiation, but decreased by the elapse of storage period. 33.6% of its content was decreased in control and 12.5% in irradiated group but $20{\sim}26$ decreased in both control and irradiated group of late variety after nine months storage. No appreciable change was observed immediately after irradiation irrespective of variety and dose, but decreased slightly with storage. Ascorbic acid content of precocious variety was increased slightly with dose immediately after irradiation, but those of late variety decreased slightly. Ascorbic acid content were generally decreased during whole storage period. An economical preservation method of onions appliable to late variety, would be to irradiate onion bulbs at dost range of $10{\sim}15$ Krad followed by storage at natural low temperature storage room.

  • PDF

Calculation of Unit Hydrograph from Discharge Curve, Determination of Sluice Dimension and Tidal Computation for Determination of the Closure curve (단위유량도와 비수갑문 단면 및 방조제 축조곡선 결정을 위한 조속계산)

  • 최귀열
    • Magazine of the Korean Society of Agricultural Engineers
    • /
    • v.7 no.1
    • /
    • pp.861-876
    • /
    • 1965
  • During my stay in the Netherlands, I have studied the following, primarily in relation to the Mokpo Yong-san project which had been studied by the NEDECO for a feasibility report. 1. Unit hydrograph at Naju There are many ways to make unit hydrograph, but I want explain here to make unit hydrograph from the- actual run of curve at Naju. A discharge curve made from one rain storm depends on rainfall intensity per houre After finriing hydrograph every two hours, we will get two-hour unit hydrograph to devide each ordinate of the two-hour hydrograph by the rainfall intensity. I have used one storm from June 24 to June 26, 1963, recording a rainfall intensity of average 9. 4 mm per hour for 12 hours. If several rain gage stations had already been established in the catchment area. above Naju prior to this storm, I could have gathered accurate data on rainfall intensity throughout the catchment area. As it was, I used I the automatic rain gage record of the Mokpo I moteorological station to determine the rainfall lntensity. In order. to develop the unit ~Ydrograph at Naju, I subtracted the basic flow from the total runoff flow. I also tried to keed the difference between the calculated discharge amount and the measured discharge less than 1O~ The discharge period. of an unit graph depends on the length of the catchment area. 2. Determination of sluice dimension Acoording to principles of design presently used in our country, a one-day storm with a frequency of 20 years must be discharged in 8 hours. These design criteria are not adequate, and several dams have washed out in the past years. The design of the spillway and sluice dimensions must be based on the maximun peak discharge flowing into the reservoir to avoid crop and structure damages. The total flow into the reservoir is the summation of flow described by the Mokpo hydrograph, the basic flow from all the catchment areas and the rainfall on the reservoir area. To calculate the amount of water discharged through the sluiceCper half hour), the average head during that interval must be known. This can be calculated from the known water level outside the sluiceCdetermined by the tide) and from an estimated water level inside the reservoir at the end of each time interval. The total amount of water discharged through the sluice can be calculated from this average head, the time interval and the cross-sectional area of' the sluice. From the inflow into the .reservoir and the outflow through the sluice gates I calculated the change in the volume of water stored in the reservoir at half-hour intervals. From the stored volume of water and the known storage capacity of the reservoir, I was able to calculate the water level in the reservoir. The Calculated water level in the reservoir must be the same as the estimated water level. Mean stand tide will be adequate to use for determining the sluice dimension because spring tide is worse case and neap tide is best condition for the I result of the calculatio 3. Tidal computation for determination of the closure curve. During the construction of a dam, whether by building up of a succession of horizontael layers or by building in from both sides, the velocity of the water flowinii through the closing gapwill increase, because of the gradual decrease in the cross sectional area of the gap. 1 calculated the . velocities in the closing gap during flood and ebb for the first mentioned method of construction until the cross-sectional area has been reduced to about 25% of the original area, the change in tidal movement within the reservoir being negligible. Up to that point, the increase of the velocity is more or less hyperbolic. During the closing of the last 25 % of the gap, less water can flow out of the reservoir. This causes a rise of the mean water level of the reservoir. The difference in hydraulic head is then no longer negligible and must be taken into account. When, during the course of construction. the submerged weir become a free weir the critical flow occurs. The critical flow is that point, during either ebb or flood, at which the velocity reaches a maximum. When the dam is raised further. the velocity decreases because of the decrease\ulcorner in the height of the water above the weir. The calculation of the currents and velocities for a stage in the closure of the final gap is done in the following manner; Using an average tide with a neglible daily quantity, I estimated the water level on the pustream side of. the dam (inner water level). I determined the current through the gap for each hour by multiplying the storage area by the increment of the rise in water level. The velocity at a given moment can be determined from the calcalated current in m3/sec, and the cross-sectional area at that moment. At the same time from the difference between inner water level and tidal level (outer water level) the velocity can be calculated with the formula $h= \frac{V^2}{2g}$ and must be equal to the velocity detertnined from the current. If there is a difference in velocity, a new estimate of the inner water level must be made and entire procedure should be repeated. When the higher water level is equal to or more than 2/3 times the difference between the lower water level and the crest of the dam, we speak of a "free weir." The flow over the weir is then dependent upon the higher water level and not on the difference between high and low water levels. When the weir is "submerged", that is, the higher water level is less than 2/3 times the difference between the lower water and the crest of the dam, the difference between the high and low levels being decisive. The free weir normally occurs first during ebb, and is due to. the fact that mean level in the estuary is higher than the mean level of . the tide in building dams with barges the maximum velocity in the closing gap may not be more than 3m/sec. As the maximum velocities are higher than this limit we must use other construction methods in closing the gap. This can be done by dump-cars from each side or by using a cable way.e or by using a cable way.

  • PDF

Chinese relationship between animation and best pole - Focused on the aesthetic principles of the Cultural Revolution period (중국 애니메이션과 모범극의 상관관계 연구 - 문화대혁명 시기의 미학 원칙을 중심으로)

  • Kong, De Wei
    • Cartoon and Animation Studies
    • /
    • s.39
    • /
    • pp.215-231
    • /
    • 2015
  • The Cultural Revolution in the history of Chinese animation hinder the development of the initial animation, and after a negative assessment instrument provided the cause is to become sluggish growth of the Chinese animation. So this time animation are things that are the subject of academic research studies or analysis has been depreciating almost uniformly without evaluation. However, of all the cultural and artistic creation it is developing in its own specific historical conditions and has the aesthetic results. This paper puts the primary purpose is to hold in consideration the aesthetic principles that led to cultural and artistic creativity and objective perspective the achievements the Chinese animation of the time period of the Cultural Revolution. Cultural Revolution is avoided to the previous period in accordance with the socialist ideology of Mao Ze-dong(毛澤東) sikindaneun highlight the culture of the proletariat and placed our goal to create a new class culture. Therefore, cultural and artistic creation of this period is often inconsistent with this part of our aesthetic principles generally accepted character has a non- elitist and anti properties. Best drama is a creative one hand as a model to implement the principles of aesthetics, art and culture Cultural Revolution period kkophimyeo reference for understanding the aesthetic principles that animated the Chinese Cultural Revolution period of orientation. This paper has San Tu Chu(三突出), Hong Guang Liang(紅光亮), and Gao Da Quan(高大全) at the time of the Cultural Revolution aesthetic principles are reflected in how the concrete work, the Cultural Revolution when the animation is how to accommodate these aesthetic principles and placed emphasis on comparative studies on best pole and correlation of the Cultural Revolution when the Chinese animation to ensure that adaptation in own way. First, after analyzing whether the aesthetic principles of focusing on the similarities of the best pole time of the Cultural Revolution and China, and how to implement animation in the works, these aesthetic principles according to the analysis of positive and negative influence on the creation of Chinese animation It was described as neutral. The detailed analysis and comparative study courses were trying to access in two significant aspects of the characters and scenes directing. In terms of character animation of the Cultural Revolution in China when a young boy or girl, emphasis should emphasize the health tinged with red lips and cheek blush to highlight the desired Gong Nong Bing(工農兵) shape as the main character and smooth texture and sophisticated highlights the glittering feeling to the touch, it was confirmed focused hayeoteum to implement the principle of 'Hong Guang Liang', highlighting the brilliant colors with a clean, bright colors. Highlighting a number of protagoniste compared to the antagonist in the animated scene of the Cultural Revolution a few times in terms of production and, among a number of protagoniste also emphasizes the outstanding hero figure, "yet three outstanding heroes heroic figures also emphasize the leading figures among the the director of the extrusion step-by-step approach "('San Tu Chu')was used. In addition, the hero figure is generally high and low angle by directing a large and perfect aesthetic appearance was to faithfully implement the principle of 'high-charged'('Gao Da Quan').

Ensemble Learning with Support Vector Machines for Bond Rating (회사채 신용등급 예측을 위한 SVM 앙상블학습)

  • Kim, Myoung-Jong
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.2
    • /
    • pp.29-45
    • /
    • 2012
  • Bond rating is regarded as an important event for measuring financial risk of companies and for determining the investment returns of investors. As a result, it has been a popular research topic for researchers to predict companies' credit ratings by applying statistical and machine learning techniques. The statistical techniques, including multiple regression, multiple discriminant analysis (MDA), logistic models (LOGIT), and probit analysis, have been traditionally used in bond rating. However, one major drawback is that it should be based on strict assumptions. Such strict assumptions include linearity, normality, independence among predictor variables and pre-existing functional forms relating the criterion variablesand the predictor variables. Those strict assumptions of traditional statistics have limited their application to the real world. Machine learning techniques also used in bond rating prediction models include decision trees (DT), neural networks (NN), and Support Vector Machine (SVM). Especially, SVM is recognized as a new and promising classification and regression analysis method. SVM learns a separating hyperplane that can maximize the margin between two categories. SVM is simple enough to be analyzed mathematical, and leads to high performance in practical applications. SVM implements the structuralrisk minimization principle and searches to minimize an upper bound of the generalization error. In addition, the solution of SVM may be a global optimum and thus, overfitting is unlikely to occur with SVM. In addition, SVM does not require too many data sample for training since it builds prediction models by only using some representative sample near the boundaries called support vectors. A number of experimental researches have indicated that SVM has been successfully applied in a variety of pattern recognition fields. However, there are three major drawbacks that can be potential causes for degrading SVM's performance. First, SVM is originally proposed for solving binary-class classification problems. Methods for combining SVMs for multi-class classification such as One-Against-One, One-Against-All have been proposed, but they do not improve the performance in multi-class classification problem as much as SVM for binary-class classification. Second, approximation algorithms (e.g. decomposition methods, sequential minimal optimization algorithm) could be used for effective multi-class computation to reduce computation time, but it could deteriorate classification performance. Third, the difficulty in multi-class prediction problems is in data imbalance problem that can occur when the number of instances in one class greatly outnumbers the number of instances in the other class. Such data sets often cause a default classifier to be built due to skewed boundary and thus the reduction in the classification accuracy of such a classifier. SVM ensemble learning is one of machine learning methods to cope with the above drawbacks. Ensemble learning is a method for improving the performance of classification and prediction algorithms. AdaBoost is one of the widely used ensemble learning techniques. It constructs a composite classifier by sequentially training classifiers while increasing weight on the misclassified observations through iterations. The observations that are incorrectly predicted by previous classifiers are chosen more often than examples that are correctly predicted. Thus Boosting attempts to produce new classifiers that are better able to predict examples for which the current ensemble's performance is poor. In this way, it can reinforce the training of the misclassified observations of the minority class. This paper proposes a multiclass Geometric Mean-based Boosting (MGM-Boost) to resolve multiclass prediction problem. Since MGM-Boost introduces the notion of geometric mean into AdaBoost, it can perform learning process considering the geometric mean-based accuracy and errors of multiclass. This study applies MGM-Boost to the real-world bond rating case for Korean companies to examine the feasibility of MGM-Boost. 10-fold cross validations for threetimes with different random seeds are performed in order to ensure that the comparison among three different classifiers does not happen by chance. For each of 10-fold cross validation, the entire data set is first partitioned into tenequal-sized sets, and then each set is in turn used as the test set while the classifier trains on the other nine sets. That is, cross-validated folds have been tested independently of each algorithm. Through these steps, we have obtained the results for classifiers on each of the 30 experiments. In the comparison of arithmetic mean-based prediction accuracy between individual classifiers, MGM-Boost (52.95%) shows higher prediction accuracy than both AdaBoost (51.69%) and SVM (49.47%). MGM-Boost (28.12%) also shows the higher prediction accuracy than AdaBoost (24.65%) and SVM (15.42%)in terms of geometric mean-based prediction accuracy. T-test is used to examine whether the performance of each classifiers for 30 folds is significantly different. The results indicate that performance of MGM-Boost is significantly different from AdaBoost and SVM classifiers at 1% level. These results mean that MGM-Boost can provide robust and stable solutions to multi-classproblems such as bond rating.

Rough Set Analysis for Stock Market Timing (러프집합분석을 이용한 매매시점 결정)

  • Huh, Jin-Nyung;Kim, Kyoung-Jae;Han, In-Goo
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.3
    • /
    • pp.77-97
    • /
    • 2010
  • Market timing is an investment strategy which is used for obtaining excessive return from financial market. In general, detection of market timing means determining when to buy and sell to get excess return from trading. In many market timing systems, trading rules have been used as an engine to generate signals for trade. On the other hand, some researchers proposed the rough set analysis as a proper tool for market timing because it does not generate a signal for trade when the pattern of the market is uncertain by using the control function. The data for the rough set analysis should be discretized of numeric value because the rough set only accepts categorical data for analysis. Discretization searches for proper "cuts" for numeric data that determine intervals. All values that lie within each interval are transformed into same value. In general, there are four methods for data discretization in rough set analysis including equal frequency scaling, expert's knowledge-based discretization, minimum entropy scaling, and na$\ddot{i}$ve and Boolean reasoning-based discretization. Equal frequency scaling fixes a number of intervals and examines the histogram of each variable, then determines cuts so that approximately the same number of samples fall into each of the intervals. Expert's knowledge-based discretization determines cuts according to knowledge of domain experts through literature review or interview with experts. Minimum entropy scaling implements the algorithm based on recursively partitioning the value set of each variable so that a local measure of entropy is optimized. Na$\ddot{i}$ve and Booleanreasoning-based discretization searches categorical values by using Na$\ddot{i}$ve scaling the data, then finds the optimized dicretization thresholds through Boolean reasoning. Although the rough set analysis is promising for market timing, there is little research on the impact of the various data discretization methods on performance from trading using the rough set analysis. In this study, we compare stock market timing models using rough set analysis with various data discretization methods. The research data used in this study are the KOSPI 200 from May 1996 to October 1998. KOSPI 200 is the underlying index of the KOSPI 200 futures which is the first derivative instrument in the Korean stock market. The KOSPI 200 is a market value weighted index which consists of 200 stocks selected by criteria on liquidity and their status in corresponding industry including manufacturing, construction, communication, electricity and gas, distribution and services, and financing. The total number of samples is 660 trading days. In addition, this study uses popular technical indicators as independent variables. The experimental results show that the most profitable method for the training sample is the na$\ddot{i}$ve and Boolean reasoning but the expert's knowledge-based discretization is the most profitable method for the validation sample. In addition, the expert's knowledge-based discretization produced robust performance for both of training and validation sample. We also compared rough set analysis and decision tree. This study experimented C4.5 for the comparison purpose. The results show that rough set analysis with expert's knowledge-based discretization produced more profitable rules than C4.5.

Analysis on Factors Influencing Welfare Spending of Local Authority : Implementing the Detailed Data Extracted from the Social Security Information System (지방자치단체 자체 복지사업 지출 영향요인 분석 : 사회보장정보시스템을 통한 접근)

  • Kim, Kyoung-June;Ham, Young-Jin;Lee, Ki-Dong
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.2
    • /
    • pp.141-156
    • /
    • 2013
  • Researchers in welfare services of local government in Korea have rather been on isolated issues as disables, childcare, aging phenomenon, etc. (Kang, 2004; Jung et al., 2009). Lately, local officials, yet, realize that they need more comprehensive welfare services for all residents, not just for above-mentioned focused groups. Still cases dealt with focused group approach have been a main research stream due to various reason(Jung et al., 2009; Lee, 2009; Jang, 2011). Social Security Information System is an information system that comprehensively manages 292 welfare benefits provided by 17 ministries and 40 thousand welfare services provided by 230 local authorities in Korea. The purpose of the system is to improve efficiency of social welfare delivery process. The study of local government expenditure has been on the rise over the last few decades after the restarting the local autonomy, but these studies have limitations on data collection. Measurement of a local government's welfare efforts(spending) has been primarily on expenditures or budget for an individual, set aside for welfare. This practice of using monetary value for an individual as a "proxy value" for welfare effort(spending) is based on the assumption that expenditure is directly linked to welfare efforts(Lee et al., 2007). This expenditure/budget approach commonly uses total welfare amount or percentage figure as dependent variables (Wildavsky, 1985; Lee et al., 2007; Kang, 2000). However, current practice of using actual amount being used or percentage figure as a dependent variable may have some limitation; since budget or expenditure is greatly influenced by the total budget of a local government, relying on such monetary value may create inflate or deflate the true "welfare effort" (Jang, 2012). In addition, government budget usually contain a large amount of administrative cost, i.e., salary, for local officials, which is highly unrelated to the actual welfare expenditure (Jang, 2011). This paper used local government welfare service data from the detailed data sets linked to the Social Security Information System. The purpose of this paper is to analyze the factors that affect social welfare spending of 230 local authorities in 2012. The paper applied multiple regression based model to analyze the pooled financial data from the system. Based on the regression analysis, the following factors affecting self-funded welfare spending were identified. In our research model, we use the welfare budget/total budget(%) of a local government as a true measurement for a local government's welfare effort(spending). Doing so, we exclude central government subsidies or support being used for local welfare service. It is because central government welfare support does not truly reflect the welfare efforts(spending) of a local. The dependent variable of this paper is the volume of the welfare spending and the independent variables of the model are comprised of three categories, in terms of socio-demographic perspectives, the local economy and the financial capacity of local government. This paper categorized local authorities into 3 groups, districts, and cities and suburb areas. The model used a dummy variable as the control variable (local political factor). This paper demonstrated that the volume of the welfare spending for the welfare services is commonly influenced by the ratio of welfare budget to total local budget, the population of infants, self-reliance ratio and the level of unemployment factor. Interestingly, the influential factors are different by the size of local government. Analysis of determinants of local government self-welfare spending, we found a significant effect of local Gov. Finance characteristic in degree of the local government's financial independence, financial independence rate, rate of social welfare budget, and regional economic in opening-to-application ratio, and sociology of population in rate of infants. The result means that local authorities should have differentiated welfare strategies according to their conditions and circumstances. There is a meaning that this paper has successfully proven the significant factors influencing welfare spending of local government in Korea.