• Title/Summary/Keyword: Logit

Search Result 937, Processing Time 0.023 seconds

An exploratory study on the characteristics of technology innovation persistence of Korean firms (한국 기업의 기술혁신 지속 특성에 대한 탐색적 연구)

  • Song, Changhyeon;Lee, Jungwoo;Jang, Pilseong
    • Journal of Technology Innovation
    • /
    • v.29 no.3
    • /
    • pp.1-31
    • /
    • 2021
  • With the growing importance of technology innovation as a key factor for firms' competitive advantage, 'innovation persistence' became also an important research subject. 'Innovation Persistence' is a concept that indicates whether or not firms' innovation activity or performance continues. However, the data used for innovation studies are carried out as cross-sectional surveys in most countries. For this reason, studies dealing with longitudinal aspect of innovation persistence are rare. In particular, there is almost no research on innovation persistence using Korean innovation survey data. This study reviews the concepts and characteristics of innovation persistence based on extant literature, and perform an empirical analysis on the status and features of Korean firms' technology innovation persistence. Based on the data of the Korean Innovation Survey (KIS) conducted every other year from 2012 to 2018, panel data on 3,379 firms which observed multiple times are constructed. As a result, only part of the firms with persistent innovation were observed (for innovation performance 10~12%, for innovation activity 15~17%), and it was found that the persistence of non-innovation was remarkable(about 52~57%). And it was confirmed that the persistence of innovation activities is stronger than that of innovation performance. Besides, some features by sub-types of innovation appeared. Product innovation showed higher persistence than process innovation, and internal R&D also showed higher persistence than joint/external R&D. As a result of additional logit analysis to identify factors, it was found that radical or gradual product innovation is the most influential factor in persisting innovation in the next period. Since the sample selection bias due to a limitations of raw data might exist in the panel data constructed in this study, it should be noted that faulty generalization of the results are not allowed. Nevertheless, this is the first study to examine the technology innovation persistence targeting Korean firms and is expected to be a starting point for follow-up studies. It is anticipated that advanced research results will be drawn through the establishment of official panel data and improved methodologies.

The Effect of Objective and Subjective Social Isolation and Interpersonal Conflict Type on the Probability of Cognitive Impairment by Age Group in Old Age (노년기 연령집단별 객관적·주관적 사회적 고립과 대인관계갈등 유형이 인지기능에 미치는 영향)

  • Lee, Sang Chul
    • 한국노년학
    • /
    • v.38 no.4
    • /
    • pp.811-835
    • /
    • 2018
  • Social relations and cognitive function in old age are closely related to each other, and social relation is classified into structural characteristics and qualitative characteristics reflecting cognitive and emotional evaluation. The concept of social isolation is the focus of attention in relation to the social relations of old age. Social isolation has a multidimensional theoretical structure that is divided into objective dimension such as social network, type of furniture, social participation, and subjective dimension such as lack of perceived social support and loneliness. There is also a close relationship between cognitive function and interpersonal conflict in old age. In this study, we examined the effect of subjective social isolation, which shows the structural characteristics of social relations, and subjective social isolation and interpersonal conflict on the dementia occurrence by age group in the elderly. The data were analyzed by applying a random effect panel logit model using 1,740 panel data from the first year to the third year of KSHAP. The results of the analysis are summarized as follows. First, the cognitive impairment increased sharply with age. Objective and subjective social isolation were both U-shaped distribution with an inflection point of 80 years old. Second, the main effect on the probability of cognitive impairment was statistically significant with objective and subjective social isolation, but the type of interpersonal conflict did not appear to be significant. Third, the results of two-way interaction effect analysis on the probability of cognitive impairment are as follows. The relationship between subjective social isolation and the probability of occurrence of cognitive impairment was significantly different according to the level of conflict with spouse. In addition, the higher the subjective social isolation, the higher the probability of cognitive impairment in the elderly(over 85) than in the young-old(65~74). In addition, as the level of conflict with spouses increases, the probability of cognitive impairment of the oldest-old(aged 85 or older) is drastically lower than that of the young-old(aged 65~74). Based on the results of this study, policy and practical implications for reducing the cognitive impairment of the elderly age group were suggested, and limitations of the study and suggestions for future research were discussed.

A Recidivism Prediction Model Based on XGBoost Considering Asymmetric Error Costs (비대칭 오류 비용을 고려한 XGBoost 기반 재범 예측 모델)

  • Won, Ha-Ram;Shim, Jae-Seung;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.127-137
    • /
    • 2019
  • Recidivism prediction has been a subject of constant research by experts since the early 1970s. But it has become more important as committed crimes by recidivist steadily increase. Especially, in the 1990s, after the US and Canada adopted the 'Recidivism Risk Assessment Report' as a decisive criterion during trial and parole screening, research on recidivism prediction became more active. And in the same period, empirical studies on 'Recidivism Factors' were started even at Korea. Even though most recidivism prediction studies have so far focused on factors of recidivism or the accuracy of recidivism prediction, it is important to minimize the prediction misclassification cost, because recidivism prediction has an asymmetric error cost structure. In general, the cost of misrecognizing people who do not cause recidivism to cause recidivism is lower than the cost of incorrectly classifying people who would cause recidivism. Because the former increases only the additional monitoring costs, while the latter increases the amount of social, and economic costs. Therefore, in this paper, we propose an XGBoost(eXtream Gradient Boosting; XGB) based recidivism prediction model considering asymmetric error cost. In the first step of the model, XGB, being recognized as high performance ensemble method in the field of data mining, was applied. And the results of XGB were compared with various prediction models such as LOGIT(logistic regression analysis), DT(decision trees), ANN(artificial neural networks), and SVM(support vector machines). In the next step, the threshold is optimized to minimize the total misclassification cost, which is the weighted average of FNE(False Negative Error) and FPE(False Positive Error). To verify the usefulness of the model, the model was applied to a real recidivism prediction dataset. As a result, it was confirmed that the XGB model not only showed better prediction accuracy than other prediction models but also reduced the cost of misclassification most effectively.

A Study on Estimation of Environmental Value of Tentatively Named 'East-West Trail' Using CVM (CVM기법을 이용한 가칭 '동서트레일'의 환경가치 추정)

  • Kee-Rae Kang;Yoon-Ho Choi;Bo-Kwang Chung;Dong-Pil Kim;Hyun-Kyeong Oh;Woo-Sung Lee;Su-Bok Chae
    • Korean Journal of Environment and Ecology
    • /
    • v.36 no.6
    • /
    • pp.676-683
    • /
    • 2022
  • Due to the effects of rapid changes in the living environment since 2000 and the recent unforeseen pandemic, people are refraining from domestic and international traveling and movement, and outdoor activities for health and the public value of forest trails, called Dullegil Trail in Korea, have become more important. This study estimated the environmental value of the tentatively named "East-West Trail," which connects the forest trails crossing Chungcheong and Gyeongsang Provinces using CVM (Contingent Valuation Method). It surveyed visitors to the East-West Trail, and 725 questionnaires were used for analysis. The average characteristics of respondents were those who exercised 2-3 times per week, visited a forest trail not far from their residence with friends or family, and showed a tendency to spend 50 thousand Korean won or more per visit. Visitors to the Dullegil Trail felt that there was a shortage of information boards on the forest trail, and they preferred a shelter in appropriate locations. We used a double-bounded dichotomous choice (BDDC) logit model proposed by Hanemann to measure the conservation value of the East-West Trail. It was estimated that the environmental value that a visitor to the East-West Trail could obtain was 30,087 won per trip. The estimated environmental value of the East-West Trail can be converted to about 94 billion won total visitors annually based on the population belonging to the direct-use zone near the East-West Trail. As there has been no study on the environmental value of forest trails using CVM, the results of this study will be able to suggest the feasibility of the government policies on forest trails.

A Time Series Graph based Convolutional Neural Network Model for Effective Input Variable Pattern Learning : Application to the Prediction of Stock Market (효과적인 입력변수 패턴 학습을 위한 시계열 그래프 기반 합성곱 신경망 모형: 주식시장 예측에의 응용)

  • Lee, Mo-Se;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.167-181
    • /
    • 2018
  • Over the past decade, deep learning has been in spotlight among various machine learning algorithms. In particular, CNN(Convolutional Neural Network), which is known as the effective solution for recognizing and classifying images or voices, has been popularly applied to classification and prediction problems. In this study, we investigate the way to apply CNN in business problem solving. Specifically, this study propose to apply CNN to stock market prediction, one of the most challenging tasks in the machine learning research. As mentioned, CNN has strength in interpreting images. Thus, the model proposed in this study adopts CNN as the binary classifier that predicts stock market direction (upward or downward) by using time series graphs as its inputs. That is, our proposal is to build a machine learning algorithm that mimics an experts called 'technical analysts' who examine the graph of past price movement, and predict future financial price movements. Our proposed model named 'CNN-FG(Convolutional Neural Network using Fluctuation Graph)' consists of five steps. In the first step, it divides the dataset into the intervals of 5 days. And then, it creates time series graphs for the divided dataset in step 2. The size of the image in which the graph is drawn is $40(pixels){\times}40(pixels)$, and the graph of each independent variable was drawn using different colors. In step 3, the model converts the images into the matrices. Each image is converted into the combination of three matrices in order to express the value of the color using R(red), G(green), and B(blue) scale. In the next step, it splits the dataset of the graph images into training and validation datasets. We used 80% of the total dataset as the training dataset, and the remaining 20% as the validation dataset. And then, CNN classifiers are trained using the images of training dataset in the final step. Regarding the parameters of CNN-FG, we adopted two convolution filters ($5{\times}5{\times}6$ and $5{\times}5{\times}9$) in the convolution layer. In the pooling layer, $2{\times}2$ max pooling filter was used. The numbers of the nodes in two hidden layers were set to, respectively, 900 and 32, and the number of the nodes in the output layer was set to 2(one is for the prediction of upward trend, and the other one is for downward trend). Activation functions for the convolution layer and the hidden layer were set to ReLU(Rectified Linear Unit), and one for the output layer set to Softmax function. To validate our model - CNN-FG, we applied it to the prediction of KOSPI200 for 2,026 days in eight years (from 2009 to 2016). To match the proportions of the two groups in the independent variable (i.e. tomorrow's stock market movement), we selected 1,950 samples by applying random sampling. Finally, we built the training dataset using 80% of the total dataset (1,560 samples), and the validation dataset using 20% (390 samples). The dependent variables of the experimental dataset included twelve technical indicators popularly been used in the previous studies. They include Stochastic %K, Stochastic %D, Momentum, ROC(rate of change), LW %R(Larry William's %R), A/D oscillator(accumulation/distribution oscillator), OSCP(price oscillator), CCI(commodity channel index), and so on. To confirm the superiority of CNN-FG, we compared its prediction accuracy with the ones of other classification models. Experimental results showed that CNN-FG outperforms LOGIT(logistic regression), ANN(artificial neural network), and SVM(support vector machine) with the statistical significance. These empirical results imply that converting time series business data into graphs and building CNN-based classification models using these graphs can be effective from the perspective of prediction accuracy. Thus, this paper sheds a light on how to apply deep learning techniques to the domain of business problem solving.

Impact of Shortly Acquired IPO Firms on ICT Industry Concentration (ICT 산업분야 신생기업의 IPO 이후 인수합병과 산업 집중도에 관한 연구)

  • Chang, YoungBong;Kwon, YoungOk
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.3
    • /
    • pp.51-69
    • /
    • 2020
  • Now, it is a stylized fact that a small number of technology firms such as Apple, Alphabet, Microsoft, Amazon, Facebook and a few others have become larger and dominant players in an industry. Coupled with the rise of these leading firms, we have also observed that a large number of young firms have become an acquisition target in their early IPO stages. This indeed results in a sharp decline in the number of new entries in public exchanges although a series of policy reforms have been promulgated to foster competition through an increase in new entries. Given the observed industry trend in recent decades, a number of studies have reported increased concentration in most developed countries. However, it is less understood as to what caused an increase in industry concentration. In this paper, we uncover the mechanisms by which industries have become concentrated over the last decades by tracing the changes in industry concentration associated with a firm's status change in its early IPO stages. To this end, we put emphasis on the case in which firms are acquired shortly after they went public. Especially, with the transition to digital-based economies, it is imperative for incumbent firms to adapt and keep pace with new ICT and related intelligent systems. For instance, after the acquisition of a young firm equipped with AI-based solutions, an incumbent firm may better respond to a change in customer taste and preference by integrating acquired AI solutions and analytics skills into multiple business processes. Accordingly, it is not unusual for young ICT firms become an attractive acquisition target. To examine the role of M&As involved with young firms in reshaping the level of industry concentration, we identify a firm's status in early post-IPO stages over the sample periods spanning from 1990 to 2016 as follows: i) being delisted, ii) being standalone firms and iii) being acquired. According to our analysis, firms that have conducted IPO since 2000s have been acquired by incumbent firms at a relatively quicker time than those that did IPO in previous generations. We also show a greater acquisition rate for IPO firms in the ICT sector compared with their counterparts in other sectors. Our results based on multinomial logit models suggest that a large number of IPO firms have been acquired in their early post-IPO lives despite their financial soundness. Specifically, we show that IPO firms are likely to be acquired rather than be delisted due to financial distress in early IPO stages when they are more profitable, more mature or less leveraged. For those IPO firms with venture capital backup have also become an acquisition target more frequently. As a larger number of firms are acquired shortly after their IPO, our results show increased concentration. While providing limited evidence on the impact of large incumbent firms in explaining the change in industry concentration, our results show that the large firms' effect on industry concentration are pronounced in the ICT sector. This result possibly captures the current trend that a few tech giants such as Alphabet, Apple and Facebook continue to increase their market share. In addition, compared with the acquisitions of non-ICT firms, the concentration impact of IPO firms in early stages becomes larger when ICT firms are acquired as a target. Our study makes new contributions. To our best knowledge, this is one of a few studies that link a firm's post-IPO status to associated changes in industry concentration. Although some studies have addressed concentration issues, their primary focus was on market power or proprietary software. Contrast to earlier studies, we are able to uncover the mechanism by which industries have become concentrated by placing emphasis on M&As involving young IPO firms. Interestingly, the concentration impact of IPO firm acquisitions are magnified when a large incumbent firms are involved as an acquirer. This leads us to infer the underlying reasons as to why industries have become more concentrated with a favor of large firms in recent decades. Overall, our study sheds new light on the literature by providing a plausible explanation as to why industries have become concentrated.

Ensemble Learning with Support Vector Machines for Bond Rating (회사채 신용등급 예측을 위한 SVM 앙상블학습)

  • Kim, Myoung-Jong
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.2
    • /
    • pp.29-45
    • /
    • 2012
  • Bond rating is regarded as an important event for measuring financial risk of companies and for determining the investment returns of investors. As a result, it has been a popular research topic for researchers to predict companies' credit ratings by applying statistical and machine learning techniques. The statistical techniques, including multiple regression, multiple discriminant analysis (MDA), logistic models (LOGIT), and probit analysis, have been traditionally used in bond rating. However, one major drawback is that it should be based on strict assumptions. Such strict assumptions include linearity, normality, independence among predictor variables and pre-existing functional forms relating the criterion variablesand the predictor variables. Those strict assumptions of traditional statistics have limited their application to the real world. Machine learning techniques also used in bond rating prediction models include decision trees (DT), neural networks (NN), and Support Vector Machine (SVM). Especially, SVM is recognized as a new and promising classification and regression analysis method. SVM learns a separating hyperplane that can maximize the margin between two categories. SVM is simple enough to be analyzed mathematical, and leads to high performance in practical applications. SVM implements the structuralrisk minimization principle and searches to minimize an upper bound of the generalization error. In addition, the solution of SVM may be a global optimum and thus, overfitting is unlikely to occur with SVM. In addition, SVM does not require too many data sample for training since it builds prediction models by only using some representative sample near the boundaries called support vectors. A number of experimental researches have indicated that SVM has been successfully applied in a variety of pattern recognition fields. However, there are three major drawbacks that can be potential causes for degrading SVM's performance. First, SVM is originally proposed for solving binary-class classification problems. Methods for combining SVMs for multi-class classification such as One-Against-One, One-Against-All have been proposed, but they do not improve the performance in multi-class classification problem as much as SVM for binary-class classification. Second, approximation algorithms (e.g. decomposition methods, sequential minimal optimization algorithm) could be used for effective multi-class computation to reduce computation time, but it could deteriorate classification performance. Third, the difficulty in multi-class prediction problems is in data imbalance problem that can occur when the number of instances in one class greatly outnumbers the number of instances in the other class. Such data sets often cause a default classifier to be built due to skewed boundary and thus the reduction in the classification accuracy of such a classifier. SVM ensemble learning is one of machine learning methods to cope with the above drawbacks. Ensemble learning is a method for improving the performance of classification and prediction algorithms. AdaBoost is one of the widely used ensemble learning techniques. It constructs a composite classifier by sequentially training classifiers while increasing weight on the misclassified observations through iterations. The observations that are incorrectly predicted by previous classifiers are chosen more often than examples that are correctly predicted. Thus Boosting attempts to produce new classifiers that are better able to predict examples for which the current ensemble's performance is poor. In this way, it can reinforce the training of the misclassified observations of the minority class. This paper proposes a multiclass Geometric Mean-based Boosting (MGM-Boost) to resolve multiclass prediction problem. Since MGM-Boost introduces the notion of geometric mean into AdaBoost, it can perform learning process considering the geometric mean-based accuracy and errors of multiclass. This study applies MGM-Boost to the real-world bond rating case for Korean companies to examine the feasibility of MGM-Boost. 10-fold cross validations for threetimes with different random seeds are performed in order to ensure that the comparison among three different classifiers does not happen by chance. For each of 10-fold cross validation, the entire data set is first partitioned into tenequal-sized sets, and then each set is in turn used as the test set while the classifier trains on the other nine sets. That is, cross-validated folds have been tested independently of each algorithm. Through these steps, we have obtained the results for classifiers on each of the 30 experiments. In the comparison of arithmetic mean-based prediction accuracy between individual classifiers, MGM-Boost (52.95%) shows higher prediction accuracy than both AdaBoost (51.69%) and SVM (49.47%). MGM-Boost (28.12%) also shows the higher prediction accuracy than AdaBoost (24.65%) and SVM (15.42%)in terms of geometric mean-based prediction accuracy. T-test is used to examine whether the performance of each classifiers for 30 folds is significantly different. The results indicate that performance of MGM-Boost is significantly different from AdaBoost and SVM classifiers at 1% level. These results mean that MGM-Boost can provide robust and stable solutions to multi-classproblems such as bond rating.