• Title/Summary/Keyword: Systems Performance

Search Result 25,976, Processing Time 0.051 seconds

Hierarchical Overlapping Clustering to Detect Complex Concepts (중복을 허용한 계층적 클러스터링에 의한 복합 개념 탐지 방법)

  • Hong, Su-Jeong;Choi, Joong-Min
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.1
    • /
    • pp.111-125
    • /
    • 2011
  • Clustering is a process of grouping similar or relevant documents into a cluster and assigning a meaningful concept to the cluster. By this process, clustering facilitates fast and correct search for the relevant documents by narrowing down the range of searching only to the collection of documents belonging to related clusters. For effective clustering, techniques are required for identifying similar documents and grouping them into a cluster, and discovering a concept that is most relevant to the cluster. One of the problems often appearing in this context is the detection of a complex concept that overlaps with several simple concepts at the same hierarchical level. Previous clustering methods were unable to identify and represent a complex concept that belongs to several different clusters at the same level in the concept hierarchy, and also could not validate the semantic hierarchical relationship between a complex concept and each of simple concepts. In order to solve these problems, this paper proposes a new clustering method that identifies and represents complex concepts efficiently. We developed the Hierarchical Overlapping Clustering (HOC) algorithm that modified the traditional Agglomerative Hierarchical Clustering algorithm to allow overlapped clusters at the same level in the concept hierarchy. The HOC algorithm represents the clustering result not by a tree but by a lattice to detect complex concepts. We developed a system that employs the HOC algorithm to carry out the goal of complex concept detection. This system operates in three phases; 1) the preprocessing of documents, 2) the clustering using the HOC algorithm, and 3) the validation of semantic hierarchical relationships among the concepts in the lattice obtained as a result of clustering. The preprocessing phase represents the documents as x-y coordinate values in a 2-dimensional space by considering the weights of terms appearing in the documents. First, it goes through some refinement process by applying stopwords removal and stemming to extract index terms. Then, each index term is assigned a TF-IDF weight value and the x-y coordinate value for each document is determined by combining the TF-IDF values of the terms in it. The clustering phase uses the HOC algorithm in which the similarity between the documents is calculated by applying the Euclidean distance method. Initially, a cluster is generated for each document by grouping those documents that are closest to it. Then, the distance between any two clusters is measured, grouping the closest clusters as a new cluster. This process is repeated until the root cluster is generated. In the validation phase, the feature selection method is applied to validate the appropriateness of the cluster concepts built by the HOC algorithm to see if they have meaningful hierarchical relationships. Feature selection is a method of extracting key features from a document by identifying and assigning weight values to important and representative terms in the document. In order to correctly select key features, a method is needed to determine how each term contributes to the class of the document. Among several methods achieving this goal, this paper adopted the $x^2$�� statistics, which measures the dependency degree of a term t to a class c, and represents the relationship between t and c by a numerical value. To demonstrate the effectiveness of the HOC algorithm, a series of performance evaluation is carried out by using a well-known Reuter-21578 news collection. The result of performance evaluation showed that the HOC algorithm greatly contributes to detecting and producing complex concepts by generating the concept hierarchy in a lattice structure.

The Efficiency Analysis of CRM System in the Hotel Industry Using DEA (DEA를 이용한 호텔 관광 서비스 업계의 CRM 도입 효율성 분석)

  • Kim, Tai-Young;Seol, Kyung-Jin;Kwak, Young-Dai
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.1
    • /
    • pp.91-110
    • /
    • 2011
  • This paper analyzes the cases where the hotels have increased their services and enhanced their work process through IT solutions to cope with computerization globalization. Also the cases have been studies where national hotels use the CRM solution internally to respond effectively to customers requests, increase customer analysis, and build marketing strategies. In particular, this study discusses the introduction of the CRM solutions and CRM sales business and marketing services using a process for utilizing the presumed, CRM by introducing effective DEA(Data Envelopment Analysis). First, the comparison has done regarding the relative efficiency of L Company with the CCR model, then compared L Company's restaurants and facilities' effectiveness through BCC model. L Company reached a conclusion that it is important to precisely create and manage sales data which are the preliminary data for CRM, and for that reason it made it possible to save sales data generated by POS system on each sales performance database. In order to do that, it newly established Oracle POS system and LORIS POS system concerned with restaurants for food and beverage as well as rooms, and made it possible to stably generate and manage sales data and manage. Moreover, it set up a composite database to control comprehensively the results of work processes during a specific period by collecting customer registration information and made it possible to systematically control the information on sales performances. By establishing a system which unifies database and managing it comprehensively, impeccability of data has been greatly enhanced and a problem which generated asymmetric data could be thoroughly solved. Using data accumulated on the comprehensive database, sales data can be analyzed, categorized, classified through data mining engine imbedded in Polaris CRM and the results can be organized on data mart to provide them in the form of CRM application data. By transforming original sales data into forms which are easy to handle and saving them on data mart separately, it enabled acquiring well-organized data with ease when engaging in various marketing operations, holding a morning meeting and working on decision-making. By using summarized data at data mart, it was possible to process marketing operations such as telemarketing, direct mailing, internet marketing service and service product developments for perceived customers; moreover, information on customer perceptions which is one of CRM's end-products could feed back into the comprehensive database. This research was undertaken to find out how effectively CRM has been employed by comparing and analyzing the management performance of each enterprise site and store after introducing CRM to Hotel enterprises using DEA technique. According to the research results, efficiency evaluation for each site was calculated through input and output factors to find out comparative CRM system usage efficiency of L's Company four sites; moreover, with regard to stores, the sizes of workforce and budget application show a huge difference and so does the each store efficiency. Furthermore, by using the DEA technique, it could assess which sites have comparatively high efficiency and which don't by comparing and evaluating hotel enterprises IT project outcomes such as CRM introduction using the CCR model for each site of the related enterprises. By using the BCC model, it could comparatively evaluate the outcome of CRM usage at each store of A site, which is representative of L Company, and as a result, it could figure out which stores maintain high efficiency in using CRM and which don't. It analyzed the cases of CRM introduction at L Company, which is a hotel enterprise, and precisely evaluated them through DEA. L Company analyzed the customer analysis system by introducing CRM and achieved to provide customers identified through client analysis data with one to one tailored services. Moreover, it could come up with a plan to differentiate the service for customers who revisit by assessing customer discernment rate. As tasks to be solved in the future, it is required to do research on the process analysis which can lead to a specific outcome such as increased sales volumes by carrying on test marketing, target marketing using CRM. Furthermore, it is also necessary to do research on efficiency evaluation in accordance with linkages between other IT solutions such as ERP and CRM system.

Bankruptcy prediction using an improved bagging ensemble (개선된 배깅 앙상블을 활용한 기업부도예측)

  • Min, Sung-Hwan
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.4
    • /
    • pp.121-139
    • /
    • 2014
  • Predicting corporate failure has been an important topic in accounting and finance. The costs associated with bankruptcy are high, so the accuracy of bankruptcy prediction is greatly important for financial institutions. Lots of researchers have dealt with the topic associated with bankruptcy prediction in the past three decades. The current research attempts to use ensemble models for improving the performance of bankruptcy prediction. Ensemble classification is to combine individually trained classifiers in order to gain more accurate prediction than individual models. Ensemble techniques are shown to be very useful for improving the generalization ability of the classifier. Bagging is the most commonly used methods for constructing ensemble classifiers. In bagging, the different training data subsets are randomly drawn with replacement from the original training dataset. Base classifiers are trained on the different bootstrap samples. Instance selection is to select critical instances while deleting and removing irrelevant and harmful instances from the original set. Instance selection and bagging are quite well known in data mining. However, few studies have dealt with the integration of instance selection and bagging. This study proposes an improved bagging ensemble based on instance selection using genetic algorithms (GA) for improving the performance of SVM. GA is an efficient optimization procedure based on the theory of natural selection and evolution. GA uses the idea of survival of the fittest by progressively accepting better solutions to the problems. GA searches by maintaining a population of solutions from which better solutions are created rather than making incremental changes to a single solution to the problem. The initial solution population is generated randomly and evolves into the next generation by genetic operators such as selection, crossover and mutation. The solutions coded by strings are evaluated by the fitness function. The proposed model consists of two phases: GA based Instance Selection and Instance based Bagging. In the first phase, GA is used to select optimal instance subset that is used as input data of bagging model. In this study, the chromosome is encoded as a form of binary string for the instance subset. In this phase, the population size was set to 100 while maximum number of generations was set to 150. We set the crossover rate and mutation rate to 0.7 and 0.1 respectively. We used the prediction accuracy of model as the fitness function of GA. SVM model is trained on training data set using the selected instance subset. The prediction accuracy of SVM model over test data set is used as fitness value in order to avoid overfitting. In the second phase, we used the optimal instance subset selected in the first phase as input data of bagging model. We used SVM model as base classifier for bagging ensemble. The majority voting scheme was used as a combining method in this study. This study applies the proposed model to the bankruptcy prediction problem using a real data set from Korean companies. The research data used in this study contains 1832 externally non-audited firms which filed for bankruptcy (916 cases) and non-bankruptcy (916 cases). Financial ratios categorized as stability, profitability, growth, activity and cash flow were investigated through literature review and basic statistical methods and we selected 8 financial ratios as the final input variables. We separated the whole data into three subsets as training, test and validation data set. In this study, we compared the proposed model with several comparative models including the simple individual SVM model, the simple bagging model and the instance selection based SVM model. The McNemar tests were used to examine whether the proposed model significantly outperforms the other models. The experimental results show that the proposed model outperforms the other models.

Using Spatial Data and Crop Growth Modeling to Predict Performance of South Korean Rice Varieties Grown in Western Coastal Plains in North Korea (공간정보와 생육모의에 의한 남한 벼 품종의 북한 서부지대 적응성 예측)

  • 김영호;김희동;한상욱;최재연;구자민;정유란;김재영;윤진일
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.4 no.4
    • /
    • pp.224-236
    • /
    • 2002
  • A long-term growth simulation was performed at 496 land units in the western coastal plains (WCP) of North Korea to test the potential adaptability of each land unit for growing South Korean rice cultivars. The land units for rice cultivation (CZU), each of them represented by a geographically referenced 5 by 5 km grid tell, were identified by analyzing satellite remote sensing data. Surfaces of monthly climatic normals for daily maximum and minimum temperature, precipitation number of rain days and solar radiation were generated at a 1 by 1 km interval by spatial statistical methods using observed data at 51 synoptic weather stations in North and South Korea during 1981-2000. Grid cells felling within a same CZU and, at the same time, corresponding to the satellite data- identified rice growing pixels were extracted and aggregated to make a spatially explicit climatic normals relevant to the rice growing area of the CZU. Daily weather dataset for 30 years was randomly generated from the monthly climatic normals of each CZU. Growth and development parameters of CERES-rice model suitable for 11 major South Korean cultivars were derived from long-term field observations. Eight treatments comprised of 2 transplanting dates $\times$ 2 cropping systems $\times$ 2 irrigation methods were assigned to each cultivar. Each treatment was simulated with the randomly generated 30 years' daily weather data (from planting to physiological maturity) for 496 land units in WCP to simulate the growth and yield responses to the interannual climate variation. The same model was run with the input data from the 3 major crop experiment stations in South Korea to obtain a 30 year normal performance of each cultivar, which was used as a "reference" for comparison. Results were analyzed with respect to spatial and temporal variation in yield and maturity, and used to evaluate the suitability of each land unit for growing a specific South Korean cultivar. The results may be utilized as decision aids for agrotechnology transfer to North Korea, for example, germplasm evaluation, resource allocation and crop calendar preparation.

SysML-Based System Modeling for Design of BIPV Electric Power Generation (건물일체형 태양광 시스템의 전력발전부 설계를 위한 SysML기반 시스템 모델링)

  • Lee, Seung-Joon;Lee, Jae-Chon
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.19 no.10
    • /
    • pp.578-589
    • /
    • 2018
  • Building Integrated Photovoltaic (BIPV) system is a typical integrated system that simultaneously performs both building function and solar power generation function. To maximize its potential advantage, however, the solar photovoltaic power generation function must be integrated from the early conceptual design stage, and maximum power generation must be designed. To cope with such requirements, preliminary research on BIPV design process based on architectural design model and computer simulation results for improving solar power generation performance have been published. However, the requirements of the BIPV system have not been clearly identified and systematically reflected in the subsequent design. Moreover, no model has verified the power generation design. To solve these problems, we systematically model the requirements of BIPV system and study power generation design based on the system requirements model. Through the study, we consistently use the standard system modeling language, SysML. Specifically, stakeholder requirements were first identified from stakeholders and related BIPV standards. Then, based on the domain model, the design requirements of the BIPV system were derived at the system level, and the functional and physical architectures of the target system were created based on the system requirements. Finally, the power generation performance of the BIPV system was evaluated through a simulated SysML model (Parametric diagram). If the SysML system model developed herein can be reinforced by reflecting the conditions resulting from building design, it will open an opportunity to study and optimize the power generation in the BIPV system in an integrated fashion.

Membership Fluidity and Knowledge Collaboration in Virtual Communities: A Multilateral Approach to Membership Fluidity (가상 커뮤니티의 멤버 유동성과 지식 협업: 멤버 유동성에 대한 다각적 접근)

  • Park, Hyun-jung;Shin, Kyung-shik
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.2
    • /
    • pp.19-47
    • /
    • 2015
  • In this era of knowledge economy, a variety of virtual communities are proliferating for the purpose of knowledge creation and utilization. Since the voluntary contributions of members are the essential source of knowledge, member turnover can have significant implications on the survival and success of virtual communities. However, there is a dearth of research on the effect of membership turnover and even the method of measurement for membership turnover is left unclear in virtual communities. In a traditional context, membership turnover is calculated as the ratio of the number of departing members to the average number of members for a given time period. In virtual communities, while the influx of newcomers can be clearly measured, the magnitude of departure is elusive since explicit withdrawals are seldom executed. In addition, there doesn't exist a common way to determine the average number of community members who return and contribute intermittently at will. This study initially examines the limitations in applying the concept of traditional turnover to virtual communities, and proposes five membership fluidity measures based on a preliminary analysis of editing behaviors of 2,978 featured articles in English Wikipedia. Subsequently, this work investigates the relationships between three selected membership fluidity measures and group collaboration performance, reflecting a moderating effect dependent on work characteristic. We obtained the following results: First, membership turnover relates to collaboration efficiency in a right-shortened U-shaped manner, with a moderating effect from work characteristic; given the same turnover rate, the promotion likelihood for a more professional task is lower than that for a less professional task, and the likelihood difference diminishes as the turnover rate increases. Second, contribution period relates to collaboration efficiency in a left-shortened U-shaped manner, with a moderating effect from work characteristic; the marginal performance change per unit change of contribution period is greater for a less professional task. Third, the number of new participants per month relates to collaboration efficiency in a left-shortened reversed U-shaped manner, for which the moderating effect from work characteristic appears to be insignificant.

A Hybrid Forecasting Framework based on Case-based Reasoning and Artificial Neural Network (사례기반 추론기법과 인공신경망을 이용한 서비스 수요예측 프레임워크)

  • Hwang, Yousub
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.4
    • /
    • pp.43-57
    • /
    • 2012
  • To enhance the competitive advantage in a constantly changing business environment, an enterprise management must make the right decision in many business activities based on both internal and external information. Thus, providing accurate information plays a prominent role in management's decision making. Intuitively, historical data can provide a feasible estimate through the forecasting models. Therefore, if the service department can estimate the service quantity for the next period, the service department can then effectively control the inventory of service related resources such as human, parts, and other facilities. In addition, the production department can make load map for improving its product quality. Therefore, obtaining an accurate service forecast most likely appears to be critical to manufacturing companies. Numerous investigations addressing this problem have generally employed statistical methods, such as regression or autoregressive and moving average simulation. However, these methods are only efficient for data with are seasonal or cyclical. If the data are influenced by the special characteristics of product, they are not feasible. In our research, we propose a forecasting framework that predicts service demand of manufacturing organization by combining Case-based reasoning (CBR) and leveraging an unsupervised artificial neural network based clustering analysis (i.e., Self-Organizing Maps; SOM). We believe that this is one of the first attempts at applying unsupervised artificial neural network-based machine-learning techniques in the service forecasting domain. Our proposed approach has several appealing features : (1) We applied CBR and SOM in a new forecasting domain such as service demand forecasting. (2) We proposed our combined approach between CBR and SOM in order to overcome limitations of traditional statistical forecasting methods and We have developed a service forecasting tool based on the proposed approach using an unsupervised artificial neural network and Case-based reasoning. In this research, we conducted an empirical study on a real digital TV manufacturer (i.e., Company A). In addition, we have empirically evaluated the proposed approach and tool using real sales and service related data from digital TV manufacturer. In our empirical experiments, we intend to explore the performance of our proposed service forecasting framework when compared to the performances predicted by other two service forecasting methods; one is traditional CBR based forecasting model and the other is the existing service forecasting model used by Company A. We ran each service forecasting 144 times; each time, input data were randomly sampled for each service forecasting framework. To evaluate accuracy of forecasting results, we used Mean Absolute Percentage Error (MAPE) as primary performance measure in our experiments. We conducted one-way ANOVA test with the 144 measurements of MAPE for three different service forecasting approaches. For example, the F-ratio of MAPE for three different service forecasting approaches is 67.25 and the p-value is 0.000. This means that the difference between the MAPE of the three different service forecasting approaches is significant at the level of 0.000. Since there is a significant difference among the different service forecasting approaches, we conducted Tukey's HSD post hoc test to determine exactly which means of MAPE are significantly different from which other ones. In terms of MAPE, Tukey's HSD post hoc test grouped the three different service forecasting approaches into three different subsets in the following order: our proposed approach > traditional CBR-based service forecasting approach > the existing forecasting approach used by Company A. Consequently, our empirical experiments show that our proposed approach outperformed the traditional CBR based forecasting model and the existing service forecasting model used by Company A. The rest of this paper is organized as follows. Section 2 provides some research background information such as summary of CBR and SOM. Section 3 presents a hybrid service forecasting framework based on Case-based Reasoning and Self-Organizing Maps, while the empirical evaluation results are summarized in Section 4. Conclusion and future research directions are finally discussed in Section 5.

Software Reliability Growth Modeling in the Testing Phase with an Outlier Stage (하나의 이상구간을 가지는 테스팅 단계에서의 소프트웨어 신뢰도 성장 모형화)

  • Park, Man-Gon;Jung, Eun-Yi
    • The Transactions of the Korea Information Processing Society
    • /
    • v.5 no.10
    • /
    • pp.2575-2583
    • /
    • 1998
  • The productionof the highly relible softwae systems and theirs performance evaluation hae become important interests in the software industry. The software evaluation has been mainly carried out in ternns of both reliability and performance of software system. Software reliability is the probability that no software error occurs for a fixed time interval during software testing phase. These theoretical software reliability models are sometimes unsuitable for the practical testing phase in which a software error at a certain testing stage occurs by causes of the imperfect debugging, abnornal software correction, and so on. Such a certatin software testing stage needs to be considered as an outlying stage. And we can assume that the software reliability does not improve by means of muisance factor in this outlying testing stage. In this paper, we discuss Bavesian software reliability growth modeling and estimation procedure in the presence of an imidentitied outlying software testing stage by the modification of Jehnski Moranda. Also we derive the Bayes estimaters of the software reliability panmeters by the assumption of prior information under the squared error los function. In addition, we evaluate the proposed software reliability growth model with an unidentified outlying stage in an exchangeable model according to the values of nuisance paramether using the accuracy, bias, trend, noise metries as the quantilative evaluation criteria through the compater simulation.

  • PDF

An Empirical Study on Successful Factor of Local Mobile App One-Person Creating Company : The Moderating Effects of Social Capital (지역 모바일 앱 1인 창조기업의 성공요인에 관한 실증분석 : 사회적 자본의 조절효과를 중심으로)

  • Cheon, Phyeong Uk;Chung, Dong Seop;Ock, Young Seok
    • Asia-Pacific Journal of Business Venturing and Entrepreneurship
    • /
    • v.9 no.2
    • /
    • pp.201-219
    • /
    • 2014
  • The Republic of Korea in the real economy to a knowledge economy, and a center of creativity and imagination in the creative economy is changing the paradigm. As the core of creating economic, creative industries with the technology and information play an important role in the industry individuals. In order to solve the problem of the polarization of the economy and high youth unemployment rate of Korea, to recognize the role of the creative industries, as objection part, dimensions pan-national and one creative companies in industries of Mobile Apps various policies that support has been promoted. Support these policies to be able to contribute to the establishment of the success of mobile apps one-person creating company, we performed this study targeting one-person company that creates mobile apps area, we conducted a demonstration study of success factors, and thus more effective and efficient in an attempt to seek out support measures. In this study, we derive a research 4 hypothesis about the success factors of one creative enterprise through literature discussion, a study was made on the basis of empirical data of one-person company that creates mobile apps. The results of the analysis, first, if the development rate of the mobile application technology is fast and a new competition associated product is appeared, it was possible to find a tendency to be higher at the performance quantitative companies. Second, if the founder is a founding for the benefit and rewarding work and come to terms with the risk, it was possible to discover tends to be higher achievement quantitative. Third, if one-person company select a target market with capture intensively, it was possible to find a tendency for higher qualitative results. Fourth, it could be found that the reliability of the contact frequency of the network related performance business environment these characteristics enterprise management strategy and act as a significant modulatory effect. Provision of information relating to management and entrepreneurship education to be one creative enterprise is required, these results suggest that there is a provision continuing need for the opportunity to be able to meet and network and reliable variety have. In this study, to take advantage to promote the elimination measures that can increase the likelihood of success of the company of institutions to support one company that creates knowledge-based, such as in the field of mobile application.

  • PDF

Early Identification of Gifted Young Children and Dynamic assessment (유아 영재의 판별과 역동적 평가)

  • 장영숙
    • Journal of Gifted/Talented Education
    • /
    • v.11 no.3
    • /
    • pp.131-153
    • /
    • 2001
  • The importance of identifying gifted children during early childhood is becoming recognized. Nonetheless, most researchers preferred to study the primary and secondary levels where children are already and more clearly demonstrating what talents they have, and where more reliable predictions of gifted may be made. Comparatively lisle work has been done in this area. When we identify giftedness during early childhood, we have to consider the potential of the young children rather than on actual achievement. Giftedness during early childhood is still developing and less stable than that of older children and this prevents us from making firm and accurate predictions based on children's actual achievement. Dynamic assessment, based on Vygotsky's concept of the zone of proximal development(ZPD), suggests a new idea in the way the gifted young children are identified. In light of dynamic assessment, for identifying the potential giftedness of young children. we need to involve measuring both unassisted and assisted performance. Dynamic assessment usually consists of a test-intervene-retest format that focuses attention on the improvement in child performance when an adult provides mediated assistance on how to master the testing task. The advantages of the dynamic assessment are as follows: First, the dynamic assessment approach can provide a useful means for assessing young gifted child who have not demonstrated high ability on traditional identification method. Second, the dynamic assessment approach can assess the learning process of young children. Third, the dynamic assessment can lead an individualized education by the early identification of young gifted children. Fourth, the dynamic assessment can be a more accurate predictor of potential by linking diagnosis and instruction. Thus, it can make us provide an educational treatment effectively for young gifted children.

  • PDF