• Title/Summary/Keyword: Input system

Search Result 11,540, Processing Time 0.037 seconds

The Pattern Analysis of Financial Distress for Non-audited Firms using Data Mining (데이터마이닝 기법을 활용한 비외감기업의 부실화 유형 분석)

  • Lee, Su Hyun;Park, Jung Min;Lee, Hyoung Yong
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.4
    • /
    • pp.111-131
    • /
    • 2015
  • There are only a handful number of research conducted on pattern analysis of corporate distress as compared with research for bankruptcy prediction. The few that exists mainly focus on audited firms because financial data collection is easier for these firms. But in reality, corporate financial distress is a far more common and critical phenomenon for non-audited firms which are mainly comprised of small and medium sized firms. The purpose of this paper is to classify non-audited firms under distress according to their financial ratio using data mining; Self-Organizing Map (SOM). SOM is a type of artificial neural network that is trained using unsupervised learning to produce a lower dimensional discretized representation of the input space of the training samples, called a map. SOM is different from other artificial neural networks as it applies competitive learning as opposed to error-correction learning such as backpropagation with gradient descent, and in the sense that it uses a neighborhood function to preserve the topological properties of the input space. It is one of the popular and successful clustering algorithm. In this study, we classify types of financial distress firms, specially, non-audited firms. In the empirical test, we collect 10 financial ratios of 100 non-audited firms under distress in 2004 for the previous two years (2002 and 2003). Using these financial ratios and the SOM algorithm, five distinct patterns were distinguished. In pattern 1, financial distress was very serious in almost all financial ratios. 12% of the firms are included in these patterns. In pattern 2, financial distress was weak in almost financial ratios. 14% of the firms are included in pattern 2. In pattern 3, growth ratio was the worst among all patterns. It is speculated that the firms of this pattern may be under distress due to severe competition in their industries. Approximately 30% of the firms fell into this group. In pattern 4, the growth ratio was higher than any other pattern but the cash ratio and profitability ratio were not at the level of the growth ratio. It is concluded that the firms of this pattern were under distress in pursuit of expanding their business. About 25% of the firms were in this pattern. Last, pattern 5 encompassed very solvent firms. Perhaps firms of this pattern were distressed due to a bad short-term strategic decision or due to problems with the enterpriser of the firms. Approximately 18% of the firms were under this pattern. This study has the academic and empirical contribution. In the perspectives of the academic contribution, non-audited companies that tend to be easily bankrupt and have the unstructured or easily manipulated financial data are classified by the data mining technology (Self-Organizing Map) rather than big sized audited firms that have the well prepared and reliable financial data. In the perspectives of the empirical one, even though the financial data of the non-audited firms are conducted to analyze, it is useful for find out the first order symptom of financial distress, which makes us to forecast the prediction of bankruptcy of the firms and to manage the early warning and alert signal. These are the academic and empirical contribution of this study. The limitation of this research is to analyze only 100 corporates due to the difficulty of collecting the financial data of the non-audited firms, which make us to be hard to proceed to the analysis by the category or size difference. Also, non-financial qualitative data is crucial for the analysis of bankruptcy. Thus, the non-financial qualitative factor is taken into account for the next study. This study sheds some light on the non-audited small and medium sized firms' distress prediction in the future.

Robo-Advisor Algorithm with Intelligent View Model (지능형 전망모형을 결합한 로보어드바이저 알고리즘)

  • Kim, Sunwoong
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.39-55
    • /
    • 2019
  • Recently banks and large financial institutions have introduced lots of Robo-Advisor products. Robo-Advisor is a Robot to produce the optimal asset allocation portfolio for investors by using the financial engineering algorithms without any human intervention. Since the first introduction in Wall Street in 2008, the market size has grown to 60 billion dollars and is expected to expand to 2,000 billion dollars by 2020. Since Robo-Advisor algorithms suggest asset allocation output to investors, mathematical or statistical asset allocation strategies are applied. Mean variance optimization model developed by Markowitz is the typical asset allocation model. The model is a simple but quite intuitive portfolio strategy. For example, assets are allocated in order to minimize the risk on the portfolio while maximizing the expected return on the portfolio using optimization techniques. Despite its theoretical background, both academics and practitioners find that the standard mean variance optimization portfolio is very sensitive to the expected returns calculated by past price data. Corner solutions are often found to be allocated only to a few assets. The Black-Litterman Optimization model overcomes these problems by choosing a neutral Capital Asset Pricing Model equilibrium point. Implied equilibrium returns of each asset are derived from equilibrium market portfolio through reverse optimization. The Black-Litterman model uses a Bayesian approach to combine the subjective views on the price forecast of one or more assets with implied equilibrium returns, resulting a new estimates of risk and expected returns. These new estimates can produce optimal portfolio by the well-known Markowitz mean-variance optimization algorithm. If the investor does not have any views on his asset classes, the Black-Litterman optimization model produce the same portfolio as the market portfolio. What if the subjective views are incorrect? A survey on reports of stocks performance recommended by securities analysts show very poor results. Therefore the incorrect views combined with implied equilibrium returns may produce very poor portfolio output to the Black-Litterman model users. This paper suggests an objective investor views model based on Support Vector Machines(SVM), which have showed good performance results in stock price forecasting. SVM is a discriminative classifier defined by a separating hyper plane. The linear, radial basis and polynomial kernel functions are used to learn the hyper planes. Input variables for the SVM are returns, standard deviations, Stochastics %K and price parity degree for each asset class. SVM output returns expected stock price movements and their probabilities, which are used as input variables in the intelligent views model. The stock price movements are categorized by three phases; down, neutral and up. The expected stock returns make P matrix and their probability results are used in Q matrix. Implied equilibrium returns vector is combined with the intelligent views matrix, resulting the Black-Litterman optimal portfolio. For comparisons, Markowitz mean-variance optimization model and risk parity model are used. The value weighted market portfolio and equal weighted market portfolio are used as benchmark indexes. We collect the 8 KOSPI 200 sector indexes from January 2008 to December 2018 including 132 monthly index values. Training period is from 2008 to 2015 and testing period is from 2016 to 2018. Our suggested intelligent view model combined with implied equilibrium returns produced the optimal Black-Litterman portfolio. The out of sample period portfolio showed better performance compared with the well-known Markowitz mean-variance optimization portfolio, risk parity portfolio and market portfolio. The total return from 3 year-period Black-Litterman portfolio records 6.4%, which is the highest value. The maximum draw down is -20.8%, which is also the lowest value. Sharpe Ratio shows the highest value, 0.17. It measures the return to risk ratio. Overall, our suggested view model shows the possibility of replacing subjective analysts's views with objective view model for practitioners to apply the Robo-Advisor asset allocation algorithms in the real trading fields.

A Deep Learning Based Approach to Recognizing Accompanying Status of Smartphone Users Using Multimodal Data (스마트폰 다종 데이터를 활용한 딥러닝 기반의 사용자 동행 상태 인식)

  • Kim, Kilho;Choi, Sangwoo;Chae, Moon-jung;Park, Heewoong;Lee, Jaehong;Park, Jonghun
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.163-177
    • /
    • 2019
  • As smartphones are getting widely used, human activity recognition (HAR) tasks for recognizing personal activities of smartphone users with multimodal data have been actively studied recently. The research area is expanding from the recognition of the simple body movement of an individual user to the recognition of low-level behavior and high-level behavior. However, HAR tasks for recognizing interaction behavior with other people, such as whether the user is accompanying or communicating with someone else, have gotten less attention so far. And previous research for recognizing interaction behavior has usually depended on audio, Bluetooth, and Wi-Fi sensors, which are vulnerable to privacy issues and require much time to collect enough data. Whereas physical sensors including accelerometer, magnetic field and gyroscope sensors are less vulnerable to privacy issues and can collect a large amount of data within a short time. In this paper, a method for detecting accompanying status based on deep learning model by only using multimodal physical sensor data, such as an accelerometer, magnetic field and gyroscope, was proposed. The accompanying status was defined as a redefinition of a part of the user interaction behavior, including whether the user is accompanying with an acquaintance at a close distance and the user is actively communicating with the acquaintance. A framework based on convolutional neural networks (CNN) and long short-term memory (LSTM) recurrent networks for classifying accompanying and conversation was proposed. First, a data preprocessing method which consists of time synchronization of multimodal data from different physical sensors, data normalization and sequence data generation was introduced. We applied the nearest interpolation to synchronize the time of collected data from different sensors. Normalization was performed for each x, y, z axis value of the sensor data, and the sequence data was generated according to the sliding window method. Then, the sequence data became the input for CNN, where feature maps representing local dependencies of the original sequence are extracted. The CNN consisted of 3 convolutional layers and did not have a pooling layer to maintain the temporal information of the sequence data. Next, LSTM recurrent networks received the feature maps, learned long-term dependencies from them and extracted features. The LSTM recurrent networks consisted of two layers, each with 128 cells. Finally, the extracted features were used for classification by softmax classifier. The loss function of the model was cross entropy function and the weights of the model were randomly initialized on a normal distribution with an average of 0 and a standard deviation of 0.1. The model was trained using adaptive moment estimation (ADAM) optimization algorithm and the mini batch size was set to 128. We applied dropout to input values of the LSTM recurrent networks to prevent overfitting. The initial learning rate was set to 0.001, and it decreased exponentially by 0.99 at the end of each epoch training. An Android smartphone application was developed and released to collect data. We collected smartphone data for a total of 18 subjects. Using the data, the model classified accompanying and conversation by 98.74% and 98.83% accuracy each. Both the F1 score and accuracy of the model were higher than the F1 score and accuracy of the majority vote classifier, support vector machine, and deep recurrent neural network. In the future research, we will focus on more rigorous multimodal sensor data synchronization methods that minimize the time stamp differences. In addition, we will further study transfer learning method that enables transfer of trained models tailored to the training data to the evaluation data that follows a different distribution. It is expected that a model capable of exhibiting robust recognition performance against changes in data that is not considered in the model learning stage will be obtained.

Index-based Searching on Timestamped Event Sequences (타임스탬프를 갖는 이벤트 시퀀스의 인덱스 기반 검색)

  • 박상현;원정임;윤지희;김상욱
    • Journal of KIISE:Databases
    • /
    • v.31 no.5
    • /
    • pp.468-478
    • /
    • 2004
  • It is essential in various application areas of data mining and bioinformatics to effectively retrieve the occurrences of interesting patterns from sequence databases. For example, let's consider a network event management system that records the types and timestamp values of events occurred in a specific network component(ex. router). The typical query to find out the temporal casual relationships among the network events is as fellows: 'Find all occurrences of CiscoDCDLinkUp that are fellowed by MLMStatusUP that are subsequently followed by TCPConnectionClose, under the constraint that the interval between the first two events is not larger than 20 seconds, and the interval between the first and third events is not larger than 40 secondsTCPConnectionClose. This paper proposes an indexing method that enables to efficiently answer such a query. Unlike the previous methods that rely on inefficient sequential scan methods or data structures not easily supported by DBMSs, the proposed method uses a multi-dimensional spatial index, which is proven to be efficient both in storage and search, to find the answers quickly without false dismissals. Given a sliding window W, the input to a multi-dimensional spatial index is a n-dimensional vector whose i-th element is the interval between the first event of W and the first occurrence of the event type Ei in W. Here, n is the number of event types that can be occurred in the system of interest. The problem of‘dimensionality curse’may happen when n is large. Therefore, we use the dimension selection or event type grouping to avoid this problem. The experimental results reveal that our proposed technique can be a few orders of magnitude faster than the sequential scan and ISO-Depth index methods.hods.

Behavioural Analysis of Password Authentication and Countermeasure to Phishing Attacks - from User Experience and HCI Perspectives (사용자의 패스워드 인증 행위 분석 및 피싱 공격시 대응방안 - 사용자 경험 및 HCI의 관점에서)

  • Ryu, Hong Ryeol;Hong, Moses;Kwon, Taekyoung
    • Journal of Internet Computing and Services
    • /
    • v.15 no.3
    • /
    • pp.79-90
    • /
    • 2014
  • User authentication based on ID and PW has been widely used. As the Internet has become a growing part of people' lives, input times of ID/PW have been increased for a variety of services. People have already learned enough to perform the authentication procedure and have entered ID/PW while ones are unconscious. This is referred to as the adaptive unconscious, a set of mental processes incoming information and producing judgements and behaviors without our conscious awareness and within a second. Most people have joined up for various websites with a small number of IDs/PWs, because they relied on their memory for managing IDs/PWs. Human memory decays with the passing of time and knowledges in human memory tend to interfere with each other. For that reason, there is the potential for people to enter an invalid ID/PW. Therefore, these characteristics above mentioned regarding of user authentication with ID/PW can lead to human vulnerabilities: people use a few PWs for various websites, manage IDs/PWs depending on their memory, and enter ID/PW unconsciously. Based on the vulnerability of human factors, a variety of information leakage attacks such as phishing and pharming attacks have been increasing exponentially. In the past, information leakage attacks exploited vulnerabilities of hardware, operating system, software and so on. However, most of current attacks tend to exploit the vulnerabilities of the human factors. These attacks based on the vulnerability of the human factor are called social-engineering attacks. Recently, malicious social-engineering technique such as phishing and pharming attacks is one of the biggest security problems. Phishing is an attack of attempting to obtain valuable information such as ID/PW and pharming is an attack intended to steal personal data by redirecting a website's traffic to a fraudulent copy of a legitimate website. Screens of fraudulent copies used for both phishing and pharming attacks are almost identical to those of legitimate websites, and even the pharming can include the deceptive URL address. Therefore, without the supports of prevention and detection techniques such as vaccines and reputation system, it is difficult for users to determine intuitively whether the site is the phishing and pharming sites or legitimate site. The previous researches in terms of phishing and pharming attacks have mainly studied on technical solutions. In this paper, we focus on human behaviour when users are confronted by phishing and pharming attacks without knowing them. We conducted an attack experiment in order to find out how many IDs/PWs are leaked from pharming and phishing attack. We firstly configured the experimental settings in the same condition of phishing and pharming attacks and build a phishing site for the experiment. We then recruited 64 voluntary participants and asked them to log in our experimental site. For each participant, we conducted a questionnaire survey with regard to the experiment. Through the attack experiment and survey, we observed whether their password are leaked out when logging in the experimental phishing site, and how many different passwords are leaked among the total number of passwords of each participant. Consequently, we found out that most participants unconsciously logged in the site and the ID/PW management dependent on human memory caused the leakage of multiple passwords. The user should actively utilize repudiation systems and the service provider with online site should support prevention techniques that the user can intuitively determined whether the site is phishing.

The Present State and Solutions for Archival Arrangement and Description of National Archives & Records Service of Korea (국가기록원의 기록물 정리기술의 현황과 개선방안)

  • Yoon, Ju-Bom
    • Journal of Korean Society of Archives and Records Management
    • /
    • v.4 no.2
    • /
    • pp.118-162
    • /
    • 2004
  • Archival description in archives has an important role in document control and reference service. Archives has made an effort to do archival description. But we have some differences and problems about a theory and practical processes comparing with advanced countries. The serious difference in a theory is that a function classification, maintenance of an original order, arrangement of multi-level description are not reflected in practical process. they are arranged in shelves after they are arranged by registration order in a unit of a volume like an arrangement of book. In addition, there are problems in history of agency change or control of index. So these can cause inconvenience for users. For improving, in this study we introduced the meaning and importance of arrangement of description, the situation and problem of arrangement of description in The National Archives, and a description guideline in other foreign countries. The next is an example for ISAD(G). This paper has chapter 8, the chapter 1 is introduction, the chapter 2 is the meaning and importance of arrangement of description, excluding the chapter 8 is conclusion we can say like this from the chapter 3 to the chapter 7. In the chapter 3, we explain GOVT we are using now and description element category in situation and problem of arrangement of description in Archives. In the chapter 4, this is about guideline from Archives in U.S.A, England and Australia. 1. Lifecycle Date Requirement Guide from NARA is introduced and of the description field, the way of the description about just one title element is introduced. 2. This is about the guideline of the description from Public Record Office. That name is National Archives Cataloguing Guidelines Introduction. We are saying "PROCAT" from this guideline and the seven procedure of description. 3. This is about Commomon Record Series from National Archives of Australia. we studied Registration & description procedures for CRS system. In the chapter 5, This is about the example which applied ISAD to. Archives introduce description of documents produced from Appeals Commission in the Ministry of Government Administration. In the chapter 6, 7. These are about the problems we pointed after using ISAD, naming for the document at procedure section in every institution, the lack of description fields category, the sort or classification of the kind or form, the reference or identified number, the absence description rule about the details, function classification, multi-level description, input format, arrangement of book shelf, authority control. The plan for improving are that problems. The best way for arrangement and description in Archives is to examine the standard, guideline, manual from archives in the advanced countries. So we suggested we need many research and study about this in the academic field.

Development of Computer Program for the Arrangement of the Forest-road Network to Maximize the Investment Effect on the Forest-road Construction (임도개설(林道開設)에 있어서 투자효과(投資效果)를 최대(最大)로 하는 임도배치(林道配置)프로그램 개발(開發))

  • Park, Sang-Jun;Son, Doo-Sik
    • Journal of Korean Society of Forest Science
    • /
    • v.90 no.4
    • /
    • pp.420-430
    • /
    • 2001
  • The object of this study is to develop a computer program for the arrangement of the forest-road network maximizing the investment effect in forest-road construction with factors such as terrains, forest physiognomy, management plan, logging system, cost of forest-road construction, capacity of inputted labour, capacity of timber production and so on. The operating system developed by this study is Korean Windows 95/98 and Microsoft Visual Basic ver. 5.0. User interface was designed as systematic structure, it is presented as a kind of GUI(graphic user interface). The developed program has result of the most suitable forest-road arrangement, has suitable forest-road density calculated with cost of logging, cost of forest-road construction, diversion ratio of forest-road, cost of walking in forest. And the most suitable forest-road arrangement was designed for forest-road arrangement network which maximized investment effect through minimizing the sum of cost of logging and cost of forest-road construction. Input data were divided into map data and control data. Digital terrain model, division of forest-road layout plan, division of forest function and the existing road network are obtained from map data. on the other hand, cost of logging related terrain division, diversion ratio of forest-road and working road, cost of forest-road construction, cost of walking, cost of labor, walking speed, capacity of inputted labor, capacity of timber production and total distance of forest-road are inputted from control data. And map data was designed to be inputted by mesh method for common matrix. This program can be used to construct a new forest-road or vice forest-road which compensate already existing forest-road for the functional forestry.

  • PDF

Predicting Regional Soybean Yield using Crop Growth Simulation Model (작물 생육 모델을 이용한 지역단위 콩 수량 예측)

  • Ban, Ho-Young;Choi, Doug-Hwan;Ahn, Joong-Bae;Lee, Byun-Woo
    • Korean Journal of Remote Sensing
    • /
    • v.33 no.5_2
    • /
    • pp.699-708
    • /
    • 2017
  • The present study was to develop an approach for predicting soybean yield using a crop growth simulation model at the regional level where the detailed and site-specific information on cultivation management practices is not easily accessible for model input. CROPGRO-Soybean model included in Decision Support System for Agrotechnology Transfer (DSSAT) was employed for this study, and Illinois which is a major soybean production region of USA was selected as a study region. As a first step to predict soybean yield of Illinois using CROPGRO-Soybean model, genetic coefficients representative for each soybean maturity group (MG I~VI) were estimated through sowing date experiments using domestic and foreign cultivars with diverse maturity in Seoul National University Farm ($37.27^{\circ}N$, $126.99^{\circ}E$) for two years. The model using the representative genetic coefficients simulated the developmental stages of cultivars within each maturity group fairly well. Soybean yields for the grids of $10km{\times}10km$ in Illinois state were simulated from 2,000 to 2,011 with weather data under 18 simulation conditions including the combinations of three maturity groups, three seeding dates and two irrigation regimes. Planting dates and maturity groups were assigned differently to the three sub-regions divided longitudinally. The yearly state yields that were estimated by averaging all the grid yields simulated under non-irrigated and fully-Irrigated conditions showed a big difference from the statistical yields and did not explain the annual trend of yield increase due to the improved cultivation technologies. Using the grain yield data of 9 agricultural districts in Illinois observed and estimated from the simulated grid yield under 18 simulation conditions, a multiple regression model was constructed to estimate soybean yield at agricultural district level. In this model a year variable was also added to reflect the yearly yield trend. This model explained the yearly and district yield variation fairly well with a determination coefficients of $R^2=0.61$ (n = 108). Yearly state yields which were calculated by weighting the model-estimated yearly average agricultural district yield by the cultivation area of each agricultural district showed very close correspondence ($R^2=0.80$) to the yearly statistical state yields. Furthermore, the model predicted state yield fairly well in 2012 in which data were not used for the model construction and severe yield reduction was recorded due to drought.

Independent Verification Program for High-Dose-Rate Brachytherapy Treatment Plans (고선량률 근접치료계획의 정도보증 프로그램)

  • Han Youngyih;Chu Sung Sil;Huh Seung Jae;Suh Chang-Ok
    • Radiation Oncology Journal
    • /
    • v.21 no.3
    • /
    • pp.238-244
    • /
    • 2003
  • Purpose: The Planning of High-Dose-Rate (HDR) brachytherapy treatments are becoming individualized and more dependent on the treatment planning system. Therefore, computer software has been developed to perform independent point dose calculations with the integration of an isodose distribution curve display into the patient anatomy images. Meterials and Methods: As primary input data, the program takes patients'planning data including the source dwell positions, dwell times and the doses at reference points, computed by an HDR treatment planning system (TPS). Dosimetric calculations were peformed in a $10\times12\times10\;Cm^3$ grid space using the Interstitial Collaborative Working Group (ICWG) formalism and an anisotropy table for the HDR Iridium-192 source. The computed doses at the reference points were automatically compared with the relevant results of the TPS. The MR and simulation film images were then imported and the isodose distributions on the axial, sagittal and coronal planes intersecting the point selected by a user were superimposed on the imported images and then displayed. The accuracy of the software was tested in three benchmark plans peformed by Gamma-Med 12i TPS (MDS Nordion, Germany). Nine patients'plans generated by Plato (Nucletron Corporation, The Netherlands) were verified by the developed software. Results: The absolute doses computed by the developed software agreed with the commercial TPS results within an accuracy of $2.8\%$ in the benchmark plans. The isodose distribution plots showed excellent agreements with the exception of the tip legion of the source's longitudinal axis where a slight deviation was observed. In clinical plans, the secondary dose calculations had, on average, about a $3.4\%$ deviation from the TPS plans. Conclusion: The accurate validation of complicate treatment plans is possible with the developed software and the qualify of the HDR treatment plan can be improved with the isodose display integrated into the patient anatomy information.

A Web-based Internet Program for Nutritional Assessment and Diet Prescription by Renal Diseases (웹기반의 신장질환별 영양평가 밑 식사처방 프로그램)

  • 한지숙;김종경;전영수
    • Journal of the Korean Society of Food Science and Nutrition
    • /
    • v.31 no.5
    • /
    • pp.847-885
    • /
    • 2002
  • The purpose of this study was to develop a web-based internet program for nutritional assessment and diet Prescription by renal diseases. Renal diseases were classified by nephrotic syndrome, renal failure, hemodialysis and peritoneal dialysis. The system consisted of five parts according to their functions and contents. The first part is to assess the general health status such as body weight, obesity index, basal metabolic rate and total energy requirement by the input of age, sex, height, weight and degree of activity. The second part was designed to investigate dietary history of patient, that is, to find out his inappropriate dietary habit and give him some suggestions for appropriate dietary behavior by investigating his dietary history. This part also offers the diet and nutrition management by personal status with renal disease, and the information for food selection, snacks, convenience foods, dine-out, behavioral modification, cooking methods, food exchange lists and terms. The third part is evaluating their energy and nutrients intake by comparing with recommended dietary allowance for Koreans or standardized data for patient with renal disease. In this part, it is also analyzing energy and nutrients of food consumed by food group and meals, and evaluating the status of nutrient intake. The fort]1 one, a major part of the system, is implementing the diet and menu planning by using food exchange lists. This Part Provides the patient with menus lists and I day menu suitable to his weight, activity and the status of renal disease. The fifth part is providing information on energy and nutrients of foods and drinks, and top 20 foods classified by nutrients. These results are finally displayed as tabular forms and graphical forms on the computer screen.