• Title/Summary/Keyword: Multi-class classification

Search Result 224, Processing Time 0.026 seconds

An Incremental Multi Partition Averaging Algorithm Based on Memory Based Reasoning (메모리 기반 추론 기법에 기반한 점진적 다분할평균 알고리즘)

  • Yih, Hyeong-Il
    • Journal of IKEEE
    • /
    • v.12 no.1
    • /
    • pp.65-74
    • /
    • 2008
  • One of the popular methods used for pattern classification is the MBR (Memory-Based Reasoning) algorithm. Since it simply computes distances between a test pattern and training patterns or hyperplanes stored in memory, and then assigns the class of the nearest training pattern, it is notorious for memory usage and can't learn additional information from new data. In order to overcome this problem, we propose an incremental learning algorithm (iMPA). iMPA divides the entire pattern space into fixed number partitions, and generates representatives from each partition. Also, due to the fact that it can not learn additional information from new data, we present iMPA which can learn additional information from new data and not require access to the original data, used to train. Proposed methods have been successfully shown to exhibit comparable performance to k-NN with a lot less number of patterns and better result than EACH system which implements the NGE theory using benchmark data sets from UCI Machine Learning Repository.

  • PDF

Multi-classification Sensitive Image Detection Method Based on Lightweight Convolutional Neural Network

  • Yueheng Mao;Bin Song;Zhiyong Zhang;Wenhou Yang;Yu Lan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.5
    • /
    • pp.1433-1449
    • /
    • 2023
  • In recent years, the rapid development of social networks has led to a rapid increase in the amount of information available on the Internet, which contains a large amount of sensitive information related to pornography, politics, and terrorism. In the aspect of sensitive image detection, the existing machine learning algorithms are confronted with problems such as large model size, long training time, and slow detection speed when auditing and supervising. In order to detect sensitive images more accurately and quickly, this paper proposes a multiclassification sensitive image detection method based on lightweight Convolutional Neural Network. On the basis of the EfficientNet model, this method combines the Ghost Module idea of the GhostNet model and adds the SE channel attention mechanism in the Ghost Module for feature extraction training. The experimental results on the sensitive image data set constructed in this paper show that the accuracy of the proposed method in sensitive information detection is 94.46% higher than that of the similar methods. Then, the model is pruned through an ablation experiment, and the activation function is replaced by Hard-Swish, which reduces the parameters of the original model by 54.67%. Under the condition of ensuring accuracy, the detection time of a single image is reduced from 8.88ms to 6.37ms. The results of the experiment demonstrate that the method put forward has successfully enhanced the precision of identifying multi-class sensitive images, significantly decreased the number of parameters in the model, and achieved higher accuracy than comparable algorithms while using a more lightweight model design.

Hierarchical Internet Application Traffic Classification using a Multi-class SVM (다중 클래스 SVM을 이용한 계층적 인터넷 애플리케이션 트래픽의 분류)

  • Yu, Jae-Hak;Lee, Han-Sung;Im, Young-Hee;Kim, Myung-Sup;Park, Dai-Hee
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.20 no.1
    • /
    • pp.7-14
    • /
    • 2010
  • In this paper, we introduce a hierarchical internet application traffic classification system based on SVM as an alternative overcoming the uppermost limit of the conventional methodology which is using the port number or payload information. After selecting an optimal attribute subset of the bidirectional traffic flow data collected from the campus, the proposed system classifies the internet application traffic hierarchically. The system is composed of three layers: the first layer quickly determines P2P traffic and non-P2P traffic using a SVM, the second layer classifies P2P traffics into file-sharing, messenger, and TV, based on three SVDDs. The third layer makes specific classification of the entire 16 application traffics. By classifying the internet application traffic finely or coarsely, the proposed system can guarantee an efficient system resource management, a stable network environment, a seamless bandwidth, and an appropriate QoS. Also, even a new application traffic is added, it is possible to have a system incremental updating and scalability by training only a new SVDD without retraining the whole system. We validate the performance of our approach with computer experiments.

Evaluation of Corporate Distress Prediction Power using the Discriminant Analysis: The Case of First-Class Hotels in Seoul (판별분석에 의한 기업부실예측력 평가: 서울지역 특1급 호텔 사례 분석)

  • Kim, Si-Joong
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.17 no.10
    • /
    • pp.520-526
    • /
    • 2016
  • This study aims to develop a distress prediction model, in order to evaluate the distress prediction power for first-class hotels and to calculate the average financial ratio in the Seoul area by using the financial ratios of hotels in 2015. The sample data was collected from 19 first-class hotels in Seoul and the financial ratios extracted from 14 of these 19 hotels. The results show firstly that the seven financial ratios, viz. the current ratio, total borrowings and bonds payable to total assets, interest coverage ratio to operating income, operating income to sales, net income to stockholders' equity, ratio of cash flows from operating activities to sales and total assets turnover, enable the top-level corporations to be discriminated from the failed corporations and, secondly, by using these seven financial ratios, a discriminant function which classifies the corporations into top-level and failed ones is estimated by linear multiple discriminant analysis. The accuracy of prediction of this discriminant capability turned out to be 87.9%. The accuracy of the estimates obtained by discriminant analysis indicates that the distress prediction model's distress prediction power is 78.95%. According to the analysis results, hotel management groups which administrate low level corporations need to focus on the classification of these seven financial ratios. Furthermore, hotel corporations have very different financial structures and failure prediction indicators from other industries. In accordance with this finding, for the development of credit evaluation systems for such hotel corporations, there is a need for systems to be developed that reflect hotel corporations' financial features.

A Study on Improvement of the Human Posture Estimation Method for Performing Robots (공연로봇을 위한 인간자세 추정방법 개선에 관한 연구)

  • Park, Cheonyu;Park, Jaehun;Han, Jeakweon
    • Journal of Broadcast Engineering
    • /
    • v.25 no.5
    • /
    • pp.750-757
    • /
    • 2020
  • One of the basic tasks for robots to interact with humans is to quickly and accurately grasp human behavior. Therefore, it is necessary to increase the accuracy of human pose recognition when the robot is estimating the human pose and to recognize it as quickly as possible. However, when the human pose is estimated using deep learning, which is a representative method of artificial intelligence technology, recognition accuracy and speed are not satisfied at the same time. Therefore, it is common to select one of a top-down method that has high inference accuracy or a bottom-up method that has high processing speed. In this paper, we propose two methods that complement the disadvantages while including both the advantages of the two methods mentioned above. The first is to perform parallel inference on the server using multi GPU, and the second is to mix bottom-up and One-class Classification. As a result of the experiment, both of the methods presented in this paper showed improvement in speed. If these two methods are applied to the entertainment robot, it is expected that a highly reliable interaction with the audience can be performed.

Comparison Between Methods for Suitability Classification of Wild Edible Greens (산채류 재배적지 기준설정 방법 간의 비교 분석)

  • Hyun, Byung-Keun;Jung, Sug-Jae;Sonn, Yeon-Kyu;Park, Chan-Won;Zhang, Young-Seon;Song, Kwan-Cheol;Kim, Lee-Hyun;Choi, Eun-Young;Hong, Suk-Young;Kwon, Sun-Ik;Jang, Byoung-Choon
    • Korean Journal of Soil Science and Fertilizer
    • /
    • v.43 no.5
    • /
    • pp.696-704
    • /
    • 2010
  • The objective of this study was analysis of two methods of land suitability classification for wild edible green. One method was Maximum limiting factor method (MLFM) and the other was Multi-regression method (MRM) for land suitability classification for wild edible green. The investigation was carried out in Pyeongchang, Hongcheong, Hoeingseong, and Yanggu regions in Korea. The obtained results showed that factors related to the decision classification of the land suitability for wild edible green cultivation were land slope, altitude, soil morphology and gravel contents so on. The classification of the best suitability soil for wild edible greens were fine loamy (silty), valley or fan of soil morphology, well drainage class, B-slope (2~7%), available soil depth deeper than 100cm, and altitude higher than 501m. Contribution of soil that influence to crop yields using Multi-regression method were slope 0.30, altitude 0.22, soil morphology 0.13, drainage classes 0.09, available soil depth 0.07, and soil texture 0.01 orders. Using MLFM, area of best suitable land was 0.2%, suitable soil 15.0%, possible soil 16.7%, and low productive soil 68.0% in Hongcheon region of Gangwon province. But, area of best suitable land was 35.1%, suitable soil 30.7%, possible soil 10.3%, and low productive soil 23.9% by MRM. There was big difference of suitable soil area between two methods (MLFM and MRM). When decision classificatin of the land suitability for wild edible green cultivation should consider enough analysis methods. Furthermore, to establishment of land suitability classification for crop would be better use MRM than MLFM.

A screening of Alzheimer's disease using basis synthesis by singular value decomposition from Raman spectra of platelet (혈소판 라만 스펙트럼에서 특이값 분해에 의한 기저 합성을 통한 알츠하이머병 검출)

  • Park, Aaron;Baek, Sung-June
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.14 no.5
    • /
    • pp.2393-2399
    • /
    • 2013
  • In this paper, we proposed a method to screening of Alzheimer's disease (AD) from Raman spectra of platelet with synthesis of basis spectra using singular value decomposition (SVD). Raman spectra of platelet from AD transgenic mice are preprocessed with denoising, removal background and normalization method. The column vectors of each data matrix consist of Raman spectrum of AD and normal (NR). The matrix is factorized using SVD algorithm and then the basis spectra of AD and NR are determined by 12 column vectors of each matrix. The classification process is completed by select the class that minimized the root-mean-square error between the validation spectrum and the linear synthesized spectrum of the basis spectra. According to the experiments involving 278 Raman spectra, the proposed method gave about 97.6% classification rate, which is better performance about 6.1% than multi-layer perceptron (MLP) with extracted features using principle components analysis (PCA). The results show that the basis spectra using SVD is well suited for the diagnosis of AD by Raman spectra from platelet.

Feasibility of Deep Learning Algorithms for Binary Classification Problems (이진 분류문제에서의 딥러닝 알고리즘의 활용 가능성 평가)

  • Kim, Kitae;Lee, Bomi;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.95-108
    • /
    • 2017
  • Recently, AlphaGo which is Bakuk (Go) artificial intelligence program by Google DeepMind, had a huge victory against Lee Sedol. Many people thought that machines would not be able to win a man in Go games because the number of paths to make a one move is more than the number of atoms in the universe unlike chess, but the result was the opposite to what people predicted. After the match, artificial intelligence technology was focused as a core technology of the fourth industrial revolution and attracted attentions from various application domains. Especially, deep learning technique have been attracted as a core artificial intelligence technology used in the AlphaGo algorithm. The deep learning technique is already being applied to many problems. Especially, it shows good performance in image recognition field. In addition, it shows good performance in high dimensional data area such as voice, image and natural language, which was difficult to get good performance using existing machine learning techniques. However, in contrast, it is difficult to find deep leaning researches on traditional business data and structured data analysis. In this study, we tried to find out whether the deep learning techniques have been studied so far can be used not only for the recognition of high dimensional data but also for the binary classification problem of traditional business data analysis such as customer churn analysis, marketing response prediction, and default prediction. And we compare the performance of the deep learning techniques with that of traditional artificial neural network models. The experimental data in the paper is the telemarketing response data of a bank in Portugal. It has input variables such as age, occupation, loan status, and the number of previous telemarketing and has a binary target variable that records whether the customer intends to open an account or not. In this study, to evaluate the possibility of utilization of deep learning algorithms and techniques in binary classification problem, we compared the performance of various models using CNN, LSTM algorithm and dropout, which are widely used algorithms and techniques in deep learning, with that of MLP models which is a traditional artificial neural network model. However, since all the network design alternatives can not be tested due to the nature of the artificial neural network, the experiment was conducted based on restricted settings on the number of hidden layers, the number of neurons in the hidden layer, the number of output data (filters), and the application conditions of the dropout technique. The F1 Score was used to evaluate the performance of models to show how well the models work to classify the interesting class instead of the overall accuracy. The detail methods for applying each deep learning technique in the experiment is as follows. The CNN algorithm is a method that reads adjacent values from a specific value and recognizes the features, but it does not matter how close the distance of each business data field is because each field is usually independent. In this experiment, we set the filter size of the CNN algorithm as the number of fields to learn the whole characteristics of the data at once, and added a hidden layer to make decision based on the additional features. For the model having two LSTM layers, the input direction of the second layer is put in reversed position with first layer in order to reduce the influence from the position of each field. In the case of the dropout technique, we set the neurons to disappear with a probability of 0.5 for each hidden layer. The experimental results show that the predicted model with the highest F1 score was the CNN model using the dropout technique, and the next best model was the MLP model with two hidden layers using the dropout technique. In this study, we were able to get some findings as the experiment had proceeded. First, models using dropout techniques have a slightly more conservative prediction than those without dropout techniques, and it generally shows better performance in classification. Second, CNN models show better classification performance than MLP models. This is interesting because it has shown good performance in binary classification problems which it rarely have been applied to, as well as in the fields where it's effectiveness has been proven. Third, the LSTM algorithm seems to be unsuitable for binary classification problems because the training time is too long compared to the performance improvement. From these results, we can confirm that some of the deep learning algorithms can be applied to solve business binary classification problems.

Clinical Manifestation and Treatment Outcome of Lupus Nephritis in Children (소아 루프스 신염의 임상양상 및 치료결과)

  • Park Jee-Min;Shin Jae-Il;Kim Pyung-Kil;Lee Jae-Seung
    • Childhood Kidney Diseases
    • /
    • v.6 no.2
    • /
    • pp.155-168
    • /
    • 2002
  • Purpose; Systemic lupus erythematosus(SLE) is an autoimmune disease with multi-system involvement and renal damage is a major cause of morbidity and mortality in children. Renal involvement is more common and severe in children than in adults. Therefore, renal biopsy plays a crucial role in planning effective therapy. In this study, we investigated the clinical and pathological findings of lupus nephritis in children to aid clinical care of the disease. Methods: The clinical and pathological data of 40 patients who were diagnosed as SLE with renal involvement in Shinchon Severance Hospital from Jan. 1990 to Sep. 2002 were analyzed retrospectively. Results: The ratio of male to female patients was 1:3 and the median age at diagnosis was 12.1(2-18) years old. FANA(95.0%), anti-ds DNA antibody(87.5%), malar rash(80.0%) were the most common findings among the classification criteria by ARA. Microscopic hematuria with proteinuria(75.0%), nephrotic syndrome(55.0%), and microscopic hematuria alone(15.0%) were the most common renal presentations in the respective order at diagnosis. There were 27 cases with WHO class IV lupus nephritis confirmed by renal biopsy and 3 cases with pathological changes of WHO class type. Different treatment modalities were carried out : prednisolone only in 5 cases, prednisol-one+azat-hioprine in 9 cases, prednisolone+azathioprine+intravenous cyclophosphamide in 14 cases, prednisolone+cyclosporine A+intravenous cyclophosphamide in 12 cases, plasma exchange in 9 cases and intravenous gamma-globulin in 2 cases. The average follow-up period was $51.8{\pm}40.5$ months. During $51.8{\pm}40.5$ months. During follow-up, 4 patients expired. The risk factors associated with mortality were male, WHO class IV and acute renal failure at diagnosis. Conclusion: Renal involvement was noted in 63.5% of childhood SLE, and 67.5% of renal lesion was WHO class IV lupus nephritis which is known to be associated with a poor prognosis. Therefore aggressive treatment employing immunosuppressant during the early stages of disease could be helpful in improving long-term prognosis. But careful attention should be given to optimize the treatment due to unique problems associated with growth, psychosocial development and gonadal toxicity, especially in children.

  • PDF

Ensemble of Nested Dichotomies for Activity Recognition Using Accelerometer Data on Smartphone (Ensemble of Nested Dichotomies 기법을 이용한 스마트폰 가속도 센서 데이터 기반의 동작 인지)

  • Ha, Eu Tteum;Kim, Jeongmin;Ryu, Kwang Ryel
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.4
    • /
    • pp.123-132
    • /
    • 2013
  • As the smartphones are equipped with various sensors such as the accelerometer, GPS, gravity sensor, gyros, ambient light sensor, proximity sensor, and so on, there have been many research works on making use of these sensors to create valuable applications. Human activity recognition is one such application that is motivated by various welfare applications such as the support for the elderly, measurement of calorie consumption, analysis of lifestyles, analysis of exercise patterns, and so on. One of the challenges faced when using the smartphone sensors for activity recognition is that the number of sensors used should be minimized to save the battery power. When the number of sensors used are restricted, it is difficult to realize a highly accurate activity recognizer or a classifier because it is hard to distinguish between subtly different activities relying on only limited information. The difficulty gets especially severe when the number of different activity classes to be distinguished is very large. In this paper, we show that a fairly accurate classifier can be built that can distinguish ten different activities by using only a single sensor data, i.e., the smartphone accelerometer data. The approach that we take to dealing with this ten-class problem is to use the ensemble of nested dichotomy (END) method that transforms a multi-class problem into multiple two-class problems. END builds a committee of binary classifiers in a nested fashion using a binary tree. At the root of the binary tree, the set of all the classes are split into two subsets of classes by using a binary classifier. At a child node of the tree, a subset of classes is again split into two smaller subsets by using another binary classifier. Continuing in this way, we can obtain a binary tree where each leaf node contains a single class. This binary tree can be viewed as a nested dichotomy that can make multi-class predictions. Depending on how a set of classes are split into two subsets at each node, the final tree that we obtain can be different. Since there can be some classes that are correlated, a particular tree may perform better than the others. However, we can hardly identify the best tree without deep domain knowledge. The END method copes with this problem by building multiple dichotomy trees randomly during learning, and then combining the predictions made by each tree during classification. The END method is generally known to perform well even when the base learner is unable to model complex decision boundaries As the base classifier at each node of the dichotomy, we have used another ensemble classifier called the random forest. A random forest is built by repeatedly generating a decision tree each time with a different random subset of features using a bootstrap sample. By combining bagging with random feature subset selection, a random forest enjoys the advantage of having more diverse ensemble members than a simple bagging. As an overall result, our ensemble of nested dichotomy can actually be seen as a committee of committees of decision trees that can deal with a multi-class problem with high accuracy. The ten classes of activities that we distinguish in this paper are 'Sitting', 'Standing', 'Walking', 'Running', 'Walking Uphill', 'Walking Downhill', 'Running Uphill', 'Running Downhill', 'Falling', and 'Hobbling'. The features used for classifying these activities include not only the magnitude of acceleration vector at each time point but also the maximum, the minimum, and the standard deviation of vector magnitude within a time window of the last 2 seconds, etc. For experiments to compare the performance of END with those of other methods, the accelerometer data has been collected at every 0.1 second for 2 minutes for each activity from 5 volunteers. Among these 5,900 ($=5{\times}(60{\times}2-2)/0.1$) data collected for each activity (the data for the first 2 seconds are trashed because they do not have time window data), 4,700 have been used for training and the rest for testing. Although 'Walking Uphill' is often confused with some other similar activities, END has been found to classify all of the ten activities with a fairly high accuracy of 98.4%. On the other hand, the accuracies achieved by a decision tree, a k-nearest neighbor, and a one-versus-rest support vector machine have been observed as 97.6%, 96.5%, and 97.6%, respectively.