• Title/Summary/Keyword: Fuzzy Mining

Search Result 120, Processing Time 0.023 seconds

Application of Fuzzy Reasoning Method for Prediction of Subsidence Occurrences in Abandoned Mine Area (폐광산 지역에서의 지반침하예측을 위한 퍼지추론기법 적용 연구)

  • Choi, Sung-O.;Kim, Jae-Dong;Choi, Gwang-Su
    • Tunnel and Underground Space
    • /
    • v.19 no.5
    • /
    • pp.463-472
    • /
    • 2009
  • Many old domestic mines were excavated with the room and pillar method or the sublevel caving method and they involve the great possibility of surface subsidence, especially in the shallow depth mines. In most of these cases, the mine roadways and openings are very irregular in shape and the information about the local geology is uncertain. Consequently it is not simple to standardize the estimation method for the possibility of subsidence, especially the sinkhole subsidence. In this study, the fuzzy reasoning method has been applied for development of estimating the possibility of subsidence occurrence in abandoned mine area. This method has the advantage in producing the reliable estimation results with a simple performance procedure even when the precise information on the local geology and mining conditions is rare. For the verification of applicability of this method, the developed method has been applied to Kumho mine in Bonghwa, Kyungbook province and the Choong-ju mine in Iryu, Choongbook province where the surface subsidence occurred already.

Malicious Codes Re-grouping Methods using Fuzzy Clustering based on Native API Frequency (Native API 빈도 기반의 퍼지 군집화를 이용한 악성코드 재그룹화 기법연구)

  • Kwon, O-Chul;Bae, Seong-Jae;Cho, Jae-Ik;Moon, Jung-Sub
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.18 no.6A
    • /
    • pp.115-127
    • /
    • 2008
  • The Native API is a system call which can only be accessed with the authentication of the administrator. It can be used to detect a variety of malicious codes which can only be executed with the administrator's authority. Therefore, much research is being done on detection methods using the characteristics of the Native API. Most of these researches are being done by using supervised learning methods of machine learning. However, the classification standards of Anti-Virus companies do not reflect the characteristics of the Native API. As a result the population data used in the supervised learning methods are not accurate. Therefore, more research is needed on the topic of classification standards using the Native API for detection. This paper proposes a method for re-grouping malicious codes using fuzzy clustering methods with the Native API standard. The accuracy of the proposed re-grouping method uses machine learning to compare detection rates with previous classifying methods for evaluation.

Actor-Critic Algorithm with Transition Cost Estimation

  • Sergey, Denisov;Lee, Jee-Hyong
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.16 no.4
    • /
    • pp.270-275
    • /
    • 2016
  • We present an approach for acceleration actor-critic algorithm for reinforcement learning with continuous action space. Actor-critic algorithm has already proved its robustness to the infinitely large action spaces in various high dimensional environments. Despite that success, the main problem of the actor-critic algorithm remains the same-speed of convergence to the optimal policy. In high dimensional state and action space, a searching for the correct action in each state takes enormously long time. Therefore, in this paper we suggest a search accelerating function that allows to leverage speed of algorithm convergence and reach optimal policy faster. In our method, we assume that actions may have their own distribution of preference, that independent on the state. Since in the beginning of learning agent act randomly in the environment, it would be more efficient if actions were taken according to the some heuristic function. We demonstrate that heuristically-accelerated actor-critic algorithm learns optimal policy faster, using Educational Process Mining dataset with records of students' course learning process and their grades.

Document Summarization via Convex-Concave Programming

  • Kim, Minyoung
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.16 no.4
    • /
    • pp.293-298
    • /
    • 2016
  • Document summarization is an important task in various areas where the goal is to select a few the most descriptive sentences from a given document as a succinct summary. Even without training data of human labeled summaries, there has been several interesting existing work in the literature that yields reasonable performance. In this paper, within the same unsupervised learning setup, we propose a more principled learning framework for the document summarization task. Specifically we formulate an optimization problem that expresses the requirements of both faithful preservation of the document contents and the summary length constraint. We circumvent the difficult integer programming originating from binary sentence selection via continuous relaxation and the low entropy penalization. We also suggest an efficient convex-concave optimization solver algorithm that guarantees to improve the original objective at every iteration. For several document datasets, we demonstrate that the proposed learning algorithm significantly outperforms the existing approaches.

Protein Named Entity Identification Based on Probabilistic Features Derived from GENIA Corpus and Medical Text on the Web

  • Sumathipala, Sagara;Yamada, Koichi;Unehara, Muneyuki;Suzuki, Izumi
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.15 no.2
    • /
    • pp.111-120
    • /
    • 2015
  • Protein named entity identification is one of the most essential and fundamental predecessor for extracting information about protein-protein interactions from biomedical literature. In this paper, we explore the use of abstracts of biomedical literature in MEDLINE for protein name identification and present the results of the conducted experiments. We present a robust and effective approach to classify biomedical named entities into protein and non-protein classes, based on a rich set of features: orthographic, keyword, morphological and newly introduced Protein-Score features. Our procedure shows significant performance in the experiments on GENIA corpus using Random Forest, achieving the highest values of precision 92.7%, recall 91.7%, and F-measure 92.2% for protein identification, while reducing the training and testing time significantly.

Developing Takagi-Sugeno Fuzzy Model-Based Estimator for Short-Term Load Forecasting (단기부하예측을 위한 Tskagi-Sugeno 퍼지 모델 기반 예측기 설계)

  • 김도완;박진배;장권규;정근호;주영훈
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2004.04a
    • /
    • pp.523-527
    • /
    • 2004
  • This paper presents a new design methods of the short-term load forecasting system (STLFS) using the data mining. The proposed predictor takes form of the convex combination of the linear time series predictors for each inputs. The problem of estimating the consequent parameters is formulated by the convex optimization problem, which is to minimize the norm distance between the real load and the output of the linear time series estimator, The problem of estimating the premise parameters is to find the parameter value minimizing the error between the real load and the overall output. Finally, to show the feasibility of the proposed method, this paper provides the short-term load forecasting example.

  • PDF

A Classification Method Using Data Reduction

  • Uhm, Daiho;Jun, Sung-Hae;Lee, Seung-Joo
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.12 no.1
    • /
    • pp.1-5
    • /
    • 2012
  • Data reduction has been used widely in data mining for convenient analysis. Principal component analysis (PCA) and factor analysis (FA) methods are popular techniques. The PCA and FA reduce the number of variables to avoid the curse of dimensionality. The curse of dimensionality is to increase the computing time exponentially in proportion to the number of variables. So, many methods have been published for dimension reduction. Also, data augmentation is another approach to analyze data efficiently. Support vector machine (SVM) algorithm is a representative technique for dimension augmentation. The SVM maps original data to a feature space with high dimension to get the optimal decision plane. Both data reduction and augmentation have been used to solve diverse problems in data analysis. In this paper, we compare the strengths and weaknesses of dimension reduction and augmentation for classification and propose a classification method using data reduction for classification. We will carry out experiments for comparative studies to verify the performance of this research.

EXTENDED ONLINE DIVISIVE AGGLOMERATIVE CLUSTERING

  • Musa, Ibrahim Musa Ishag;Lee, Dong-Gyu;Ryu, Keun-Ho
    • Proceedings of the KSRS Conference
    • /
    • 2008.10a
    • /
    • pp.406-409
    • /
    • 2008
  • Clustering data streams has an importance over many applications like sensor networks. Existing hierarchical methods follow a semi fuzzy clustering that yields duplicate clusters. In order to solve the problems, we propose an extended online divisive agglomerative clustering on data streams. It builds a tree-like top-down hierarchy of clusters that evolves with data streams using geometric time frame for snapshots. It is an enhancement of the Online Divisive Agglomerative Clustering (ODAC) with a pruning strategy to avoid duplicate clusters. Our main features are providing update time and memory space which is independent of the number of examples on data streams. It can be utilized for clustering sensor data and network monitoring as well as web click streams.

  • PDF

A Prediction Model Based on Relevance Vector Machine and Granularity Analysis

  • Cho, Young Im
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.16 no.3
    • /
    • pp.157-162
    • /
    • 2016
  • In this paper, a yield prediction model based on relevance vector machine (RVM) and a granular computing model (quotient space theory) is presented. With a granular computing model, massive and complex meteorological data can be analyzed at different layers of different grain sizes, and new meteorological feature data sets can be formed in this way. In order to forecast the crop yield, a grey model is introduced to label the training sample data sets, which also can be used for computing the tendency yield. An RVM algorithm is introduced as the classification model for meteorological data mining. Experiments on data sets from the real world using this model show an advantage in terms of yield prediction compared with other models.

A Co-Evolutionary Computing for Statistical Learning Theory

  • Jun Sung-Hae
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.5 no.4
    • /
    • pp.281-285
    • /
    • 2005
  • Learning and evolving are two basics for data mining. As compared with classical learning theory based on objective function with minimizing training errors, the recently evolutionary computing has had an efficient approach for constructing optimal model without the minimizing training errors. The global search of evolutionary computing in solution space can settle the local optima problems of learning models. In this research, combining co-evolving algorithm into statistical learning theory, we propose an co-evolutionary computing for statistical learning theory for overcoming local optima problems of statistical learning theory. We apply proposed model to classification and prediction problems of the learning. In the experimental results, we verify the improved performance of our model using the data sets from UCI machine learning repository and KDD Cup 2000.