• Title/Summary/Keyword: Hybrid Algorithms

Search Result 586, Processing Time 0.025 seconds

The Analysis and Design of Advanced Neurofuzzy Polynomial Networks (고급 뉴로퍼지 다항식 네트워크의 해석과 설계)

  • Park, Byeong-Jun;O, Seong-Gwon
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.39 no.3
    • /
    • pp.18-31
    • /
    • 2002
  • In this study, we introduce a concept of advanced neurofuzzy polynomial networks(ANFPN), a hybrid modeling architecture combining neurofuzzy networks(NFN) and polynomial neural networks(PNN). These networks are highly nonlinear rule-based models. The development of the ANFPN dwells on the technologies of Computational Intelligence(Cl), namely fuzzy sets, neural networks and genetic algorithms. NFN contributes to the formation of the premise part of the rule-based structure of the ANFPN. The consequence part of the ANFPN is designed using PNN. At the premise part of the ANFPN, NFN uses both the simplified fuzzy inference and error back-propagation learning rule. The parameters of the membership functions, learning rates and momentum coefficients are adjusted with the use of genetic optimization. As the consequence structure of ANFPN, PNN is a flexible network architecture whose structure(topology) is developed through learning. In particular, the number of layers and nodes of the PNN are not fixed in advance but is generated in a dynamic way. In this study, we introduce two kinds of ANFPN architectures, namely the basic and the modified one. Here the basic and the modified architecture depend on the number of input variables and the order of polynomial in each layer of PNN structure. Owing to the specific features of two combined architectures, it is possible to consider the nonlinear characteristics of process system and to obtain the better output performance with superb predictive ability. The availability and feasibility of the ANFPN are discussed and illustrated with the aid of two representative numerical examples. The results show that the proposed ANFPN can produce the model with higher accuracy and predictive ability than any other method presented previously.

Bitmap Indexes and Query Processing Strategies for Relational XML Twig Queries (관계형 XML 가지 패턴 질의를 위한 비트맵 인덱스와 질의 처리 기법)

  • Lee, Kyong-Ha;Moon, Bong-Ki;Lee, Kyu-Chul
    • Journal of KIISE:Databases
    • /
    • v.37 no.3
    • /
    • pp.146-164
    • /
    • 2010
  • Due to an increasing volume of XML data, it is considered prudent to store XML data on an industry-strength database system instead of relying on a domain specific application or a file system. For shredded XML data stored in relational tables, however, it may not be straightforward to apply existing algorithms for twig query processing, since most of the algorithms require XML data to be accessed in a form of streams of elements grouped by their tags and sorted in a particular order. In order to support XML query processing within the common framework of relational database systems, we first propose several bitmap indexes and their strategies for supporting holistic twig joining on XML data stored in relational tables. Since bitmap indexes are well supported in most of the commercial and open-source database systems, the proposed bitmapped indexes and twig query processing strategies can be incorporated into relational query processing framework with more ease. The proposed query processing strategies are efficient in terms of both time and space, because the compressed bitmap indexes stay compressed during data access. In addition, we propose a hybrid index which computes twig query solutions with only bit-vectors, without accessing labeled XML elements stored in the relational tables.

A Study on the Improvement of Injection Molding Process Using CAE and Decision-tree (CAE와 Decision-tree를 이용한 사출성형 공정개선에 관한 연구)

  • Hwang, Soonhwan;Han, Seong-Ryeol;Lee, Hoojin
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.22 no.4
    • /
    • pp.580-586
    • /
    • 2021
  • The CAT methodology is a numerical analysis technique using CAE. Recently, a methodology of applying artificial intelligence techniques to a simulation has been studied. A previous study compared the deformation results according to the injection molding process using a machine learning technique. Although MLP has excellent prediction performance, it lacks an explanation of the decision process and is like a black box. In this study, data was generated using Autodesk Moldflow 2018, an injection molding analysis software. Several Machine Learning Algorithms models were developed using RapidMiner version 9.5, a machine learning platform software, and the root mean square error was compared. The decision-tree showed better prediction performance than other machine learning techniques with the RMSE values. The classification criterion can be increased according to the Maximal Depth that determines the size of the Decision-tree, but the complexity also increases. The simulation showed that by selecting an intermediate value that satisfies the constraint based on the changed position, there was 7.7% improvement compared to the previous simulation.

White striping degree assessment using computer vision system and consumer acceptance test

  • Kato, Talita;Mastelini, Saulo Martiello;Campos, Gabriel Fillipe Centini;Barbon, Ana Paula Ayub da Costa;Prudencio, Sandra Helena;Shimokomaki, Massami;Soares, Adriana Lourenco;Barbon, Sylvio Jr.
    • Asian-Australasian Journal of Animal Sciences
    • /
    • v.32 no.7
    • /
    • pp.1015-1026
    • /
    • 2019
  • Objective: The objective of this study was to evaluate three different degrees of white striping (WS) addressing their automatic assessment and customer acceptance. The WS classification was performed based on a computer vision system (CVS), exploring different machine learning (ML) algorithms and the most important image features. Moreover, it was verified by consumer acceptance and purchase intent. Methods: The samples for image analysis were classified by trained specialists, according to severity degrees regarding visual and firmness aspects. Samples were obtained with a digital camera, and 25 features were extracted from these images. ML algorithms were applied aiming to induce a model capable of classifying the samples into three severity degrees. In addition, two sensory analyses were performed: 75 samples properly grilled were used for the first sensory test, and 9 photos for the second. All tests were performed using a 10-cm hybrid hedonic scale (acceptance test) and a 5-point scale (purchase intention). Results: The information gain metric ranked 13 attributes. However, just one type of image feature was not enough to describe the phenomenon. The classification models support vector machine, fuzzy-W, and random forest showed the best results with similar general accuracy (86.4%). The worst performance was obtained by multilayer perceptron (70.9%) with the high error rate in normal (NORM) sample predictions. The sensory analysis of acceptance verified that WS myopathy negatively affects the texture of the broiler breast fillets when grilled and the appearance attribute of the raw samples, which influenced the purchase intention scores of raw samples. Conclusion: The proposed system has proved to be adequate (fast and accurate) for the classification of WS samples. The sensory analysis of acceptance showed that WS myopathy negatively affects the tenderness of the broiler breast fillets when grilled, while the appearance attribute of the raw samples eventually influenced purchase intentions.

Improved Resource Allocation Model for Reducing Interference among Secondary Users in TV White Space for Broadband Services

  • Marco P. Mwaimu;Mike Majham;Ronoh Kennedy;Kisangiri Michael;Ramadhani Sinde
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.4
    • /
    • pp.55-68
    • /
    • 2023
  • In recent years, the Television White Space (TVWS) has attracted the interest of many researchers due to its propagation characteristics obtainable between 470MHz and 790MHz spectrum bands. The plenty of unused channels in the TV spectrum allows the secondary users (SUs) to use the channels for broadband services especially in rural areas. However, when the number of SUs increases in the TVWS wireless network the aggregate interference also increases. Aggregate interferences are the combined harmful interferences that can include both co-channel and adjacent interferences. The aggregate interference on the side of Primary Users (PUs) has been extensively scrutinized. Therefore, resource allocation (power and spectrum) is crucial when designing the TVWS network to avoid interferences from Secondary Users (SUs) to PUs and among SUs themselves. This paper proposes a model to improve the resource allocation for reducing the aggregate interface among SUs for broadband services in rural areas. The proposed model uses joint power and spectrum hybrid Firefly algorithm (FA), Genetic algorithm (GA), and Particle Swarm Optimization algorithm (PSO) which is considered the Co-channel interference (CCI) and Adjacent Channel Interference (ACI). The algorithm is integrated with the admission control algorithm so that; there is a possibility to remove some of the SUs in the TVWS network whenever the SINR threshold for SUs and PU are not met. We considered the infeasible system whereby all SUs and PU may not be supported simultaneously. Therefore, we proposed a joint spectrum and power allocation with an admission control algorithm whose better complexity and performance than the ones which have been proposed in the existing algorithms in the literature. The performance of the proposed algorithm is compared using the metrics such as sum throughput, PU SINR, algorithm running time and SU SINR less than threshold and the results show that the PSOFAGA with ELGR admission control algorithm has best performance compared to GA, PSO, FA, and FAGAPSO algorithms.

Development and application of prediction model of hyperlipidemia using SVM and meta-learning algorithm (SVM과 meta-learning algorithm을 이용한 고지혈증 유병 예측모형 개발과 활용)

  • Lee, Seulki;Shin, Taeksoo
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.111-124
    • /
    • 2018
  • This study aims to develop a classification model for predicting the occurrence of hyperlipidemia, one of the chronic diseases. Prior studies applying data mining techniques for predicting disease can be classified into a model design study for predicting cardiovascular disease and a study comparing disease prediction research results. In the case of foreign literatures, studies predicting cardiovascular disease were predominant in predicting disease using data mining techniques. Although domestic studies were not much different from those of foreign countries, studies focusing on hypertension and diabetes were mainly conducted. Since hypertension and diabetes as well as chronic diseases, hyperlipidemia, are also of high importance, this study selected hyperlipidemia as the disease to be analyzed. We also developed a model for predicting hyperlipidemia using SVM and meta learning algorithms, which are already known to have excellent predictive power. In order to achieve the purpose of this study, we used data set from Korea Health Panel 2012. The Korean Health Panel produces basic data on the level of health expenditure, health level and health behavior, and has conducted an annual survey since 2008. In this study, 1,088 patients with hyperlipidemia were randomly selected from the hospitalized, outpatient, emergency, and chronic disease data of the Korean Health Panel in 2012, and 1,088 nonpatients were also randomly extracted. A total of 2,176 people were selected for the study. Three methods were used to select input variables for predicting hyperlipidemia. First, stepwise method was performed using logistic regression. Among the 17 variables, the categorical variables(except for length of smoking) are expressed as dummy variables, which are assumed to be separate variables on the basis of the reference group, and these variables were analyzed. Six variables (age, BMI, education level, marital status, smoking status, gender) excluding income level and smoking period were selected based on significance level 0.1. Second, C4.5 as a decision tree algorithm is used. The significant input variables were age, smoking status, and education level. Finally, C4.5 as a decision tree algorithm is used. In SVM, the input variables selected by genetic algorithms consisted of 6 variables such as age, marital status, education level, economic activity, smoking period, and physical activity status, and the input variables selected by genetic algorithms in artificial neural network consist of 3 variables such as age, marital status, and education level. Based on the selected parameters, we compared SVM, meta learning algorithm and other prediction models for hyperlipidemia patients, and compared the classification performances using TP rate and precision. The main results of the analysis are as follows. First, the accuracy of the SVM was 88.4% and the accuracy of the artificial neural network was 86.7%. Second, the accuracy of classification models using the selected input variables through stepwise method was slightly higher than that of classification models using the whole variables. Third, the precision of artificial neural network was higher than that of SVM when only three variables as input variables were selected by decision trees. As a result of classification models based on the input variables selected through the genetic algorithm, classification accuracy of SVM was 88.5% and that of artificial neural network was 87.9%. Finally, this study indicated that stacking as the meta learning algorithm proposed in this study, has the best performance when it uses the predicted outputs of SVM and MLP as input variables of SVM, which is a meta classifier. The purpose of this study was to predict hyperlipidemia, one of the representative chronic diseases. To do this, we used SVM and meta-learning algorithms, which is known to have high accuracy. As a result, the accuracy of classification of hyperlipidemia in the stacking as a meta learner was higher than other meta-learning algorithms. However, the predictive performance of the meta-learning algorithm proposed in this study is the same as that of SVM with the best performance (88.6%) among the single models. The limitations of this study are as follows. First, various variable selection methods were tried, but most variables used in the study were categorical dummy variables. In the case with a large number of categorical variables, the results may be different if continuous variables are used because the model can be better suited to categorical variables such as decision trees than general models such as neural networks. Despite these limitations, this study has significance in predicting hyperlipidemia with hybrid models such as met learning algorithms which have not been studied previously. It can be said that the result of improving the model accuracy by applying various variable selection techniques is meaningful. In addition, it is expected that our proposed model will be effective for the prevention and management of hyperlipidemia.

Combining Model-based and Heuristic Techniques for Fast Tracking the Global Maximum Power Point of a Photovoltaic String

  • Shi, Ji-Ying;Xue, Fei;Ling, Le-Tao;Li, Xiao-Fei;Qin, Zi-Jian;Li, Ya-Jing;Yang, Ting
    • Journal of Power Electronics
    • /
    • v.17 no.2
    • /
    • pp.476-489
    • /
    • 2017
  • Under partial shading conditions (PSCs), multiple maximums may be exhibited on the P-U curve of string inverter photovoltaic (PV) systems. Under such conditions, heuristic methods are invalid for extracting a global maximum power point (GMPP); intelligent algorithms are time-consuming; and model-based methods are complex and costly. To overcome these shortcomings, a novel hybrid MPPT (MPF-IP&O) based on a model-based peak forecasting (MPF) method and an improved perturbation and observation (IP&O) method is proposed. The MPF considers the influence of temperature and does not require solar radiation measurements. In addition, it can forecast all of the peak values of the PV string without complex computation under PSCs, and it can determine the candidate GMPP after a comparison. Hence, the MPF narrows the searching range tremendously and accelerates the convergence to the GMPP. Additionally, the IP&O with a successive approximation strategy searches for the real GMPP in the neighborhood of the candidate one, which can significantly enhance the tracking efficiency. Finally, simulation and experiment results show that the proposed method has a higher tracking speed and accuracy than the perturbation and observation (P&O) and particle swarm optimization (PSO) methods under PSCs.

An Intelligent Framework for Test Case Prioritization Using Evolutionary Algorithm

  • Dobuneh, Mojtaba Raeisi Nejad;Jawawi, Dayang N.A.
    • Journal of Internet Computing and Services
    • /
    • v.17 no.5
    • /
    • pp.89-95
    • /
    • 2016
  • In a software testing domain, test case prioritization techniques improve the performance of regression testing, and arrange test cases in such a way that maximum available faults be detected in a shorter time. User-sessions and cookies are unique features of web applications that are useful in regression testing because they have precious information about the application state before and after making changes to software code. This approach is in fact a user-session based technique. The user session will collect from the database on the server side, and test cases are released by the small change configuration of a user session data. The main challenges are the effectiveness of Average Percentage Fault Detection rate (APFD) and time constraint in the existing techniques, so in this paper developed an intelligent framework which has three new techniques use to manage and put test cases in group by applying useful criteria for test case prioritization in web application regression testing. In dynamic weighting approach the hybrid criteria which set the initial weight to each criterion determines optimal weight of combination criteria by evolutionary algorithms. The weight of each criterion is based on the effectiveness of finding faults in the application. In this research the priority is given to test cases that are performed based on most common http requests in pages, the length of http request chains, and the dependency of http requests. To verify the new technique some fault has been seeded in subject application, then applying the prioritization criteria on test cases for comparing the effectiveness of APFD rate with existing techniques.

Development and Usability Testing of a User-Centered 3D Virtual Liver Surgery Planning System

  • Yang, Xiaopeng;Yu, Hee Chul;Choi, Younggeun;Yang, Jae Do;Cho, Baik Hwan;You, Heecheon
    • Journal of the Ergonomics Society of Korea
    • /
    • v.36 no.1
    • /
    • pp.37-52
    • /
    • 2017
  • Objective: The present study developed a user-centered 3D virtual liver surgery planning (VLSP) system called Dr. Liver to provide preoperative information for safe and rational surgery. Background: Preoperative 3D VLSP is needed for patients' safety in liver surgery. Existing systems either do not provide functions specialized for liver surgery planning or do not provide functions for cross-check of the accuracy of analysis results. Method: Use scenarios of Dr. Liver were developed through literature review, benchmarking, and interviews with surgeons. User interfaces of Dr. Liver with various user-friendly features (e.g., context-sensitive hotkey menu and 3D view navigation box) was designed. Novel image processing algorithms (e.g., hybrid semi-automatic algorithm for liver extraction and customized region growing algorithm for vessel extraction) were developed for accurate and efficient liver surgery planning. Usability problems of a preliminary version of Dr. Liver were identified by surgeons and system developers and then design changes were made to resolve the identified usability problems. Results: A usability testing showed that the revised version of Dr. Liver achieved a high level of satisfaction ($6.1{\pm}0.8$ out of 7) and an acceptable time efficiency ($26.7{\pm}0.9 min$) in liver surgery planning. Conclusion: Involvement of usability testing in system development process from the beginning is useful to identify potential usability problems to improve for shortening system development period and cost. Application: The development and evaluation process of Dr. Liver in this study can be referred in designing a user-centered system.

A Study on Hybrid Split-Spectrum Processing Technique for Enhanced Reliability in Ultrasonic Signal Analysis (초음파 신호 해석의 신뢰도 개선을 위한 하이브리드 스플릿-스펙트럼 신호 처리 기술에 관한 연구)

  • Huh, H.;Koo, K.M.;Kim, G.J.
    • Journal of the Korean Society for Nondestructive Testing
    • /
    • v.16 no.1
    • /
    • pp.1-9
    • /
    • 1996
  • Many signal-processing techniques have been found to be useful in ultrasonic and nondestructive evaluation. Among the most popular techniques are signal averaging, spatial compounding, matched filters and homomorphic processing. One of the significant new process is split-spectrum processing(SSP), which can be equally useful in signal-to-noise ratio(SNR) improvement and grain characterization in several specimens. The purpose of this paper is to explore the utility of SSP in ultrasonic NDE. A wide variety of engineering problems are reviewed, and suggestions for implementation of the technique are provided. SSP uses the frequency-dependent response of the interfering coherent noise produced by unresolvable scatters in the resolution range cell of a transducer. It is implemented by splitting the frequency spectrum of the received signal by using gaussian bandpass filter. The theoretical basis for the potential of SSP for grain characterization in SUS 304 material is discussed, and some experimental evidence for the feasibility of the approach is presented. Results of SNR enhancement in signals obtained from real four samples of SUS 304. The influence of various processing parameters on the performance of the processing technique is also discussed. The minimization algorithm, which provides an excellent SNR enhancement when used either in conjunction with other SSP algorithms like polarity-check or by itself, is also presented.

  • PDF