• Title/Summary/Keyword: Feature selection algorithm

Search Result 336, Processing Time 0.028 seconds

Digital Watermarking Technique in Wavelet Domain for Protecting Copyright of Contents (컨텐츠의 저작권 보호를 위한 DWT영역에서의 디지털 워터마킹 기법)

  • Seo, Young-Ho;Choi, Hyun-Jun;Kim, Dong-Wook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.14 no.6
    • /
    • pp.1409-1415
    • /
    • 2010
  • In this paper we proposed the watermarking technique using the markspace which is selected by tree-structure between the subbands in the wavelet domain and feature information in the spatial domain. The watermarking candidate region in the wavelet domain is obtained by the markspace selection algorithm divides the highest frequency subband to several segments and calculates theirs energy and the averages value of the total energy of the subband. Also the markspace of the spatial domain is obtained by the boundary information of a image. The final markspace is selected by the markspaces of the wavelet and spatial domain. The watermark is embedded into the selected markspace using the random addresses by LFSR. Finally the watermarking image is generated using the inverse wavelet transform. The proposed watermarking algorithm shows the robustness against the attacks such as JPEG, blurring, sharpening, and gaussian noise.

An Adaptive Method For Face Recognition Based Filters and Selection of Features (필터 및 특징 선택 기반의 적응형 얼굴 인식 방법)

  • Cho, Byoung-Mo;Kim, Gi-Han;Rhee, Phill-Kyu
    • The Journal of the Korea Contents Association
    • /
    • v.9 no.6
    • /
    • pp.1-8
    • /
    • 2009
  • There are a lot of influences, such as location of camera, luminosity, brightness, and direction of light, which affect the performance of 2-dimensional image recognition. This paper suggests an adaptive method for face-image recognition in noisy environments using evolvable filtering and feature extraction which uses one sample image from camera. This suggested method consists of two main parts. One is the environmental-adjustment module which determines optimum sets of filters, filter parameters, and dimensions of features by using "steady state genetic algorithm". The other another part is for face recognition module which performs recognition of face-image using the previous results. In the processing, we used Gabor wavelet for extracting features in the images and k-Nearest Neighbor method for the classification. For testing of the adaptive face recognition method, we tested the adaptive method in the brightness noise, in the impulse noise and in the composite noise and verified that the adaptive method protects face recognition-rate's rapidly decrease which can be occurred generally in the noisy environments.

Comparison of Stock Price Prediction Using Time Series and Non-Time Series Data

  • Min-Seob Song;Junghye Min
    • Journal of the Korea Society of Computer and Information
    • /
    • v.28 no.8
    • /
    • pp.67-75
    • /
    • 2023
  • Stock price prediction is an important topic extensively discussed in the financial market, but it is considered a challenging subject due to numerous factors that can influence it. In this research, performance was compared and analyzed by applying time series prediction models (LSTM, GRU) and non-time series prediction models (RF, SVR, KNN, LGBM) that do not take into account the temporal dependence of data into stock price prediction. In addition, various data such as stock price data, technical indicators, financial statements indicators, buy sell indicators, short selling, and foreign indicators were combined to find optimal predictors and analyze major factors affecting stock price prediction by industry. Through the hyperparameter optimization process, the process of improving the prediction performance for each algorithm was also conducted to analyze the factors affecting the performance. As a result of feature selection and hyperparameter optimization, it was found that the forecast accuracy of the time series prediction algorithm GRU and LSTM+GRU was the highest.

Investigating Dynamic Mutation Process of Issues Using Unstructured Text Analysis (부도예측을 위한 KNN 앙상블 모형의 동시 최적화)

  • Min, Sung-Hwan
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.1
    • /
    • pp.139-157
    • /
    • 2016
  • Bankruptcy involves considerable costs, so it can have significant effects on a country's economy. Thus, bankruptcy prediction is an important issue. Over the past several decades, many researchers have addressed topics associated with bankruptcy prediction. Early research on bankruptcy prediction employed conventional statistical methods such as univariate analysis, discriminant analysis, multiple regression, and logistic regression. Later on, many studies began utilizing artificial intelligence techniques such as inductive learning, neural networks, and case-based reasoning. Currently, ensemble models are being utilized to enhance the accuracy of bankruptcy prediction. Ensemble classification involves combining multiple classifiers to obtain more accurate predictions than those obtained using individual models. Ensemble learning techniques are known to be very useful for improving the generalization ability of the classifier. Base classifiers in the ensemble must be as accurate and diverse as possible in order to enhance the generalization ability of an ensemble model. Commonly used methods for constructing ensemble classifiers include bagging, boosting, and random subspace. The random subspace method selects a random feature subset for each classifier from the original feature space to diversify the base classifiers of an ensemble. Each ensemble member is trained by a randomly chosen feature subspace from the original feature set, and predictions from each ensemble member are combined by an aggregation method. The k-nearest neighbors (KNN) classifier is robust with respect to variations in the dataset but is very sensitive to changes in the feature space. For this reason, KNN is a good classifier for the random subspace method. The KNN random subspace ensemble model has been shown to be very effective for improving an individual KNN model. The k parameter of KNN base classifiers and selected feature subsets for base classifiers play an important role in determining the performance of the KNN ensemble model. However, few studies have focused on optimizing the k parameter and feature subsets of base classifiers in the ensemble. This study proposed a new ensemble method that improves upon the performance KNN ensemble model by optimizing both k parameters and feature subsets of base classifiers. A genetic algorithm was used to optimize the KNN ensemble model and improve the prediction accuracy of the ensemble model. The proposed model was applied to a bankruptcy prediction problem by using a real dataset from Korean companies. The research data included 1800 externally non-audited firms that filed for bankruptcy (900 cases) or non-bankruptcy (900 cases). Initially, the dataset consisted of 134 financial ratios. Prior to the experiments, 75 financial ratios were selected based on an independent sample t-test of each financial ratio as an input variable and bankruptcy or non-bankruptcy as an output variable. Of these, 24 financial ratios were selected by using a logistic regression backward feature selection method. The complete dataset was separated into two parts: training and validation. The training dataset was further divided into two portions: one for the training model and the other to avoid overfitting. The prediction accuracy against this dataset was used to determine the fitness value in order to avoid overfitting. The validation dataset was used to evaluate the effectiveness of the final model. A 10-fold cross-validation was implemented to compare the performances of the proposed model and other models. To evaluate the effectiveness of the proposed model, the classification accuracy of the proposed model was compared with that of other models. The Q-statistic values and average classification accuracies of base classifiers were investigated. The experimental results showed that the proposed model outperformed other models, such as the single model and random subspace ensemble model.

An Effective Extraction Algorithm of Pulmonary Regions Using Intensity-level Maps in Chest X-ray Images (흉부 X-ray 영상에서의 명암 레벨지도를 이용한 효과적인 폐 영역 추출 알고리즘)

  • Jang, Geun-Ho;Park, Ho-Hyun;Lee, Seok-Lyong;Kim, Deok-Hwan;Lim, Myung-Kwan
    • Journal of Korea Multimedia Society
    • /
    • v.13 no.7
    • /
    • pp.1062-1075
    • /
    • 2010
  • In the medical image application the difference of intensity is widely used for the image segmentation and feature extraction, and a well known method is the threshold technique that determines a threshold value and generates a binary image based on the threshold. A frequently-used threshold technique is the Otsu algorithm that provides efficient processing and effective selection criterion for choosing the threshold value. However, we cannot get good segmentation results by applying the Otsu algorithm to chest X-ray images. It is because there are various organic structures around lung regions such as ribs and blood vessels, causing unclear distribution of intensity levels. To overcome the ambiguity, we propose in this paper an effective algorithm to extract pulmonary regions that utilizes the Otsu algorithm after removing the background of an X-ray image, constructs intensity-level maps, and uses them for segmenting the X-ray image. To verify the effectiveness of our method, we compared it with the existing 1-dimensional and 2-dimensional Otsu algorithms, and also the results by expert's naked eyes. The experimental result showed that our method achieved the more accurate extraction of pulmonary regions compared to the Otsu methods and showed the similar result as the naked eye's one.

The Design of Feature Selection Classifier based on Physiological Signal for Emotion Detection (감성판별을 위한 생체신호기반 특징선택 분류기 설계)

  • Lee, JeeEun;Yoo, Sun K.
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.11
    • /
    • pp.206-216
    • /
    • 2013
  • The emotion plays a critical role in human's daily life including learning, action, decision and communication. In this paper, emotion discrimination classifier is designed to reduce system complexity through reduced selection of dominant features from biosignals. The photoplethysmography(PPG), skin temperature, skin conductance, fontal and parietal electroencephalography(EEG) signals were measured during 4 types of movie watching associated with the induction of neutral, sad, fear joy emotions. The genetic algorithm with support vector machine(SVM) based fitness function was designed to determine dominant features among 24 parameters extracted from measured biosignals. It shows maximum classification accuracy of 96.4%, which is 17% higher than that of SVM alone. The minimum error features selected are the mean and NN50 of heart rate variability from PPG signal, the mean of PPG induced pulse transit time, the mean of skin resistance, and ${\delta}$ and ${\beta}$ frequency band powers of parietal EEG. The combination of parietal EEG, PPG, and skin resistance is recommendable in high accuracy instrumentation, while the combinational use of PPG and skin conductance(79% accuracy) is affordable in simplified instrumentation.

State of the Art Technology Trends and Case Analysis of Leading Research in Harmony Search Algorithm (하모니 탐색 알고리즘의 선도 연구에 관한 최첨단 기술 동향과 사례 분석)

  • Kim, Eun-Sung;Shin, Seung-Soo;Kim, Yong-Hyuk;Yoon, Yourim
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.11
    • /
    • pp.81-90
    • /
    • 2021
  • There are various optimization problems in real world and research continues to solve them. An optimization problem is the problem of finding a combination of parameters that maximizes or minimizes the objective function. Harmony search is a population-based metaheuristic algorithm for solving optimization problems and it is designed to mimic the improvisation of jazz music. Harmony search has been actively applied to optimization problems in various fields such as civil engineering, computer science, energy, medical science, and water quality engineering. Harmony search has a simple working principle and it has the advantage of finding good solutions quickly in constrained optimization problems. Especially there are various application cases showing high accuracy with a low number of iterations by improving the solution through the empirical derivative. In this paper, we explain working principle of Harmony search and classify the leading research in recent 3 years, review them according to category, and suggest future research directions. The research is divided into review by field, algorithmic analysis and theory, and application to real world problems. Application to real world problems is classified according to the purpose of optimization and whether or not they are hybridized with other metaheuristic algorithms.

Extraction of Classification Boundary for Fuzzy Partitions and Its Application to Pattern Classification (퍼지 분할을 위한 분류 경계의 추출과 패턴 분류에의 응용)

  • Son, Chang-S.;Seo, Suk-T.;Chung, Hwan-M.;Kwon, Soon-H.
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.18 no.5
    • /
    • pp.685-691
    • /
    • 2008
  • The selection of classification boundaries in fuzzy rule- based classification systems is an important and difficult problem. So various methods based on learning processes such as neural network, genetic algorithm, and so on have been proposed for it. In a previous study, we pointed out the limitation of the methods and discussed a method for fuzzy partitioning in the overlapped region on feature space in order to overcome the time-consuming when the additional parameters for tuning fuzzy membership functions are necessary. In this paper, we propose a method to determine three types of classification boundaries(i.e., non-overlapping, overlapping, and a boundary point) on the basis of statistical information of the given dataset without learning by extending the method described in the study. Finally, we show the effectiveness of the proposed method through experimental results applied to pattern classification problems using the modified IRIS and standard IRIS datasets.

Mosaic image generation of AISA Eagle hyperspectral sensor using SIFT method (SIFT 기법을 이용한 AISA Eagle 초분광센서의 모자이크영상 생성)

  • Han, You Kyung;Kim, Yong Il;Han, Dong Yeob;Choi, Jae Wan
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.31 no.2
    • /
    • pp.165-172
    • /
    • 2013
  • In this paper, high-quality mosaic image is generated by high-resolution hyperspectral strip images using scale-invariant feature transform (SIFT) algorithm, which is one of the representative image matching methods. The experiments are applied to AISA Eagle images geo-referenced by using GPS/INS information acquired when it was taken on flight. The matching points between three strips of hyperspectral images are extracted using SIFT method, and the transformation models between images are constructed from the points. Mosaic image is, then, generated using the transformation models constructed from corresponding images. Optimal band appropriate for the matching point extraction is determined by selecting representative bands of hyperspectral data and analyzing the matched results based on each band. Mosaic image generated by proposed method is visually compared with the mosaic image generated from initial geo-referenced AISA hyperspectral images. From the comparison, we could estimate geometrical accuracy of generated mosaic image and analyze the efficiency of our methodology.

An Efficient Multidimensional Scaling Method based on CUDA and Divide-and-Conquer (CUDA 및 분할-정복 기반의 효율적인 다차원 척도법)

  • Park, Sung-In;Hwang, Kyu-Baek
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.4
    • /
    • pp.427-431
    • /
    • 2010
  • Multidimensional scaling (MDS) is a widely used method for dimensionality reduction, of which purpose is to represent high-dimensional data in a low-dimensional space while preserving distances among objects as much as possible. MDS has mainly been applied to data visualization and feature selection. Among various MDS methods, the classical MDS is not readily applicable to data which has large numbers of objects, on normal desktop computers due to its computational complexity. More precisely, it needs to solve eigenpair problems on dissimilarity matrices based on Euclidean distance. Thus, running time and required memory of the classical MDS highly increase as n (the number of objects) grows up, restricting its use in large-scale domains. In this paper, we propose an efficient approximation algorithm for the classical MDS based on divide-and-conquer and CUDA. Through a set of experiments, we show that our approach is highly efficient and effective for analysis and visualization of data consisting of several thousands of objects.