• Title/Summary/Keyword: dimension reduction method

Search Result 250, Processing Time 0.025 seconds

Forecasting Electric Power Demand Using Census Information and Electric Power Load (센서스 정보 및 전력 부하를 활용한 전력 수요 예측)

  • Lee, Heon Gyu;Shin, Yong Ho
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.18 no.3
    • /
    • pp.35-46
    • /
    • 2013
  • In order to develop an accurate analytical model for domestic electricity demand forecasting, we propose a prediction method of the electric power demand pattern by combining SMO classification techniques and a dimension reduction conceptualized subspace clustering techniques suitable for high-dimensional data cluster analysis. In terms of electricity demand pattern prediction, hourly electricity load patterns and the demographic and geographic characteristics can be analyzed by integrating the wireless load monitoring data as well as sub-regional unit of census information. There are composed of a total of 18 characteristics clusters in the prediction result for the sub-regional demand pattern by using census information and power load of Seoul metropolitan area. The power demand pattern prediction accuracy was approximately 85%.

Monte Carlo analysis of the induced cracked zone by single-hole rock explosion

  • Shadabfar, Mahdi;Huang, Hongwei;Wang, Yuan;Wu, Chenglong
    • Geomechanics and Engineering
    • /
    • v.21 no.3
    • /
    • pp.289-300
    • /
    • 2020
  • Estimating the damage induced by an explosion around a blast hole has always been a challenging issue in geotechnical engineering. It is difficult to determine an exact dimension for damage zone since many parameters are involved in the formation of failures, and there are some uncertainties lying in these parameters. Thus, the present study adopted a probabilistic approach towards this problem. First, a reliability model of the problem was established and the failure probability of induced damage was calculated. Then, the corresponding exceedance risk curve was developed indicating the relation between the failure probability and the cracked zone radius. The obtained risk curve indicated that the failure probability drops dramatically by increasing the cracked zone radius so that the probability of exceedance for any crack length greater than 4.5 m is less than 5%. Moreover, the effect of each parameter involved in the probability of failure, including blast hole radius, explosive density, detonation velocity, and tensile strength of the rock, was evaluated by using a sensitivity analysis. Finally, the impact of the decoupling ratio on the reduction of failures was investigated and the location of its maximum influence was demonstrated around the blast point.

Bandwidth Enhanced Miniaturization Method of Parallel Coupled-Line Filter (대역폭 특성이 개선된 평행 결합 선로 필터의 소형화 기법)

  • Myoung, Seong-Sik;Yook, Jong-Gwan
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.18 no.2 s.117
    • /
    • pp.126-135
    • /
    • 2007
  • This paper proposes a new miniaturization method for a parallel coupled line filter with enhanced bandwidth characteristics. A previous method incorporated several advantages, such as size reduction through the use of only a small number of capacitors, in addition to grounding, suppression of harmonic characteristics, and improved skirt characteristics for the parallel coupled line filter, which is conventional in the field of RE filters due to its design and fabrication simplicity. However, the previous method also has disadvantages related to the bandwidth shrinkage of the miniaturized filters. In this paper, the amount of bandwidth shrinkage is analyzed in terms of the relationship between the loaded Q(quality factor) and the group delay of a resonator. Moreover, the reduction in the bandwidth is solved by a design with new design equations. To show the validity of the proposed method, a hairpin filter with a center frequency of 5.2 GHz and an fractional bandwidth(FBW) of 10% was scaled down to half its original dimension by the proposed method with the enhanced bandwidth characteristics. The measured result shows a high level of agreement with theoretical results.

Clickstream Big Data Mining for Demographics based Digital Marketing (인구통계특성 기반 디지털 마케팅을 위한 클릭스트림 빅데이터 마이닝)

  • Park, Jiae;Cho, Yoonho
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.3
    • /
    • pp.143-163
    • /
    • 2016
  • The demographics of Internet users are the most basic and important sources for target marketing or personalized advertisements on the digital marketing channels which include email, mobile, and social media. However, it gradually has become difficult to collect the demographics of Internet users because their activities are anonymous in many cases. Although the marketing department is able to get the demographics using online or offline surveys, these approaches are very expensive, long processes, and likely to include false statements. Clickstream data is the recording an Internet user leaves behind while visiting websites. As the user clicks anywhere in the webpage, the activity is logged in semi-structured website log files. Such data allows us to see what pages users visited, how long they stayed there, how often they visited, when they usually visited, which site they prefer, what keywords they used to find the site, whether they purchased any, and so forth. For such a reason, some researchers tried to guess the demographics of Internet users by using their clickstream data. They derived various independent variables likely to be correlated to the demographics. The variables include search keyword, frequency and intensity for time, day and month, variety of websites visited, text information for web pages visited, etc. The demographic attributes to predict are also diverse according to the paper, and cover gender, age, job, location, income, education, marital status, presence of children. A variety of data mining methods, such as LSA, SVM, decision tree, neural network, logistic regression, and k-nearest neighbors, were used for prediction model building. However, this research has not yet identified which data mining method is appropriate to predict each demographic variable. Moreover, it is required to review independent variables studied so far and combine them as needed, and evaluate them for building the best prediction model. The objective of this study is to choose clickstream attributes mostly likely to be correlated to the demographics from the results of previous research, and then to identify which data mining method is fitting to predict each demographic attribute. Among the demographic attributes, this paper focus on predicting gender, age, marital status, residence, and job. And from the results of previous research, 64 clickstream attributes are applied to predict the demographic attributes. The overall process of predictive model building is compose of 4 steps. In the first step, we create user profiles which include 64 clickstream attributes and 5 demographic attributes. The second step performs the dimension reduction of clickstream variables to solve the curse of dimensionality and overfitting problem. We utilize three approaches which are based on decision tree, PCA, and cluster analysis. We build alternative predictive models for each demographic variable in the third step. SVM, neural network, and logistic regression are used for modeling. The last step evaluates the alternative models in view of model accuracy and selects the best model. For the experiments, we used clickstream data which represents 5 demographics and 16,962,705 online activities for 5,000 Internet users. IBM SPSS Modeler 17.0 was used for our prediction process, and the 5-fold cross validation was conducted to enhance the reliability of our experiments. As the experimental results, we can verify that there are a specific data mining method well-suited for each demographic variable. For example, age prediction is best performed when using the decision tree based dimension reduction and neural network whereas the prediction of gender and marital status is the most accurate by applying SVM without dimension reduction. We conclude that the online behaviors of the Internet users, captured from the clickstream data analysis, could be well used to predict their demographics, thereby being utilized to the digital marketing.

Band Selection Algorithm based on Expected Value for Pixel Classification (픽셀 분류를 위한 기댓값 기반 밴드 선택 알고리즘)

  • Chang, Duhyeuk;Jung, Byeonghyeon;Heo, Junyoung
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.22 no.6
    • /
    • pp.107-112
    • /
    • 2022
  • In an embedded system such as a drone, it is difficult to store, transfer and analyze the entire hyper-spectral image to a server in real time because it takes a lot of power and time. Therefore, the hyper-spectral image data is transmitted to the server through dimension reduction or compression pre-processing. Feature selection method are used to send only the bands for analysis purpose, and these algorithms usually take a lot of processing time depending on the size of the image, even though the efficiency is high. In this paper, by improving the temporal disadvantage of the band selection algorithm, the time taken 24 hours was reduced to around 60-180 seconds based on the 40000*682 image resolution of 8GB data, and the use of 7.6GB RAM was significantly reduced to 2.3GB using 45 out of 150 bands. However, in terms of pixel classification performance, more than 98% of analysis results were derived similarly to the previous one.

Feature selection for text data via sparse principal component analysis (희소주성분분석을 이용한 텍스트데이터의 단어선택)

  • Won Son
    • The Korean Journal of Applied Statistics
    • /
    • v.36 no.6
    • /
    • pp.501-514
    • /
    • 2023
  • When analyzing high dimensional data such as text data, if we input all the variables as explanatory variables, statistical learning procedures may suffer from over-fitting problems. Furthermore, computational efficiency can deteriorate with a large number of variables. Dimensionality reduction techniques such as feature selection or feature extraction are useful for dealing with these problems. The sparse principal component analysis (SPCA) is one of the regularized least squares methods which employs an elastic net-type objective function. The SPCA can be used to remove insignificant principal components and identify important variables from noisy observations. In this study, we propose a dimension reduction procedure for text data based on the SPCA. Applying the proposed procedure to real data, we find that the reduced feature set maintains sufficient information in text data while the size of the feature set is reduced by removing redundant variables. As a result, the proposed procedure can improve classification accuracy and computational efficiency, especially for some classifiers such as the k-nearest neighbors algorithm.

Face Recognitions Using Centroid Shift and Neural Network-based Principal Component Analysis (중심이동과 신경망 기반 주요성분분석을 이용한 얼굴인식)

  • Cho Yong-Hyun
    • The KIPS Transactions:PartB
    • /
    • v.12B no.6 s.102
    • /
    • pp.715-720
    • /
    • 2005
  • This paper presents a hybrid recognition method of first moment of face image and principal component analysis(PCA). First moment is applied to reduce the dimension by shifting to the centroid of image, which is to exclude the needless backgrounds in the face recognitions. PCA is implemented by single layer neural network which has a teaming rule of Foldiak algorithm. It has been used as an alternative method for numerical PCA. PCA is to derive an orthonormal basis which directly leads to dimensionality reduction and possibly to feature extraction of face image. The proposed method has been applied to the problems for recognizing the 48 face images(12 Persons $\ast$ 4 scenes) of 64$\ast$64 pixels. The 3 distances such as city-block, Euclidean, negative angle are used as measures when match the probe images to the nearest gallery images. The experimental results show that the proposed method has a superior recognition performances(speed, rate). The negative angle has been relatively achieved more an accurate similarity than city-block or Euclidean.

A New Image Analysis Method based on Regression Manifold 3-D PCA (회귀 매니폴드 3-D PCA 기반 새로운 이미지 분석 방법)

  • Lee, Kyung-Min;Lin, Chi-Ho
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.22 no.2
    • /
    • pp.103-108
    • /
    • 2022
  • In this paper, we propose a new image analysis method based on regression manifold 3-D PCA. The proposed method is a new image analysis method consisting of a regression analysis algorithm with a structure designed based on an autoencoder capable of nonlinear expansion of manifold 3-D PCA and PCA for efficient dimension reduction when entering large-capacity image data. With the configuration of an autoencoder, a regression manifold 3-DPCA, which derives the best hyperplane through three-dimensional rotation of image pixel values, and a Bayesian rule structure similar to a deep learning structure, are applied. Experiments are performed to verify performance. The image is improved by utilizing the fine dust image, and accuracy performance evaluation is performed through the classification model. As a result, it can be confirmed that it is effective for deep learning performance.

Adaptive quantization for effective data-rate reduction in ultrafast ultrasound imaging (초고속 초음파 영상의 효과적인 데이터율 저감을 위한 적응 양자화)

  • Doyoung Jang;Heechul Yoon
    • The Journal of the Acoustical Society of Korea
    • /
    • v.42 no.5
    • /
    • pp.422-428
    • /
    • 2023
  • Ultrafast ultrasound imaging has been applied to various imaging approaches, including shear wave elastography, ultrafast Doppler, and super-resolution imaging. However, these methods are still challenging in real-time implementation for three Dimension (3D) or portable applications because of their massive data rate required. In this paper, we proposed an adaptive quantization method that effectively reduces the data rate of large Radio Frequency (RF) data. In soft tissue, ultrasound backscatter signals require a high dynamic range, and thus typical quantization used in the current systems uses the quantization level of 10 bits to 14 bits. To alleviate the quantization level to expand the application of ultrafast ultrasound imaging, this study proposed a depth-sectional quantization approach that reduces the quantization errors. For quantitative evaluation, Field II simulations, phantom experiments, and in vivo imaging were conducted and CNR, spatial resolution, and SSIM values were compared with the proposed method and fixed quantization method. We demonstrated that our proposed method is capable of effectively reducing the quantization level down to 3-bit while minimizing the image quality degradation.

Performance Improvement of Automatic Basal Cell Carcinoma Detection Using Half Hanning Window (Half Hanning 윈도우 전처리를 통한 기저 세포암 자동 검출 성능 개선)

  • Park, Aa-Ron;Baek, Seong-Joong;Min, So-Hee;You, Hong-Yoen;Kim, Jin-Young;Hong, Sung-Hoon
    • The Journal of the Korea Contents Association
    • /
    • v.6 no.12
    • /
    • pp.105-112
    • /
    • 2006
  • In this study, we propose a simple preprocessing method for classification of basal cell carcinoma (BCC), which is one of the most common skin cancer. The preprocessing step consists of data clipping with a half Hanning window and dimension reduction with principal components analysis (PCA). The application of the half Hanning window deemphasizes the peak near $1650cm^{-1}$ and improves classification performance by lowering the false negative ratio. Classification results with various classifiers are presented to show the effectiveness of the proposed method. The classifiers include maximum a posteriori probability (MAP), k-nearest neighbor (KNN), probabilistic neural network (PNN), multilayer perceptron(MLP), support vector machine (SVM) and minimum squared error (MSE) classification. Classification results with KNN involving 216 spectra preprocessed with the proposed method gave 97.3% sensitivity, which is very promising results for automatic BCC detection.

  • PDF