• Title/Summary/Keyword: Matrix factorization

Search Result 309, Processing Time 0.019 seconds

A Study on the Sparse Matrix Method Useful to the Solution of a Large Power System (전력계통 해석에 유용한 "스파스"행렬법에 관한 연구)

  • 한만춘;신명철
    • 전기의세계
    • /
    • v.23 no.3
    • /
    • pp.43-52
    • /
    • 1974
  • The matrix inversion is very inefficient for computing direct solutions of the large spare systems of linear equations that arise in many network problems as a large electrical power system. Optimally ordered triangular factorization of sparse matrices is more efficient and offers the other important computational advantages in some applications with this method. The direct solutions are computed from sparse matrix factors instead of a full inverse matrix, thereby gaining a significant advantage is speed and computer memory requirements. In this paper, it is shown that the sparse matrix method is superior to the inverse matrix method to solve the linear equations of large sparse networks. In addition, it is shown that the sparse matrix method is superior to the inverse matrix method to solve the linear equations of large sparse networks. In addition, it is shown that the solutions may be applied directly to sove the load flow in an electrical power system. The result of this study should lead to many aplications including short circuit, transient stability, network reduction, reactive optimization and others.

  • PDF

Speed-up of the Matrix Computation on the Ridge Regression

  • Lee, Woochan;Kim, Moonseong;Park, Jaeyoung
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.10
    • /
    • pp.3482-3497
    • /
    • 2021
  • Artificial intelligence has emerged as the core of the 4th industrial revolution, and large amounts of data processing, such as big data technology and rapid data analysis, are inevitable. The most fundamental and universal data interpretation technique is an analysis of information through regression, which is also the basis of machine learning. Ridge regression is a technique of regression that decreases sensitivity to unique or outlier information. The time-consuming calculation portion of the matrix computation, however, basically includes the introduction of an inverse matrix. As the size of the matrix expands, the matrix solution method becomes a major challenge. In this paper, a new algorithm is introduced to enhance the speed of ridge regression estimator calculation through series expansion and computation recycle without adopting an inverse matrix in the calculation process or other factorization methods. In addition, the performances of the proposed algorithm and the existing algorithm were compared according to the matrix size. Overall, excellent speed-up of the proposed algorithm with good accuracy was demonstrated.

Underdetermined blind source separation using normalized spatial covariance matrix and multichannel nonnegative matrix factorization (멀티채널 비음수 행렬분해와 정규화된 공간 공분산 행렬을 이용한 미결정 블라인드 소스 분리)

  • Oh, Son-Mook;Kim, Jung-Han
    • The Journal of the Acoustical Society of Korea
    • /
    • v.39 no.2
    • /
    • pp.120-130
    • /
    • 2020
  • This paper solves the problem in underdetermined convolutive mixture by improving the disadvantages of the multichannel nonnegative matrix factorization technique widely used in blind source separation. In conventional researches based on Spatial Covariance Matrix (SCM), each element composed of values such as power gain of single channel and correlation tends to degrade the quality of the separated sources due to high variance. In this paper, level and frequency normalization is performed to effectively cluster the estimated sources. Therefore, we propose a novel SCM and an effective distance function for cluster pairs. In this paper, the proposed SCM is used for the initialization of the spatial model and used for hierarchical agglomerative clustering in the bottom-up approach. The proposed algorithm was experimented using the 'Signal Separation Evaluation Campaign 2008 development dataset'. As a result, the improvement in most of the performance indicators was confirmed by utilizing the 'Blind Source Separation Eval toolbox', an objective source separation quality verification tool, and especially the performance superiority of the typical SDR of 1 dB to 3.5 dB was verified.

Speech Denoising via Low-Rank and Sparse Matrix Decomposition

  • Huang, Jianjun;Zhang, Xiongwei;Zhang, Yafei;Zou, Xia;Zeng, Li
    • ETRI Journal
    • /
    • v.36 no.1
    • /
    • pp.167-170
    • /
    • 2014
  • In this letter, we propose an unsupervised framework for speech noise reduction based on the recent development of low-rank and sparse matrix decomposition. The proposed framework directly separates the speech signal from noisy speech by decomposing the noisy speech spectrogram into three submatrices: the noise structure matrix, the clean speech structure matrix, and the residual noise matrix. Evaluations on the Noisex-92 dataset show that the proposed method achieves a signal-to-distortion ratio approximately 2.48 dB and 3.23 dB higher than that of the robust principal component analysis method and the non-negative matrix factorization method, respectively, when the input SNR is -5 dB.

A CLASS OF MULTILEVEL RECURSIVE INCOMPLETE LU PRECONDITIONING TECHNIQUES

  • Zhang, Jun
    • Journal of applied mathematics & informatics
    • /
    • v.8 no.2
    • /
    • pp.305-326
    • /
    • 2001
  • We introduce a class of multilevel recursive incomplete LU preconditioning techniques (RILUM) for solving general sparse matrices. This techniques is based on a recursive two by two block incomplete LU factorization on the coefficient martix. The coarse level system is constructed as an (approximate) Schur complement. A dynamic preconditioner is obtained by solving the Schur complement matrix approximately. The novelty of the proposed techniques is to solve the Schur complement matrix by a preconditioned Krylov subspace method. Such a reduction process is repeated to yield a multilevel recursive preconditioner.

An Efficient Implementation of the Supernodal Multifrontal Method (초마디 멀티프런탈 방법의 효율적인 구현)

  • 박찬규;박순달
    • Korean Management Science Review
    • /
    • v.19 no.2
    • /
    • pp.155-168
    • /
    • 2002
  • In this paper, some efficient implementation techniques for the multifrontal method, which can be used to compute the Cholesky factor of a symmetric positive definite matrix, are presented. In order to use the cache effect in the cache-based computer architecture, a hybrid method for factorizing a frontal matrix is considered. This hybrid method uses the column Cholesky method and the submatrix Cholesky method alternatively. Experiments show that the hybrid method speeds up the performance of the supernodal multifrontal method by 5%~10%, and it is superior to the Cholesky method in some problems with dense columns or large frontal matrices.

A Network Reduction using Weak Coupling Method (Weak Coupling Method를 이용한 계통 축약)

  • Lee, H.M.;Rho, K.M.;Kwon, S.H.
    • Proceedings of the KIEE Conference
    • /
    • 1999.07c
    • /
    • pp.1067-1069
    • /
    • 1999
  • This paper presents a network reduction using weak coupling method. Weak coupling method of identifying coherent generator groups are proposed. The partitioning technique used in this paper is based on a property of sparse matrix factorization. When a matrix has been factorized, a system is divided into study area, boundary buses and external area. A reduction process for external system starts with the load bus elimination and coherent generator aggregation. An identification of coherent generator group, network partitioning and network reduction are presented.

  • PDF

Rank-preserver of Matrices over Chain Semiring

  • Song, Seok-Zun;Kang, Kyung-Tae
    • Kyungpook Mathematical Journal
    • /
    • v.46 no.1
    • /
    • pp.89-96
    • /
    • 2006
  • For a rank-1 matrix A, there is a factorization as $A=ab^t$, the product of two vectors a and b. We characterize the linear operators that preserve rank and some equivalent condition of rank-1 matrices over a chain semiring. We also obtain a linear operator T preserves the rank of rank-1 matrices if and only if it is a form (P, Q, B)-operator with appropriate permutation matrices P and Q, and a matrix B with all nonzero entries.

  • PDF

Unified Parametric Approaches for Observer Design in Matrix Second-order Linear Systems

  • Wu Yun-Li;Duan Guang-Ren
    • International Journal of Control, Automation, and Systems
    • /
    • v.3 no.2
    • /
    • pp.159-165
    • /
    • 2005
  • This paper designs observers for matrix second-order linear systems on the basis of generalized eigenstructure assignment via unified parametric approach. It is shown that the problem is closely related with a type of so-called generalized matrix second-order Sylvester matrix equations. Through establishing two general parametric solutions to this type of matrix equations, two unified complete parametric methods for the proposed observer design problem are presented. Both methods give simple complete parametric expressions for the observer gain matrices. The first one mainly depends on a series of singular value decompositions, and is thus numerically simple and reliable; the second one utilizes the right factorization of the system, and allows eigenvalues of the error system to be set undetermined and sought via certain optimization procedures. A spring-mass system is utilized to show the effect of the proposed approaches.

BOOLEAN RANK INEQUALITIES AND THEIR EXTREME PRESERVERS

  • Song, Seok-Zun;Kang, Mun-Hwan
    • Journal of applied mathematics & informatics
    • /
    • v.29 no.5_6
    • /
    • pp.1525-1532
    • /
    • 2011
  • The $m{\times}n$ Boolean matrix A is said to be of Boolean rank r if there exist $m{\times}r$ Boolean matrix B and $r{\times}n$ Boolean matrix C such that A = BC and r is the smallest positive integer that such a factorization exists. We consider the the sets of matrix ordered pairs which satisfy extremal properties with respect to Boolean rank inequalities of matrices over nonbinary Boolean algebra. We characterize linear operators that preserve these sets of matrix ordered pairs as the form of $T(X)=PXP^T$ with some permutation matrix P.