• Title/Summary/Keyword: sparse regularization

Search Result 26, Processing Time 0.021 seconds

A time delay estimation method using canonical correlation analysis and log-sum regularization (로그-합 규준화와 정준형 상관 분석을 이용한 시간 지연 추정에 관한 연구)

  • Lim, Jun-Seok;Pyeon, Yong-Gook;Lee, Seokjin;Cheong, MyoungJun
    • The Journal of the Acoustical Society of Korea
    • /
    • v.36 no.4
    • /
    • pp.279-284
    • /
    • 2017
  • The localization of sources has a numerous number of applications. To estimate the position of sources, the relative time delay between two or more received signals for the direct signal must be determined. Although the GCC (Generalized Cross-Correlation) method is the most popular technique, an approach based on CCA (Canonical Correlation Analysis) was also proposed for the TDE (Time Delay Estimation). In this paper, we propose a new adaptive algorithm based on CCA in order to utilized the sparsity in the eigenvector of CCA based time delay estimator. The proposed algorithm uses the eigenvector corresponding to the maximum eigenvalue with log-sum regularization in order to utilize the sparsity in the eigenvector. We have performed simulations for several SNR(signal to noise ratio)s, showing that the new CCA based algorithm can estimate the time delays more accurately than the conventional CCA and GCC based TDE algorithms.

Feature selection for text data via sparse principal component analysis (희소주성분분석을 이용한 텍스트데이터의 단어선택)

  • Won Son
    • The Korean Journal of Applied Statistics
    • /
    • v.36 no.6
    • /
    • pp.501-514
    • /
    • 2023
  • When analyzing high dimensional data such as text data, if we input all the variables as explanatory variables, statistical learning procedures may suffer from over-fitting problems. Furthermore, computational efficiency can deteriorate with a large number of variables. Dimensionality reduction techniques such as feature selection or feature extraction are useful for dealing with these problems. The sparse principal component analysis (SPCA) is one of the regularized least squares methods which employs an elastic net-type objective function. The SPCA can be used to remove insignificant principal components and identify important variables from noisy observations. In this study, we propose a dimension reduction procedure for text data based on the SPCA. Applying the proposed procedure to real data, we find that the reduced feature set maintains sufficient information in text data while the size of the feature set is reduced by removing redundant variables. As a result, the proposed procedure can improve classification accuracy and computational efficiency, especially for some classifiers such as the k-nearest neighbors algorithm.

Reconstruction of Collagen Using Tensor-Voting & Graph-Cuts

  • Park, Doyoung
    • Journal of Advanced Information Technology and Convergence
    • /
    • v.9 no.1
    • /
    • pp.89-102
    • /
    • 2019
  • Collagen can be used in building artificial skin replacements for treatment of burns and towards the reconstruction of bone as well as researching cell behavior and cellular interaction. The strength of collagen in connective tissue rests on the characteristics of collagen fibers. 3D confocal imaging of collagen fibers enables the characterization of their spatial distribution as related to their function. However, the image stacks acquired with confocal laser-scanning microscope does not clearly show the collagen architecture in 3D. Therefore, we developed a new method to reconstruct, visualize and characterize collagen fibers from fluorescence confocal images. First, we exploit the tensor voting framework to extract sparse reliable information about collagen structure in a 3D image and therefore denoise and filter the acquired image stack. We then propose to segment the collagen fibers by defining an energy term based on the Hessian matrix. This energy term is minimized by a min cut-max flow algorithm that allows adaptive regularization. We demonstrate the efficacy of our methods by visualizing reconstructed collagen from specific 3D image stack.

QoS Constrained Optimization of Cell Association and Resource Allocation for Load Balancing in Downlink Heterogeneous Cellular Networks

  • Su, Gongchao;Chen, Bin;Lin, Xiaohui;Wang, Hui;Li, Lemin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.5
    • /
    • pp.1569-1586
    • /
    • 2015
  • This paper considers the optimal cell association and resource allocation for load balancing in a heterogeneous cellular network subject to user's quality-of-service (QoS) constraints. We adopt the proportional fairness (PF) utility maximization formulation which also accommodates the QoS constraints in terms of minimum rate requirements. With equal resource allocation this joint optimization problem is either infeasible or requires relaxation that yields a solution which is difficult to implement. Nevertheless, we show that this joint optimization problem can be effectively solved without any priori assumption on resource allocation and yields a cell association scheme which enforces single BS association for each user. We re-formulated the joint optimization problem as a network-wide resource allocation problem with cardinality constraints. A reweighted heuristic l1-norm regularization method is used to obtain a sparse solution to the re-formulated problem. The cell association scheme is then derived from the sparsity pattern of the solution, which guarantees a single BS association for each user. Compared with the previously proposed method based on equal resource allocation, the proposed framework results in a feasible cell association scheme and yields a robust solution on resource allocation that satisfies the QoS constraints. Our simulations illustrate the impact of user's minimum rate requirements on cell association and demonstrate that the proposed approach achieves load balancing and enforces single BS association for users.

GEase-K: Linear and Nonlinear Autoencoder-based Recommender System with Side Information (GEase-K: 부가 정보를 활용한 선형 및 비선형 오토인코더 기반의 추천시스템)

  • Taebeom Lee;Seung-hak Lee;Min-jeong Ma;Yoonho Cho
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.3
    • /
    • pp.167-183
    • /
    • 2023
  • In the recent field of recommendation systems, various studies have been conducted to model sparse data effectively. Among these, GLocal-K(Global and Local Kernels for Recommender Systems) is a research endeavor combining global and local kernels to provide personalized recommendations by considering global data patterns and individual user characteristics. However, due to its utilization of kernel tricks, GLocal-K exhibits diminished performance on highly sparse data and struggles to offer recommendations for new users or items due to the absence of side information. In this paper, to address these limitations of GLocal-K, we propose the GEase-K (Global and EASE kernels for Recommender Systems) model, incorporating the EASE(Embarrassingly Shallow Autoencoders for Sparse Data) model and leveraging side information. Initially, we substitute EASE for the local kernel in GLocal-K to enhance recommendation performance on highly sparse data. EASE, functioning as a simple linear operational structure, is an autoencoder that performs highly on extremely sparse data through regularization and learning item similarity. Additionally, we utilize side information to alleviate the cold-start problem. We enhance the understanding of user-item similarities by employing a conditional autoencoder structure during the training process to incorporate side information. In conclusion, GEase-K demonstrates resilience in highly sparse data and cold-start situations by combining linear and nonlinear structures and utilizing side information. Experimental results show that GEase-K outperforms GLocal-K based on the RMSE and MAE metrics on the highly sparse GoodReads and ModCloth datasets. Furthermore, in cold-start experiments divided into four groups using the GoodReads and ModCloth datasets, GEase-K denotes superior performance compared to GLocal-K.

Target-free vision-based approach for vibration measurement and damage identification of truss bridges

  • Dong Tan;Zhenghao Ding;Jun Li;Hong Hao
    • Smart Structures and Systems
    • /
    • v.31 no.4
    • /
    • pp.421-436
    • /
    • 2023
  • This paper presents a vibration displacement measurement and damage identification method for a space truss structure from its vibration videos. Features from Accelerated Segment Test (FAST) algorithm is combined with adaptive threshold strategy to detect the feature points of high quality within the Region of Interest (ROI), around each node of the truss structure. Then these points are tracked by Kanade-Lucas-Tomasi (KLT) algorithm along the video frame sequences to obtain the vibration displacement time histories. For some cases with the image plane not parallel to the truss structural plane, the scale factors cannot be applied directly. Therefore, these videos are processed with homography transformation. After scale factor adaptation, tracking results are expressed in physical units and compared with ground truth data. The main operational frequencies and the corresponding mode shapes are identified by using Subspace Stochastic Identification (SSI) from the obtained vibration displacement responses and compared with ground truth data. Structural damages are quantified by elemental stiffness reductions. A Bayesian inference-based objective function is constructed based on natural frequencies to identify the damage by model updating. The Success-History based Adaptive Differential Evolution with Linear Population Size Reduction (L-SHADE) is applied to minimise the objective function by tuning the damage parameter of each element. The locations and severities of damage in each case are then identified. The accuracy and effectiveness are verified by comparison of the identified results with the ground truth data.