• Title/Summary/Keyword: adaptive scaling

Search Result 101, Processing Time 0.025 seconds

Audio Fingerprinting Using a Robust Hash Function Based on the MCLT Peak-Pair (MCLT 피크쌍 기반의 강인한 해시 함수를 이용한 오디오 핑거프린팅)

  • Lee, Jun-Yong;Kim, Hyoung-Gook
    • The Journal of the Acoustical Society of Korea
    • /
    • v.34 no.2
    • /
    • pp.157-162
    • /
    • 2015
  • In this paper, we propose an audio fingerprinting using robust hash based on the MCLT (Modulated Complex Lapped Transform) peak-pair. In existing methods, the robust audio fingerprinting is not generated if various distortions occurred; time-scaling, pith-shifting and equalization. To solve this problem, we used the spectrum of the MCLT, an adaptive thresholding method for detection of prominent peaks and the novel hash function in the audio fingerprinting. Experimental results show that the proposed method is highly robust in various distorted environments and achieves better identification rates compared to other methods.

Rotation-Scale-Translation-Intensity Invariant Algorithm for Fingerprint Identigfication (RSTI 불변 지문인식 알고리즘)

  • Kim, Hyun;Kim, Hak-Il
    • Journal of the Korean Institute of Telematics and Electronics S
    • /
    • v.35S no.6
    • /
    • pp.88-100
    • /
    • 1998
  • In this paper, an algorithm for a real-time automatic fingerprint identification system is proposed. The fingerprint feature volume is extracted by considering distinct and local characteristics(such as intensity and image quality difference etc.) in fingerprint images, which makes the algorithm properly adaptive to various image acquisitionj methods. Also the matching technique is designed to be invariant on rotation, scaling and translation (RST) changes while being capable of real-time processing. And the classification of fingerprints is performed based on the ridge flow and the relations among singular points such as cores and deltas. The developed fingerprint identification algorithm has been applied to various sets of fingerprint images such as one from NIST(National Institute of Standards and Technology, USA), a pressed fingerprint database constructed according to Korean population distributions in sex, ages and jobs, and a set of rolled-than-scanned fingerprint images. The overall performance of the algorithm has been analyzed and evaluated to the false rejection ratio of 0.07% while holding the false acceptance ratio of 0%.

  • PDF

Numerical evaluation of gamma radiation monitoring

  • Rezaei, Mohsen;Ashoor, Mansour;Sarkhosh, Leila
    • Nuclear Engineering and Technology
    • /
    • v.51 no.3
    • /
    • pp.807-817
    • /
    • 2019
  • Airborne Gamma Ray Spectrometry (AGRS) with its important applications such as gathering radiation information of ground surface, geochemistry measuring of the abundance of Potassium, Thorium and Uranium in outer earth layer, environmental and nuclear site surveillance has a key role in the field of nuclear science and human life. The Broyden-Fletcher-Goldfarb-Shanno (BFGS), with its advanced numerical unconstrained nonlinear optimization in collaboration with Artificial Neural Networks (ANNs) provides a noteworthy opportunity for modern AGRS. In this study a new AGRS system empowered by ANN-BFGS has been proposed and evaluated on available empirical AGRS data. To that effect different architectures of adaptive ANN-BFGS were implemented for a sort of published experimental AGRS outputs. The selected approach among of various training methods, with its low iteration cost and nondiagonal scaling allocation is a new powerful algorithm for AGRS data due to its inherent stochastic properties. Experiments were performed by different architectures and trainings, the selected scheme achieved the smallest number of epochs, the minimum Mean Square Error (MSE) and the maximum performance in compare with different types of optimization strategies and algorithms. The proposed method is capable to be implemented on a cost effective and minimum electronic equipment to present its real-time process, which will let it to be used on board a light Unmanned Aerial Vehicle (UAV). The advanced adaptation properties and models of neural network, the training of stochastic process and its implementation on DSP outstands an affordable, reliable and low cost AGRS design. The main outcome of the study shows this method increases the quality of curvature information of AGRS data while cost of the algorithm is reduced in each iteration so the proposed ANN-BFGS is a trustworthy appropriate model for Gamma-ray data reconstruction and analysis based on advanced novel artificial intelligence systems.

Seismic damage assessment of a large concrete gravity dam

  • Lounis Guechari;Abdelghani Seghir;Ouassila Kada;Abdelhamid Becheur
    • Earthquakes and Structures
    • /
    • v.25 no.2
    • /
    • pp.125-134
    • /
    • 2023
  • In the present work, a new global damage index is proposed for the seismic performance and failure analysis of concrete gravity dams. Unlike the existing indices of concrete structures, this index doesn't need scaling with an ultimate or an upper value. For this purpose, the Beni-Haroun dam in north-eastern Algeria, is considered as a case study, for which an average seismic capacity curve is first evaluated by performing several incremental dynamic analyses. The seismic performance point of the dam is then determined using the N2 method, considering multiple modes and taking into account the stiffness degradation. The seismic demand is obtained from the design spectrum of the Algerian seismic regulations. A series of recorded and artificial accelerograms are used as dynamic loads to evaluate the nonlinear responses of the dam. The nonlinear behaviour of the concrete mass is modelled by using continuum damage mechanics, where material damage is represented by a scalar field damage variable. This modelling, which is suitable for cyclic loading, uses only a single damage parameter to describe the stiffness degradation of the concrete. The hydrodynamic and the sediment pressures are included in the analyses. The obtained results show that the proposed damage index faithfully describes the successive brittle failures of the dam which increase with increasing applied ground accelerations. It is found that minor damage can occur for ground accelerations less than 0.3 g, and complete failure can be caused by accelerations greater than 0.45 g.

Capacity Comparison of Two Uplink OFDMA Systems Considering Synchronization Error among Multiple Users and Nonlinear Distortion of Amplifiers (사용자간 동기오차와 증폭기의 비선형 왜곡을 동시에 고려한 두 상향링크 OFDMA 기법의 채널용량 비교 분석)

  • Lee, Jin-Hui;Kim, Bong-Seok;Choi, Kwonhue
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.39A no.5
    • /
    • pp.258-270
    • /
    • 2014
  • In this paper, we investigate channel capacity of two kinds of uplink OFDMA (Orthogonal Frequency Division Multiple Access) schemes, i.e. ZCZ (Zero Correlation Zone) code time-spread OFDMA and sparse SC-FDMA (Single Carrier Frequency Division Mmultiple Access) robust to access timing offset (TO) among multiple users. In order to reflect the practical condition, we consider not only access TO among multiple users but also peak to average power ratio (PAPR) which is one of hot issues of uplink OFDMA. In the case with access TO among multiple users, the amplified signal of users by power control might affect a severe interference to signals of other users. Meanwhile, amplified signal by considering distance between user and base station might be distorted due to the limit of amplifier and thus the performance might degrade. In order to achieve the maximum channel capacity, we investigate the combinations of transmit power so called ASF (adaptive scaling factor) by numerical simulations. We check that the channel capacity of the case with ASF increases compared to the case with considering only distance i.e. ASF=1. From the simulation results, In the case of high signal to noise ratio (SNR), ZCZ code time-spread OFDMA achieves higher channel capacity compared to sparse block SC-FDMA. On the other hand, in the case of low SNR, the sparse block SC-FDMA achieves better performance compared to ZCZ time-spread OFDMA.

Optimal Spatial Scale for Land Use Change Modelling : A Case Study in a Savanna Landscape in Northern Ghana (지표피복변화 연구에서 최적의 공간스케일의 문제 : 가나 북부지역의 사바나 지역을 사례로)

  • Nick van de Giesen;Paul L. G. Vlek;Park Soo Jin
    • Journal of the Korean Geographical Society
    • /
    • v.40 no.2 s.107
    • /
    • pp.221-241
    • /
    • 2005
  • Land Use and Land Cover Changes (LUCC) occur over a wide range of space and time scales, and involve complex natural, socio-economic, and institutional processes. Therefore, modelling and predicting LUCC demands an understanding of how various measured properties behave when considered at different scales. Understanding spatial and temporal variability of driving forces and constraints on LUCC is central to understanding the scaling issues. This paper aims to 1) assess the heterogeneity of land cover change processes over the landscape in northern Ghana, where intensification of agricultural activities has been the dominant land cover change process during the past 15 years, 2) characterise dominant land cover change mechanisms for various spatial scales, and 3) identify the optimal spatial scale for LUCC modelling in a savanna landscape. A multivariate statistical method was first applied to identify land cover change intensity (LCCI), using four time-sequenced NDVI images derived from LANDSAT scenes. Three proxy land use change predictors: distance from roads, distance from surface water bodies, and a terrain characterisation index, were regressed against the LCCI using a multi-scale hierarchical adaptive model to identify scale dependency and spatial heterogeneity of LUCC processes. High spatial associations between the LCCI and land use change predictors were mostly limited to moving windows smaller than 10$\times$10km. With increasing window size, LUCC processes within the window tend to be too diverse to establish clear trends, because changes in one part of the window are compensated elsewhere. This results in a reduced correlation between LCCI and land use change predictors at a coarser spatial extent. The spatial coverage of 5-l0km is incidentally equivalent to a village or community area in the study region. In order to reduce spatial variability of land use change processes for regional or national level LUCC modelling, we suggest that the village level is the optimal spatial investigation unit in this savanna landscape.

An adaptive digital watermark using the spatial masking (공간 마스킹을 이용한 적응적 디지털 워터 마크)

  • 김현태
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.9 no.3
    • /
    • pp.39-52
    • /
    • 1999
  • In this paper we propose a new watermarking technique for copyright protection of images. The proposed technique is based on a spatial masking method with a spatial scale parameter. In general it becomes more robust against various attacks but with some degradations on the image quality as the amplitude of the watermark increases. On the other hand it becomes perceptually more invisible but more vulnerable to various attacks as the amplitude of the watermark decreases. Thus it is quite complex to decide the compromise between the robustness of watermark and its visibility. We note that watermarking using the spread spectrum is not robust enought. That is there may be some areas in the image that are tolerable to strong watermark signals. However large smooth areas may not be strong enough. Thus in order to enhance the invisibility of watermarked image for those areas the spatial masking characteristics of the HVS(Human Visual System) should be exploited. That is for texture regions the magnitude of the watermark can be large whereas for those smooth regions the magnitude of the watermark can be small. As a result the proposed watermarking algorithm is intend to satisfy both the robustness of watermark and the quality of the image. The experimental results show that the proposed algorithm is robust to image deformations(such as compression adding noise image scaling clipping and collusion attack).

Reference Frame Memory Compression Using Selective Processing Unit Merging Method (선택적 수행블록 병합을 이용한 참조 영상 메모리 압축 기법)

  • Hong, Soon-Gi;Choe, Yoon-Sik;Kim, Yong-Goo
    • Journal of Broadcast Engineering
    • /
    • v.16 no.2
    • /
    • pp.339-349
    • /
    • 2011
  • IBDI (Internal Bit Depth Increase) is able to significantly improve the coding efficiency of high definition video compression by increasing the bit depth (or precision) of internal arithmetic operation. However the scheme also increases required internal memory for storing decoded reference frames and this can be significant for higher definition of video contents. So, the reference frame memory compression method is proposed to reduce such internal memory requirement. The reference memory compression is performed on 4x4 block called the processing unit to compress the decoded image using the correlation of nearby pixel values. This method has successively reduced the reference frame memory while preserving the coding efficiency of IBDI. However, additional information of each processing unit has to be stored also in internal memory, the amount of additional information could be a limitation of the effectiveness of memory compression scheme. To relax this limitation of previous memory compression scheme, we propose a selective merging-based reference frame memory compression algorithm, dramatically reducing the amount of additional information. Simulation results show that the proposed algorithm provides much smaller overhead than that of the previous algorithm while keeping the coding efficiency of IBDI.

Joint Optimization of the Motion Estimation Module and the Up/Down Scaler in Transcoders television (트랜스코더의 해상도 변환 모듈과 움직임 추정 모듈의 공동 최적화)

  • Han, Jong-Ki;Kwak, Sang-Min;Jun, Dong-San;Kim, Jae-Gon
    • Journal of Broadcast Engineering
    • /
    • v.10 no.3
    • /
    • pp.270-285
    • /
    • 2005
  • A joint design scheme is proposed to optimize the up/down scaler and the motion vector estimation module in the transcoder system. The proposed scheme first optimizes the resolution scaler for a fixed motion vector, and then a new motion vector is estimated for the fixed scaler. These two steps are iteratively repeated until they reach a local optimum solution. In the optimization of the scaler, we derive an adaptive version of a cubic convolution interpolator to enlarge or reduce digital images by arbitrary scaling factors. The adaptation is performed at each macroblock of an image. In order to estimate the optimal motion vector, a temporary motion vector is composed from the given motion vectors. Then the motion vector is refined over a narrow search range. It is well-known that this refinement scheme provides the comparable performance compared to the full search method. Simulation results show that a jointly optimized system based on the proposed algorithms outperforms the conventional systems. We can also see that the algorithms exhibit significant improvement in the minimization of information loss compared with other techniques.

Speech extraction based on AuxIVA with weighted source variance and noise dependence for robust speech recognition (강인 음성 인식을 위한 가중화된 음원 분산 및 잡음 의존성을 활용한 보조함수 독립 벡터 분석 기반 음성 추출)

  • Shin, Ui-Hyeop;Park, Hyung-Min
    • The Journal of the Acoustical Society of Korea
    • /
    • v.41 no.3
    • /
    • pp.326-334
    • /
    • 2022
  • In this paper, we propose speech enhancement algorithm as a pre-processing for robust speech recognition in noisy environments. Auxiliary-function-based Independent Vector Analysis (AuxIVA) is performed with weighted covariance matrix using time-varying variances with scaling factor from target masks representing time-frequency contributions of target speech. The mask estimates can be obtained using Neural Network (NN) pre-trained for speech extraction or diffuseness using Coherence-to-Diffuse power Ratio (CDR) to find the direct sounds component of a target speech. In addition, outputs for omni-directional noise are closely chained by sharing the time-varying variances similarly to independent subspace analysis or IVA. The speech extraction method based on AuxIVA is also performed in Independent Low-Rank Matrix Analysis (ILRMA) framework by extending the Non-negative Matrix Factorization (NMF) for noise outputs to Non-negative Tensor Factorization (NTF) to maintain the inter-channel dependency in noise output channels. Experimental results on the CHiME-4 datasets demonstrate the effectiveness of the presented algorithms.