• Title/Summary/Keyword: quantization errors

Search Result 93, Processing Time 0.032 seconds

Background Subtraction Algorithm by Using the Local Binary Pattern Based on Hexagonal Spatial Sampling (육각화소 기반의 지역적 이진패턴을 이용한 배경제거 알고리즘)

  • Choi, Young-Kyu
    • The KIPS Transactions:PartB
    • /
    • v.15B no.6
    • /
    • pp.533-542
    • /
    • 2008
  • Background subtraction from video data is one of the most important task in various realtime machine vision applications. In this paper, a new scheme for background subtraction based on the hexagonal pixel sampling is proposed. Generally it has been found that hexagonal spatial sampling yields smaller quantization errors and remarkably improves the understanding of connectivity. We try to apply the hexagonally sampled image to the LBP based non-parametric background subtraction algorithm. Our scheme makes it possible to omit the bilinear pixel interpolation step during the local binary pattern generation process, and, consequently, can reduce the computation time. Experimental results revealed that our approach based on hexagonal spatial sampling is very efficient and can be utilized in various background subtraction applications.

Discrete Multiwavelet-Based Video Watermarking Scheme Using SURF

  • Narkedamilly, Leelavathy;Evani, Venkateswara Prasad;Samayamantula, Srinivas Kumar
    • ETRI Journal
    • /
    • v.37 no.3
    • /
    • pp.595-605
    • /
    • 2015
  • This paper proposes a robust, imperceptible block-based digital video watermarking algorithm that makes use of the Speeded Up Robust Feature (SURF) technique. The SURF technique is used to extract the most important features of a video. A discrete multiwavelet transform (DMWT) domain in conjunction with a discrete cosine transform is used for embedding a watermark into feature blocks. The watermark used is a binary image. The proposed algorithm is further improved for robustness by an error-correction code to protect the watermark against bit errors. The same watermark is embedded temporally for every set of frames of an input video to improve the decoded watermark correlation. Extensive experimental results demonstrate that the proposed DMWT domain video watermarking using SURF features is robust against common image processing attacks, motion JPEG2000 compression, frame averaging, and frame swapping attacks. The quality of a watermarked video under the proposed algorithm is high, demonstrating the imperceptibility of an embedded watermark.

An Embedding /Extracting Method of Audio Watermark Information for High Quality Stereo Music (고품질 스테레오 음악을 위한 오디오 워터마크 정보 삽입/추출 기술)

  • Bae, Kyungyul
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.21-35
    • /
    • 2018
  • Since the introduction of MP3 players, CD recordings have gradually been vanishing, and the music consuming environment of music users is shifting to mobile devices. The introduction of smart devices has increased the utilization of music through music playback, mass storage, and search functions that are integrated into smartphones and tablets. At the time of initial MP3 player supply, the bitrate of the compressed music contents generally was 128 Kbps. However, as increasing of the demand for high quality music, sound quality of 384 Kbps appeared. Recently, music content of FLAC (Free License Audio Codec) format using lossless compression method is becoming popular. The download service of many music sites in Korea has classified by unlimited download with technical protection and limited download without technical protection. Digital Rights Management (DRM) technology is used as a technical protection measure for unlimited download, but it can only be used with authenticated devices that have DRM installed. Even if music purchased by the user, it cannot be used by other devices. On the contrary, in the case of music that is limited in quantity but not technically protected, there is no way to enforce anyone who distributes it, and in the case of high quality music such as FLAC, the loss is greater. In this paper, the author proposes an audio watermarking technology for copyright protection of high quality stereo music. Two kinds of information, "Copyright" and "Copy_free", are generated by using the turbo code. The two watermarks are composed of 9 bytes (72 bits). If turbo code is applied for error correction, the amount of information to be inserted as 222 bits increases. The 222-bit watermark was expanded to 1024 bits to be robust against additional errors and finally used as a watermark to insert into stereo music. Turbo code is a way to recover raw data if the damaged amount is less than 15% even if part of the code is damaged due to attack of watermarked content. It can be extended to 1024 bits or it can find 222 bits from some damaged contents by increasing the probability, the watermark itself has made it more resistant to attack. The proposed algorithm uses quantization in DCT so that watermark can be detected efficiently and SNR can be improved when stereo music is converted into mono. As a result, on average SNR exceeded 40dB, resulting in sound quality improvements of over 10dB over traditional quantization methods. This is a very significant result because it means relatively 10 times improvement in sound quality. In addition, the sample length required for extracting the watermark can be extracted sufficiently if the length is shorter than 1 second, and the watermark can be completely extracted from music samples of less than one second in all of the MP3 compression having a bit rate of 128 Kbps. The conventional quantization method can extract the watermark with a length of only 1/10 compared to the case where the sampling of the 10-second length largely fails to extract the watermark. In this study, since the length of the watermark embedded into music is 72 bits, it provides sufficient capacity to embed necessary information for music. It is enough bits to identify the music distributed all over the world. 272 can identify $4*10^{21}$, so it can be used as an identifier and it can be used for copyright protection of high quality music service. The proposed algorithm can be used not only for high quality audio but also for development of watermarking algorithm in multimedia such as UHD (Ultra High Definition) TV and high-resolution image. In addition, with the development of digital devices, users are demanding high quality music in the music industry, and artificial intelligence assistant is coming along with high quality music and streaming service. The results of this study can be used to protect the rights of copyright holders in these industries.

The Development of Dynamic Forecasting Model for Short Term Power Demand using Radial Basis Function Network (Radial Basis 함수를 이용한 동적 - 단기 전력수요예측 모형의 개발)

  • Min, Joon-Young;Cho, Hyung-Ki
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.7
    • /
    • pp.1749-1758
    • /
    • 1997
  • This paper suggests the development of dynamic forecasting model for short-term power demand based on Radial Basis Function Network and Pal's GLVQ algorithm. Radial Basis Function methods are often compared with the backpropagation training, feed-forward network, which is the most widely used neural network paradigm. The Radial Basis Function Network is a single hidden layer feed-forward neural network. Each node of the hidden layer has a parameter vector called center. This center is determined by clustering algorithm. Theatments of classical approached to clustering methods include theories by Hartigan(K-means algorithm), Kohonen(Self Organized Feature Maps %3A SOFM and Learning Vector Quantization %3A LVQ model), Carpenter and Grossberg(ART-2 model). In this model, the first approach organizes the load pattern into two clusters by Pal's GLVQ clustering algorithm. The reason of using GLVQ algorithm in this model is that GLVQ algorithm can classify the patterns better than other algorithms. And the second approach forecasts hourly load patterns by radial basis function network which has been constructed two hidden nodes. These nodes are determined from the cluster centers of the GLVQ in first step. This model was applied to forecast the hourly loads on Mar. $4^{th},\;Jun.\;4^{th},\;Jul.\;4^{th},\;Sep.\;4^{th},\;Nov.\;4^{th},$ 1995, after having trained the data for the days from Mar. $1^{th}\;to\;3^{th},\;from\;Jun.\;1^{th}\;to\;3^{th},\;from\;Jul.\;1^{th}\;to\;3^{th},\;from\;Sep.\;1^{th}\;to\;3^{th},\;and\;from\;Nov.\;1^{th}\;to\;3^{th},$ 1995, respectively. In the experiments, the average absolute errors of one-hour ahead forecasts on utility actual data are shown to be 1.3795%.

  • PDF

A Study on the Mixed Model Approach and Symbol Probability Weighting Function for Maximization of Inter-Speaker Variation (화자간 변별력 최대화를 위한 혼합 모델 방식과 심볼 확률 가중함수에 관한 연구)

  • Chin Se-Hoon;Kang Chul-Ho
    • The Journal of the Acoustical Society of Korea
    • /
    • v.24 no.7
    • /
    • pp.410-415
    • /
    • 2005
  • Recently, most of the speaker verification systems are based on the pattern recognition approach method. And performance of the pattern-classifier depends on how to classify a variety of speakers' feature parameters. In order to classify feature parameters efficiently and effectively, it is of great importance to enlarge variations between speakers and effectively measure distances between feature parameters. Therefore, this paper would suggest the positively mixed model scheme that can enlarge inter-speaker variation by searching the individual model with world model at the same time. During decision procedure, we can maximize inter-speaker variation by using the proposed mixed model scheme. We also make use of a symbol probability weighting function in this system so as to reduce vector quantization errors by measuring symbol probability derived from the distance rate of between the world codebook and individual codebook. As the result of our experiment using this method, we could halve the Detection Cost Function (DCF) of the system from $2.37\%\;to\;1.16\%$.

Distance Estimation Method using Enhanced Adaptive Fuzzy Strong Tracking Kalman Filter Based on Stereo Vision (스테레오 비전에서 향상된 적응형 퍼지 칼만 필터를 이용한 거리 추정 기법)

  • Lim, Young-Chul;Lee, Chung-Hee;Kwon, Soon;Lee, Jong-Hoon
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.45 no.6
    • /
    • pp.108-116
    • /
    • 2008
  • In this paper, we propose an algorithm that can estimate the distance using disparity based on stereo vision system, even though the obstacle is located in long ranges as well as short ranges. We use sub-pixel interpolation to minimize quantization errors which deteriorate the distance accuracy when calculating the distance with integer disparity, and also we use enhanced adaptive fuzzy strong tracking Kalman filter(EAFSTKF) to improve the distance accuracy and track the path optimally. The proposed method can solve the divergence problem caused by nonlinear dynamics such as various vehicle movements in the conventional Kalman filter(CKF), and also enhance the distance accuracy and reliability. Our simulation results show that the performance of our method improves by about 13.5% compared to other methods in point of root mean square error rate(RMSER).

The Modeling of the Optimal Data Format for JPEG2000 CODEC on the Fixed Compression Ratio (고정 압축률에서의 JPEG2000 코덱을 위한 최적의 데이터 형식 모델링)

  • Seo, Choon-Weon
    • Journal of the Korean Institute of Illuminating and Electrical Installation Engineers
    • /
    • v.19 no.6
    • /
    • pp.109-116
    • /
    • 2005
  • Recently, images/videos have been preferred as the communication media because of their information-implication and easy recognizability. But the amount of their data is so large that it has been the major research area to compress their data. This paper is related to optimization in th image data format, which can make a great effect in performance of data compression and is based on the wavelet transform and JPEG2000. This paper established a criterion to decide the data format to be used in wavelet transform, which is on the bases of the data errors in frequency transform and quantization. This criterion has been used to extract the optimal data format experimentally. The result were(1, 9) of 10-bit fixed-point format for filter coefficients and (9, 7) of 16-bit fixed-point data dormat for wavelet coefficients and their optimality was confirmed.

Illumination estimation based on valid pixel selection from CCD camera response (CCD카메라 응답으로부터 유효 화소 선택에 기반한 광원 추정)

  • 권오설;조양호;김윤태;송근호;하영호
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.41 no.5
    • /
    • pp.251-258
    • /
    • 2004
  • This paper proposes a method for estimating the illuminant chromaticity using the distributions of the camera responses obtained by a CCD camera in a real-world scene. Illuminant estimation using a highlight method is based on the geometric relation between a body and its surface reflection. In general, the pixels in a highlight region are affected by an illuminant geometric difference, camera quantization errors, and the non-uniformity of the CCD sensor. As such, this leads to inaccurate results if an illuminant is estimated using the pixels of a CCD camera without any preprocessing. Accordingly, to solve this problem the proposed method analyzes the distribution of the CCD camera responses and selects pixels using the Mahalanobis distance in highlight regions. The use of the Mahalanobis distance based on the camera responses enables the adaptive selection of valid pixels among the pixels distributed in the highlight regions. Lines are then determined based on the selected pixels with r-g chromaticity coordinates using a principal component analysis(PCA). Thereafter, the illuminant chromaticity is estimated based on the intersection points of the lines. Experimental results using the proposed method demonstrated a reduced estimation error compared with the conventional method.

The Modeling of the Optimal Data Format for JPEG2000 CODEC (JPEG2000 코덱을 위한 최적의 데이터 형식 모델링)

  • Kang, Chang-Soo;Seo, Choon-Weon
    • Journal of the Institute of Electronics Engineers of Korea TE
    • /
    • v.42 no.4
    • /
    • pp.7-14
    • /
    • 2005
  • Recently, images/videos have been preferred as the communication media because of their information-implication and easy recognizability. But the amount of their data is so large that it has been the major research area to compress their data. This paper is related to optimization in the image data format which can make a great effect in performance of data compression and is based on the wavelet transform and JPEG2000. This paper established a criterion to decide the data format to he used in wavelet transform which is on the bases of the data errors in frequency transform and quantization. This criterion has been used to extract the optimal data format experimentally. The result were (1, 9) of 10-bit fixed-point format for filter coefficients and (9, 7) of 16-bit fixed-point data format for wavelet coefficients and their optimality was confirmed.

Edge Enhanced Error Diffusion with Blue Noise Mask Threshold Modulation (청색잡음 마스크 임계값변조를 이용한 경계강조 오차확산법)

  • Lee, Eul-Hwan;Park, Jang-Sik;Park, Chang-Dae;Kim, Jae-Ho
    • Journal of the Korean Institute of Telematics and Electronics S
    • /
    • v.36S no.10
    • /
    • pp.72-82
    • /
    • 1999
  • The error diffusion algorithm is excellent for reproducing continuous gray-scale images to bianry images and also has good edge characteristics. However, it is well known that artifacts with objectionable patterns can occur in the halftoned images. On the other hand, a halftone algorithm using blue noise mask has been proposed. where the halftoning is achieved by a pixelwise comparison of gray-scale image with an array, the blue noise mask. It doesn't have pattern artifacts, but the halftoned image looks unclear because the quantization errors are not feedbacked compared to the error diffusion. In this paper, edge enhanced error diffusion which dithers the threshold with the blue noise mask is proposed. We show that the proposed algorithm can produce unstructured and edge enhanced halftone images. This algorithm is analyzed by the concept of an equivalent input image. The performace of the proposed algorithm is compared with that of the conventional halftoning by measuring the radially averaged power spectrum and edge correlation.

  • PDF