• Title/Summary/Keyword: Error Correcting Code

Search Result 178, Processing Time 0.031 seconds

On algorithm for finding primitive polynomials over GF(q) (GF(q)상의 원시다항식 생성에 관한 연구)

  • 최희봉;원동호
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.11 no.1
    • /
    • pp.35-42
    • /
    • 2001
  • The primitive polynomial on GF(q) is used in the area of the scrambler, the error correcting code and decode, the random generator and the cipher, etc. The algorithm that generates efficiently the primitive polynomial on GF(q) was proposed by A.D. Porto. The algorithm is a method that generates the sequence of the primitive polynomial by repeating to find another primitive polynomial with a known primitive polynomial. In this paper, we propose the algorithm that is improved in the A.D. Porto algorithm. The running rime of the A.D. Porto a1gorithm is O($\textrm{km}^2$), the running time of the improved algorithm is 0(m(m+k)). Here, k is gcd(k, $q^m$-1). When we find the primitive polynomial with m odor, it is efficient that we use the improved algorithm in the condition k, m>>1.

Design and Implementation of Smart Self-Learning Aid: Micro Dot Pattern Recognition based Information Embedding Solution (스마트 학습지: 미세 격자 패턴 인식 기반의 지능형 학습 도우미 시스템의 설계와 구현)

  • Shim, Jae-Youen;Kim, Seong-Whan
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2011.04a
    • /
    • pp.346-349
    • /
    • 2011
  • In this paper, we design a perceptually invisible dot pattern layout and its recognition scheme, and we apply the recognition scheme into a smart self learning aid for interactive learning aid. To increase maximum information capacity and also increase robustness to the noises, we design a ECC (error correcting code) based dot pattern with directional vector indicator. To make a smart self-learning aid, we embed the micro dot pattern (20 information bit + 15 ECC bits + 9 layout information bit) using K ink (CMYK) and extract the dot pattern using IR (infrared) LED and IR filter based camera, which is embedded in the smart pen. The reason we use K ink is that K ink is a carbon based ink in nature, and carbon is easily recognized with IR even without light. After acquiring IR camera images for the dot patterns, we perform layout adjustment using the 9 layout information bit, and extract 20 information bits from 35 data bits which is composed of 20 information bits and 15 ECC bits. To embed and extract information bits, we use topology based dot pattern recognition scheme which is robust to geometric distortion which is very usual in camera based recognition scheme. Topology based pattern recognition traces next information bit symbols using topological distance measurement from the pivot information bit. We implemented and experimented with sample patterns, and it shows that we can achieve almost 99% recognition for our embedding patterns.

Low Complexity Video Encoding Using Turbo Decoding Error Concealments for Sensor Network Application (센서네트워크상의 응용을 위한 터보 복호화 오류정정 기법을 이용한 경량화 비디오 부호화 방법)

  • Ko, Bong-Hyuck;Shim, Hyuk-Jae;Jeon, Byeung-Woo
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.45 no.1
    • /
    • pp.11-21
    • /
    • 2008
  • In conventional video coding, the complexity of encoder is much higher than that of decoder. However, as more needs arises for extremely simple encoder in environments having constrained energy such as sensor network, much investigation has been carried out for eliminating motion prediction/compensation claiming most complexity and energy in encoder. The Wyner-Ziv coding, one of the representative schemes for the problem, reconstructs video at decoder by correcting noise on side information using channel coding technique such as turbo code. Since the encoder generates only parity bits without performing any type of processes extracting correlation information between frames, it has an extremely simple structure. However, turbo decoding errors occur in noisy side information. When there are high-motion or occlusion between frames, more turbo decoding errors appear in reconstructed frame and look like Salt & Pepper noise. This severely deteriorates subjective video quality even though such noise rarely occurs. In this paper, we propose a computationally extremely light encoder based on symbol-level Wyner-Ziv coding technique and a new corresponding decoder which, based on a decision whether a pixel has error or not, applies median filter selectively in order to minimize loss of texture detail from filtering. The proposed method claims extremely low encoder complexity and shows improvements both in subjective quality and PSNR. Our experiments have verified average PSNR gain of up to 0.8dB.

Implementation of Turbo Decoder Based on Two-step SOVA with a Scaling Factor (비례축소인자를 가진 2단 SOVA를 이용한 터보 복호기의 설계)

  • Kim, Dae-Won;Choi, Jun-Rim
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.39 no.11
    • /
    • pp.14-23
    • /
    • 2002
  • Two implementation methods for SOVA (Soft Output Viterbi Algorithm)of Turbo decoder are applied and verfied. The first method is the combination of a trace back (TB) logic for the survivor state and a double trace back logic for the weight value in two-step SOVA. This architecure of two-setp SOVA decoder allows important savings in area and high-speed processing compared with that of one-step SOVA decoding using register exchange (RE) or trace-back (TB) method. Second method is adjusting the reliability value with a scaling factor between 0.25 and 0.33 in order to compensate for the distortion for a rate 1/3 and 8-state SOVA decoder with a 256-bit frame size. The proposed schemes contributed to higher SNR performance by 2dB at the BER 10E-4 than that of SOVA decoder without a scaling factor. In order to verify the suggested schemes, the SOVA decoder is testd using Xillinx XCV 1000E FPGA, which runs at 33.6MHz of the maximum speed with 845 latencies and it features 175K gates in the case of 256-bit frame size.

New Methods for Correcting the Atmospheric Effects in Landsat Imagery over Turbid (Case-2) Waters

  • Ahn Yu-Hwan;Shanmugam P.
    • Korean Journal of Remote Sensing
    • /
    • v.20 no.5
    • /
    • pp.289-305
    • /
    • 2004
  • Atmospheric correction of Landsat Visible and Near Infrared imagery (VIS/NIR) over aquatic environment is more demanding than over land because the signal from the water column is small and it carries immense information about biogeochemical variables in the ocean. This paper introduces two methods, a modified dark-pixel substraction technique (path--extraction) and our spectral shape matching method (SSMM), for the correction of the atmospheric effects in the Landsat VIS/NIR imagery in relation to the retrieval of meaningful information about the ocean color, especially from Case-2 waters (Morel and Prieur, 1977) around Korean peninsula. The results of these methods are compared with the classical atmospheric correction approaches based on the 6S radiative transfer model and standard SeaWiFS atmospheric algorithm. The atmospheric correction scheme using 6S radiative transfer code assumes a standard atmosphere with constant aerosol loading and a uniform, Lambertian surface, while the path-extraction assumes that the total radiance (L/sub TOA/) of a pixel of the black ocean (referred by Antoine and Morel, 1999) in a given image is considered as the path signal, which remains constant over, at least, the sub scene of Landsat VIS/NIR imagery. The assumption of SSMM is nearly similar, but it extracts the path signal from the L/sub TOA/ by matching-up the in-situ data of water-leaving radiance, for typical clear and turbid waters, and extrapolate it to be the spatially homogeneous contribution of the scattered signal after complex interaction of light with atmospheric aerosols and Raleigh particles, and direct reflection of light on the sea surface. The overall shape and magnitude of radiance or reflectance spectra of the atmospherically corrected Landsat VIS/NIR imagery by SSMM appears to have good agreement with the in-situ spectra collected for clear and turbid waters, while path-extraction over turbid waters though often reproduces in-situ spectra, but yields significant errors for clear waters due to the invalid assumption of zero water-leaving radiance for the black ocean pixels. Because of the standard atmosphere with constant aerosols and models adopted in 6S radiative transfer code, a large error is possible between the retrieved and in-situ spectra. The efficiency of spectral shape matching has also been explored, using SeaWiFS imagery for turbid waters and compared with that of the standard SeaWiFS atmospheric correction algorithm, which falls in highly turbid waters, due to the assumption that values of water-leaving radiance in the two NIR bands are negligible to enable retrieval of aerosol reflectance in the correction of ocean color imagery. Validation suggests that accurate the retrieval of water-leaving radiance is not feasible with the invalid assumption of the classical algorithms, but is feasible with SSMM.

A Numerical Study on the Optimization of Urea Solution Injection to Maximize Conversion Efficiency of NH3 (NH3 전환효율 극대화를 위한 Urea 인젝터의 분사 최적화에 관한 수치적 연구)

  • Moon, Seongjoon;Jo, Nakwon;Oh, Sedoo;Jeong, Soojin;Park, Kyoungwoo
    • Transactions of the Korean Society of Automotive Engineers
    • /
    • v.22 no.3
    • /
    • pp.171-178
    • /
    • 2014
  • From now on, in order to meet more stringer diesel emission standard, diesel vehicle should be equipped with emission after-treatment devices as NOx reduction catalyst and particulate filters. Urea-SCR is being developed as the most efficient method of reducing NOx emissions in the after-treatment devices of diesel engines, and recent studies have begun to mount the urea-SCR device for diesel passenger cars and light duty vehicles. That is because their operational characteristics are quite different from heavy duty vehicles, urea solution injection should be changed with other conditions. Therefore, the number and diameter of the nozzle, injection directions, mounting positions in front of the catalytic converter are important design factors. In this study, major design parameters concerning urea solution injection in front of SCR are optimized by using a CFD analysis and Taguchi method. The computational prediction of internal flow and spray characteristics in front of SCR was carried out by using STAR-CCM+7.06 code that used to evaluate $NH_3$ uniformity index($NH_3$ UI). The design parameters are optimized by using the $L_{16}$ orthogonal array and small-the-better characteristics of the Taguchi method. As a result, the optimal values are confirmed to be valid in 95% confidence and 5% significance level through analysis of variance(ANOVA). The compared maximize $NH_3$ UI and activation time($NH_3$ UI 0.82) are numerically confirmed that the optimal model provides better conversion efficiency of $NH_3$. In addition, we propose a method to minimize wall-wetting around the urea injector in order to prevent injector blocks caused by solid urea loading. Consequently, the thickness reduction of fluid film in front of mixer is numerically confirmed through the mounting mixer and correcting injection direction by using the trial and error method.

Wyner-Ziv Video Compression using Noise Model Selection (잡음 모델 선택을 이용한 Wyner-Ziv 비디오 압축)

  • Park, Chun-Ho;Shim, Hiuk-Jae;Jeon, Byeung-Woo
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.46 no.4
    • /
    • pp.58-66
    • /
    • 2009
  • Recently the emerging demands of the light-video encoder promotes lots of research efforts on DVC (Distributed Video Coding). As an appropriate video compression method, DVC has been studied, and Wyner-Ziv (WZ) video compression is its one representative structure. The WZ encoder splits the image into two kinds of frames, one is key frame which is compressed by conventional intra coding, and the other is WZ frame which is encoded by WZ coding. The WZ decoder decodes the key frame first, and estimates the WZ frame using temporal correlation between key frames. Estimated WZ frame (Side Information) cannot be the same as the original WZ frame due to the absence of the WZ frame information at decoder. As a result, the difference between the estimated and original WZ frames are regarded as virtual channel noise. The WZ frame is reconstructed by removing noise in side information. Therefore precise noise estimation produces good performance gain in WZ video compression by improving error correcting capability by channel code. But noise cannot be estimated precisely at WZ decoder unless there is good WZ frame information, and generally it is estimated from the difference of corresponding key frames. Also the estimated noise is limited by comparing with frame level noise to reduce the uncertainty of the estimation method. However these methods cannot provide good noise estimation for every frame or each bit plane. In this paper, we propose a noise nodel selection method which chooses a better noise model for each bit plane after generating candidate noise models. Experimental result shows PSNR gain up to 0.8 dB.

Corporate Bond Rating Using Various Multiclass Support Vector Machines (다양한 다분류 SVM을 적용한 기업채권평가)

  • Ahn, Hyun-Chul;Kim, Kyoung-Jae
    • Asia pacific journal of information systems
    • /
    • v.19 no.2
    • /
    • pp.157-178
    • /
    • 2009
  • Corporate credit rating is a very important factor in the market for corporate debt. Information concerning corporate operations is often disseminated to market participants through the changes in credit ratings that are published by professional rating agencies, such as Standard and Poor's (S&P) and Moody's Investor Service. Since these agencies generally require a large fee for the service, and the periodically provided ratings sometimes do not reflect the default risk of the company at the time, it may be advantageous for bond-market participants to be able to classify credit ratings before the agencies actually publish them. As a result, it is very important for companies (especially, financial companies) to develop a proper model of credit rating. From a technical perspective, the credit rating constitutes a typical, multiclass, classification problem because rating agencies generally have ten or more categories of ratings. For example, S&P's ratings range from AAA for the highest-quality bonds to D for the lowest-quality bonds. The professional rating agencies emphasize the importance of analysts' subjective judgments in the determination of credit ratings. However, in practice, a mathematical model that uses the financial variables of companies plays an important role in determining credit ratings, since it is convenient to apply and cost efficient. These financial variables include the ratios that represent a company's leverage status, liquidity status, and profitability status. Several statistical and artificial intelligence (AI) techniques have been applied as tools for predicting credit ratings. Among them, artificial neural networks are most prevalent in the area of finance because of their broad applicability to many business problems and their preeminent ability to adapt. However, artificial neural networks also have many defects, including the difficulty in determining the values of the control parameters and the number of processing elements in the layer as well as the risk of over-fitting. Of late, because of their robustness and high accuracy, support vector machines (SVMs) have become popular as a solution for problems with generating accurate prediction. An SVM's solution may be globally optimal because SVMs seek to minimize structural risk. On the other hand, artificial neural network models may tend to find locally optimal solutions because they seek to minimize empirical risk. In addition, no parameters need to be tuned in SVMs, barring the upper bound for non-separable cases in linear SVMs. Since SVMs were originally devised for binary classification, however they are not intrinsically geared for multiclass classifications as in credit ratings. Thus, researchers have tried to extend the original SVM to multiclass classification. Hitherto, a variety of techniques to extend standard SVMs to multiclass SVMs (MSVMs) has been proposed in the literature Only a few types of MSVM are, however, tested using prior studies that apply MSVMs to credit ratings studies. In this study, we examined six different techniques of MSVMs: (1) One-Against-One, (2) One-Against-AIL (3) DAGSVM, (4) ECOC, (5) Method of Weston and Watkins, and (6) Method of Crammer and Singer. In addition, we examined the prediction accuracy of some modified version of conventional MSVM techniques. To find the most appropriate technique of MSVMs for corporate bond rating, we applied all the techniques of MSVMs to a real-world case of credit rating in Korea. The best application is in corporate bond rating, which is the most frequently studied area of credit rating for specific debt issues or other financial obligations. For our study the research data were collected from National Information and Credit Evaluation, Inc., a major bond-rating company in Korea. The data set is comprised of the bond-ratings for the year 2002 and various financial variables for 1,295 companies from the manufacturing industry in Korea. We compared the results of these techniques with one another, and with those of traditional methods for credit ratings, such as multiple discriminant analysis (MDA), multinomial logistic regression (MLOGIT), and artificial neural networks (ANNs). As a result, we found that DAGSVM with an ordered list was the best approach for the prediction of bond rating. In addition, we found that the modified version of ECOC approach can yield higher prediction accuracy for the cases showing clear patterns.