• Title/Summary/Keyword: Data Normalization

Search Result 481, Processing Time 0.033 seconds

Study on Prediction of Internal Quality of Cherry Tomato using Vis/NIR Spectroscopy (가시광 및 근적외선 분광기법을 이용한 방울토마토의 내부품질 예측에 관한 연구)

  • Kim, Dae-Yong;Cho, Byoung-Kwan;Mo, Chang-Yeun;Kim, Young-Sik
    • Journal of Biosystems Engineering
    • /
    • v.35 no.6
    • /
    • pp.450-457
    • /
    • 2010
  • Although cherry tomato is one of major vegetables consumed in fresh vegetable market, the quality grading method is mostly dependant on size measurement using drum shape sorting machines. Using Visible/Near-infrared spectroscopy, apparatus to be able to acquire transmittance spectrum data was made and used to estimate firmness, sugar content, and acidity of cherry tomatoes grown at hydroponic and soil culture. Partial least square (PLS) models were performed to predict firmness, sugar content, and acidity for the acquired transmittance spectra. To enhance accuracy of the PLS models, several preprocessing methods were carried out, such as normalization, multiplicative scatter correction (MSC), standard normal variate (SNV), and derivatives, etc. The coefficient of determination ($R^2_p$) and standard error of prediction (SEP) for the prediction of firmness, sugar, and acidity of cherry tomatoes from green to red ripening stages were 0.859 and 1.899 kgf, with a preprocessing of normalization, 0.790 and $0.434^{\circ}Brix$ with a preprocessing of the 1st derivative of Savitzky Golay, and 0.518 and 0.229% with a preprocessing normalization, respectively.

The Attenuation Structure of the South Korea: A review

  • Chung, T. W.;Noh, M. H.;Matsumoto, S.
    • Journal of the Korean Geophysical Society
    • /
    • v.9 no.3
    • /
    • pp.199-207
    • /
    • 2006
  • Fukuoka earthquake on March 20, 2005 showed the potential hazard of large events out of S. Korea. From the viewpoint of seismic hazard, seismic amplitude decrease Q-1 is very important. Related to the crustal cracks induced by the earthquakes, the value of Q-1- high Q-1 regions are more attenuating than low Q-1 regions - shows a correlation with seismic activity; relatively higher values of Q-1 have been observed in seismically active areas than in stable areas. For the southeastern and central S. Korea, we first simultaneously estimated QP-1 and QS-1 by applying the extended coda-normalization method to KIGAM and KNUE network data. Estimated QP-1 and QS-1 values are 0.009 f-1.05 and 0.004 f-0.70 for southeastern S. Korea and 0.003 f -0.54 and 0.003 f -0.42 for central S. Korea, respectively. These values agree with those of seismically inactive regions such as shield. The low QLg-1 value, 0.0018f -0.54 was also obtained by the coda normalization method. In addition, we studied QLg-1 by applying the source pair/receiver pair (SPRP) method to both domestic and far-regional events. The obtained QLg-1 for all Fc is less than 0.002, which is reasonable value for a seismically inactive region.

  • PDF

Improving methods for normalizing biomedical text entities with concepts from an ontology with (almost) no training data at BLAH5 the CONTES

  • Ferre, Arnaud;Ba, Mouhamadou;Bossy, Robert
    • Genomics & Informatics
    • /
    • v.17 no.2
    • /
    • pp.20.1-20.5
    • /
    • 2019
  • Entity normalization, or entity linking in the general domain, is an information extraction task that aims to annotate/bind multiple words/expressions in raw text with semantic references, such as concepts of an ontology. An ontology consists minimally of a formally organized vocabulary or hierarchy of terms, which captures knowledge of a domain. Presently, machine-learning methods, often coupled with distributional representations, achieve good performance. However, these require large training datasets, which are not always available, especially for tasks in specialized domains. CONTES (CONcept-TErm System) is a supervised method that addresses entity normalization with ontology concepts using small training datasets. CONTES has some limitations, such as it does not scale well with very large ontologies, it tends to overgeneralize predictions, and it lacks valid representations for the out-of-vocabulary words. Here, we propose to assess different methods to reduce the dimensionality in the representation of the ontology. We also propose to calibrate parameters in order to make the predictions more accurate, and to address the problem of out-of-vocabulary words, with a specific method.

Performance Evaluation of Various Normalization Methods and Score-level Fusion Algorithms for Multiple-Biometric System (다중 생체 인식 시스템을 위한 정규화함수와 결합알고리즘의 성능 평가)

  • Woo Na-Young;Kim Hak-Il
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.16 no.3
    • /
    • pp.115-127
    • /
    • 2006
  • The purpose of this paper is evaluation of various normalization methods and fusion algorithms in addition to pattern classification algorithms for multi-biometric systems. Experiments are performed using various normalization functions, fusion algorithms and pattern classification algorithms based on Biometric Scores Set-Releasel(BSSR1) provided by NIST. The performance results are presented by Half Total Error Rate (WTER). This study gives base data for the study on performance enhancement of multiple-biometric system by showing performance results using single database and metrics.

A Study on Object Based Image Analysis Methods for Land Use and Land Cover Classification in Agricultural Areas (변화지역 탐지를 위한 시계열 KOMPSAT-2 다중분광 영상의 MAD 기반 상대복사 보정에 관한 연구)

  • Yeon, Jong-Min;Kim, Hyun-Ok;Yoon, Bo-Yeol
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.15 no.3
    • /
    • pp.66-80
    • /
    • 2012
  • It is necessary to normalize spectral image values derived from multi-temporal satellite data to a common scale in order to apply remote sensing methods for change detection, disaster mapping, crop monitoring and etc. There are two main approaches: absolute radiometric normalization and relative radiometric normalization. This study focuses on the multi-temporal satellite image processing by the use of relative radiometric normalization. Three scenes of KOMPSAT-2 imagery were processed using the Multivariate Alteration Detection(MAD) method, which has a particular advantage of selecting PIFs(Pseudo Invariant Features) automatically by canonical correlation analysis. The scenes were then applied to detect disaster areas over Sendai, Japan, which was hit by a tsunami on 11 March 2011. The case study showed that the automatic extraction of changed areas after the tsunami using relatively normalized satellite data via the MAD method was done within a high accuracy level. In addition, the relative normalization of multi-temporal satellite imagery produced better results to rapidly map disaster-affected areas with an increased confidence level.

A Desirability Function-Based Multi-Characteristic Robust Design Optimization Technique (호감도 함수 기반 다특성 강건설계 최적화 기법)

  • Jong Pil Park;Jae Hun Jo;Yoon Eui Nahm
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.46 no.4
    • /
    • pp.199-208
    • /
    • 2023
  • Taguchi method is one of the most popular approaches for design optimization such that performance characteristics become robust to uncontrollable noise variables. However, most previous Taguchi method applications have addressed a single-characteristic problem. Problems with multiple characteristics are more common in practice. The multi-criteria decision making(MCDM) problem is to select the optimal one among multiple alternatives by integrating a number of criteria that may conflict with each other. Representative MCDM methods include TOPSIS(Technique for Order of Preference by Similarity to Ideal Solution), GRA(Grey Relational Analysis), PCA(Principal Component Analysis), fuzzy logic system, and so on. Therefore, numerous approaches have been conducted to deal with the multi-characteristic design problem by combining original Taguchi method and MCDM methods. In the MCDM problem, multiple criteria generally have different measurement units, which means that there may be a large difference in the physical value of the criteria and ultimately makes it difficult to integrate the measurements for the criteria. Therefore, the normalization technique is usually utilized to convert different units of criteria into one identical unit. There are four normalization techniques commonly used in MCDM problems, including vector normalization, linear scale transformation(max-min, max, or sum). However, the normalization techniques have several shortcomings and do not adequately incorporate the practical matters. For example, if certain alternative has maximum value of data for certain criterion, this alternative is considered as the solution in original process. However, if the maximum value of data does not satisfy the required degree of fulfillment of designer or customer, the alternative may not be considered as the solution. To solve this problem, this paper employs the desirability function that has been proposed in our previous research. The desirability function uses upper limit and lower limit in normalization process. The threshold points for establishing upper or lower limits let us know what degree of fulfillment of designer or customer is. This paper proposes a new design optimization technique for multi-characteristic design problem by integrating the Taguchi method and our desirability functions. Finally, the proposed technique is able to obtain the optimal solution that is robust to multi-characteristic performances.

A note on Box-Cox transformation and application in microarray data

  • Rahman, Mezbahur;Lee, Nam-Yong
    • Journal of the Korean Data and Information Science Society
    • /
    • v.22 no.5
    • /
    • pp.967-976
    • /
    • 2011
  • The Box-Cox transformation is a well known family of power transformations that brings a set of data into agreement with the normality assumption of the residuals and hence the response variable of a postulated model in regression analysis. Normalization (studentization) of the regressors is a common practice in analyzing microarray data. Here, we implement Box-Cox transformation in normalizing regressors in microarray data. Pridictabilty of the model can be improved using data transformation compared to studentization.

An Algorithm for Baseline Correction of SELDI/MALDI Mass Spectrometry Data

  • Lee, Kyeong-Eun
    • Journal of the Korean Data and Information Science Society
    • /
    • v.17 no.4
    • /
    • pp.1289-1297
    • /
    • 2006
  • Before other statistical data analysis the preprocessing steps should be performed adequately to have meaningful results. These steps include processes such as baseline correction, normalization, denoising, and multiple alignment. In this paper an algorithm for baseline correction is proposed with using the piecewise cubic Hermite interpolation with block-selected points and local minima after denoising for SELDI or MALDI mass spectrometry data.

  • PDF

Quantization Method for Normalization of JPEG Pleno Hologram (JPEG Pleno 홀로그램 데이터의 정규화를 위한 양자화)

  • Kim, Kyung-Jin;Kim, Jin-Kyum;Oh, Kwan-Jung;Kim, Jin-Woong;Kim, Dong-Wook;Seo, Young-Ho
    • Journal of Broadcast Engineering
    • /
    • v.25 no.4
    • /
    • pp.587-597
    • /
    • 2020
  • In this paper, we analyze the normalization that occurs when processing digital hologram and propose an optimized quantization method. In JPEG Pleno, which standardizes the compression of holograms, full complex holograms are defined as complex numbers with 32-bit or 64-bit precision, and the range of values varies greatly depending on the method of hologram generation and object type. Such data with high precision and wide dynamic range are converted to fixed-point or integer numbers with lower precision for signal processing and compression. In addition, in order to reconstruct the hologram to the SLM (spatial light modulator), it is approximated with a precision of a value that can be expressed by the pixels of the SLM. This process can be refereed as a normalization process using quantization. In this paper, we introduce a method for normalizing high precision and wide range hologram using quantization technique and propose an optimized method.

A Study on Appearance-Based Facial Expression Recognition Using Active Shape Model (Active Shape Model을 이용한 외형기반 얼굴표정인식에 관한 연구)

  • Kim, Dong-Ju;Shin, Jeong-Hoon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.1
    • /
    • pp.43-50
    • /
    • 2016
  • This paper introduces an appearance-based facial expression recognition method using ASM landmarks which is used to acquire a detailed face region. In particular, EHMM-based algorithm and SVM classifier with histogram feature are employed to appearance-based facial expression recognition, and performance evaluation of proposed method was performed with CK and JAFFE facial expression database. In addition, performance comparison was achieved through comparison with distance-based face normalization method and a geometric feature-based facial expression approach which employed geometrical features of ASM landmarks and SVM algorithm. As a result, the proposed method using ASM-based face normalization showed performance improvements of 6.39% and 7.98% compared to previous distance-based face normalization method for CK database and JAFFE database, respectively. Also, the proposed method showed higher performance compared to geometric feature-based facial expression approach, and we confirmed an effectiveness of proposed method.