• Title/Summary/Keyword: data extraction

Search Result 3,350, Processing Time 0.034 seconds

Stroke Extraction of Chinese Character using Mechanism of Optical Neural Field (시각신경 메커니즘을 이용한 한자 획의 분리 및 추출)

  • Son, Jin-U;Lee, Uk-Jae;Lee, Haeng-Se
    • The Transactions of the Korea Information Processing Society
    • /
    • v.1 no.3
    • /
    • pp.311-318
    • /
    • 1994
  • In this paper, a new stroke extraction method of Chinese character base on the human optical field(the Receptive Field of Cell) is proposed. In processing the feature extraction of the chinese character, needed are more perfect extraction methods for separated informations and its data base. This method can be applied to processing neural cell using conventional feature extraction mechanism in the optical boundary of retina and cerebrum. With this method, its applicability and effectiveness were demonstrated extracting strokes from Chinese character.

  • PDF

Effect of Salts on the Extraction Characteristics of Succinic Acid by Predispersed Solvent Extraction

  • Kim, Bong-Seock;Hong, Yeon-Ki;Hong, Won-Hi
    • Biotechnology and Bioprocess Engineering:BBE
    • /
    • v.9 no.3
    • /
    • pp.207-211
    • /
    • 2004
  • Predispersed solvent extraction (PDSE) of succinic acid with Tri-n-octylamine (TOA) dissolved in 1-octanol from aqueous solutions of 50 g/L succinic acid was examined. It was found that the equilibrium data in PDSE was equal to that in conventional solvent extraction in spite of the lack of mechanical mixing in PDSE. The influence of salts on succinic acid extraction and the stability of colloidal liquid aphrons (CLAs) were also investigated. Results indicated that in the presence of sodium chloride, less succinic acid was extracted by CLAs and the stability of CLAs decreased. However, the stability of CLAs was sufficient to make PDSE practically applicable to real fermentation broth, considering the concentration range of salts in the fermentation process for succinic acid.

Study of Nonlinear Feature Extraction for Faults Diagnosis of Rotating Machinery (회전기계의 결함진단을 위한 비선형 특징 추출 방법의 연구)

  • Widodo, Achmad;Yang, Bo-Suk
    • Proceedings of the Korean Society for Noise and Vibration Engineering Conference
    • /
    • 2005.11a
    • /
    • pp.127-130
    • /
    • 2005
  • There are many methods in feature extraction have been developed. Recently, principal components analysis (PCA) and independent components analysis (ICA) is introduced for doing feature extraction. PCA and ICA linearly transform the original input into new uncorrelated and independent features space respectively In this paper, the feasibility of using nonlinear feature extraction will be studied. This method will employ the PCA and ICA procedure and adopt the kernel trick to nonlinearly map the data into a feature space. The goal of this study is to seek effectively useful feature for faults classification.

  • PDF

Efficient extraction and recovery of Lignosulfonate using sunflower oil as green solvent in liquid membrane transport: Equilibrium and kinetic study

  • Kumar, Vikas;Singh, Raghubansh K.;Chowdhury, Pradip
    • Journal of Industrial and Engineering Chemistry
    • /
    • v.67
    • /
    • pp.109-122
    • /
    • 2018
  • This work highlights extraction and removal of Lignosulfonate using sunflower oil-Tri-n-octylamine (TOA) system in bulk liquid membrane transport. Maximum extraction and recovery percentages of 92.4% and 75.2% were achieved. Optimum manifold operating conditions were: 4 vol.% TOA, $2{\pm}0.1$ feed phase pH, 300 rpm stirring speed, at $40^{\circ}C$ with 0.2 (M) $Na_2SO_4$ solution. 1:2 (organic/aqueous) and 1:1 (aqueous/aqueous) phase ratios produced best results. Extraction (36.85 kJ/mol) was found to be intermediate controlled and stripping (54.79 kJ/mol) was chemical reaction controlled. Kinetic estimation of data with higher rate constants for stripping vis-${\grave{a}}$-vis extraction showed latter to be rate determining.

Fine-tuning BERT Models for Keyphrase Extraction in Scientific Articles

  • Lim, Yeonsoo;Seo, Deokjin;Jung, Yuchul
    • Journal of Advanced Information Technology and Convergence
    • /
    • v.10 no.1
    • /
    • pp.45-56
    • /
    • 2020
  • Despite extensive research, performance enhancement of keyphrase (KP) extraction remains a challenging problem in modern informatics. Recently, deep learning-based supervised approaches have exhibited state-of-the-art accuracies with respect to this problem, and several of the previously proposed methods utilize Bidirectional Encoder Representations from Transformers (BERT)-based language models. However, few studies have investigated the effective application of BERT-based fine-tuning techniques to the problem of KP extraction. In this paper, we consider the aforementioned problem in the context of scientific articles by investigating the fine-tuning characteristics of two distinct BERT models - BERT (i.e., base BERT model by Google) and SciBERT (i.e., a BERT model trained on scientific text). Three different datasets (WWW, KDD, and Inspec) comprising data obtained from the computer science domain are used to compare the results obtained by fine-tuning BERT and SciBERT in terms of KP extraction.

One-Class Classification Model Based on Lexical Information and Syntactic Patterns (어휘 정보와 구문 패턴에 기반한 단일 클래스 분류 모델)

  • Lee, Hyeon-gu;Choi, Maengsik;Kim, Harksoo
    • Journal of KIISE
    • /
    • v.42 no.6
    • /
    • pp.817-822
    • /
    • 2015
  • Relation extraction is an important information extraction technique that can be widely used in areas such as question-answering and knowledge population. Previous studies on relation extraction have been based on supervised machine learning models that need a large amount of training data manually annotated with relation categories. Recently, to reduce the manual annotation efforts for constructing training data, distant supervision methods have been proposed. However, these methods suffer from a drawback: it is difficult to use these methods for collecting negative training data that are necessary for resolving classification problems. To overcome this drawback, we propose a one-class classification model that can be trained without using negative data. The proposed model determines whether an input data item is included in an inner category by using a similarity measure based on lexical information and syntactic patterns in a vector space. In the experiments conducted in this study, the proposed model showed higher performance (an F1-score of 0.6509 and an accuracy of 0.6833) than a representative one-class classification model, one-class SVM(Support Vector Machine).

Investigation of Building Extraction Methodologies within the Framework of Sensory Data

  • Seo, Su-Young
    • Spatial Information Research
    • /
    • v.16 no.4
    • /
    • pp.479-488
    • /
    • 2008
  • This paper performs investigation of the state-of-the-art approaches to building extraction in terms of their sensory input data and methodologies. For the last decades, there have been many types of sensory input data introduced into the mapping science and engineering field, which are considerably diverse in aspects of spatial resolution and data processing. With the cutting-edge technology in this field, accordingly, one of the key issues in GIS is to reconstruct three -dimensional virtual models of the real world to meet the requirements occurring in spatial applications such as urban design, disaster management, and civil works. Thus, this study investigates the strengths and weaknesses of previous approaches to automating building extraction with two categories - building detection and modeling and with sensor types categorized. The findings in this study can be utilized in enhancing automation algorithms and choosing suitable sensors, so that they can be optimized for a specific purpose.

  • PDF

A Method for Compound Noun Extraction to Improve Accuracy of Keyword Analysis of Social Big Data

  • Kim, Hyeon Gyu
    • Journal of the Korea Society of Computer and Information
    • /
    • v.26 no.8
    • /
    • pp.55-63
    • /
    • 2021
  • Since social big data often includes new words or proper nouns, statistical morphological analysis methods have been widely used to process them properly which are based on the frequency of occurrence of each word. However, these methods do not properly recognize compound nouns, and thus have a problem in that the accuracy of keyword extraction is lowered. This paper presents a method to extract compound nouns in keyword analysis of social big data. The proposed method creates a candidate group of compound nouns by combining the words obtained through the morphological analysis step, and extracts compound nouns by examining their frequency of appearance in a given review. Two algorithms have been proposed according to the method of constructing the candidate group, and the performance of each algorithm is expressed and compared with formulas. The comparison result is verified through experiments on real data collected online, where the results also show that the proposed method is suitable for real-time processing.

Change Detection of Buildings Using High Resolution Remotely Sensed Data

  • Zeng, Yu;Zhang, Jixian;Wang, Guangliang
    • Proceedings of the KSRS Conference
    • /
    • 2002.10a
    • /
    • pp.530-535
    • /
    • 2002
  • An approach for quickly updating GIS building data using high resolution remotely sensed data is proposed in this paper. High resolution remotely sensed data could be aerial photographs, satellite images and airborne laser scanning data. Data from different types of sensors are integrated in building extraction. Based on the extracted buildings and the outdated GIS database, the change-detection-template can be automatically created. Then, GIS building data can be fast updated by semiautomatically processing the change-detection-temp late. It is demonstrated that this approach is quick, effective and applicable.

  • PDF

A Study on the 3-D Digital Modelling of the Sea Bottom Topography (3차원 해저지형 수치모델에 관한 연구)

  • 양승윤;김정훈;김병준;김경섭
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.5 no.3
    • /
    • pp.33-44
    • /
    • 2002
  • In this study, 3-dimensional virtual visualization was performed for a rapid and accurate analysis of sea bottom topography. The visualization was done through the extracted data using the developed program and the generated data using the gridding method. The data extraction program was developed with AutoLISP programming language and this program was able to extract the needed sample bathymetry data from the electronic sea chart systematically as well as effectively The gridded bathymetry data were generated by the interpolation or extrapolation method from the spatially-irregular sample data. As the result of realization for the 3-dimensional virtual visualization, it was shown a proper feasibility in the analysis of the sea bottom topography to determine the route of submarine cable burial.