• Title/Summary/Keyword: Syntactic Parser

Search Result 35, Processing Time 0.021 seconds

A Study on Automatic Classification of Fingerprint Images (지문 영상의 자동 분류에 관한 연구)

  • Lim, In-Sic;Sin, Tae-Min;Park, Goo-Man;Lee, Byeong-Rae;Park, Kyu-Tae
    • Proceedings of the KIEE Conference
    • /
    • 1988.07a
    • /
    • pp.628-631
    • /
    • 1988
  • This paper describes a fingerprint classification on the basis of feature points(whorl, core) and feature vector and uses a syntactic approach to identify the shape of flow line around the core. Fingerprint image is divided into 8 by 8 subregions and fingerprint region is separated from background. For each subregion of fingerprint region, the dominant ridge direction is obtained to use the slit window quantized in 8 direction and relaxation is performed to correct ridge direction code. Feature points(whorl, core, delta) are found from the ridge direction code. First classification procedure divides the types of fingerprint into 4 class based on whorl and cores. The shape of flow line around the core is obtained by tracing for the fingerprint which has one core or two core and is represented as string. If the string is acceptable by LR(1) parser, feature vector is obtained from feature points(whorl, core, delta) and the shape of flow line around the core. Feature vector is used hierarchically and linearly to classify fingerprint again. The experiment resulted in 97.3 percentages of sucessful classification for 71 fingerprint impressions.

  • PDF

Linguistic Productivity and Chomskyan Grammar: A Critique (언어창조성과 춈스키 문법 비판)

  • Bong-rae Seok
    • Lingua Humanitatis
    • /
    • v.1 no.1
    • /
    • pp.235-251
    • /
    • 2001
  • According to Chomskyan grammar, humans can generate and understand an unbounded number of grammatical sentences. Against the background of pure and idealized linguistic competence, this linguistic productivity is argued and understood. In actual utterances, however, there are many limitations of productivity but they are said to come from the general constraints on performances such as capacity of short term memory or attention. In this paper I discuss a problem raised against idealized productivity. I argue that linguistic productivity idealizes our linguistic competence too much. By separating idealized competence from the various constraints of performance, Chomskyan theorists can argue for unlimited productivity. However, the absolute distinction between grammar (pure competence) and parser (actual psychological processes) makes little sense when we explain the low acceptability(intelligibility) of center embedded sentences. Usually, the problem of center embedded sentence is explained in terms of memory shortage or other performance constraints. To explain the low acceptability, however, we need to assume specialized memory structure because the low acceptability occurs only with a specific type of syntactic pattern. 1 argue that this special memory structure should not be considered as a general performance constraint. It is a domain specific (specifically linguistic) constraints and an intrinsic part of human language processing. Recent development of Chomskyan grammar, i.e., minimalist approach seems to close the gap between pure competence and this type of specialized constraints. Chomsky's earlier approach of generative grammar focuses on end result of the generative derivation. However, economy principle (of minimalist approach) focuses on actual derivational processes. By having less mathematical or less idealized grammar, we can come closer to the actual computational processes that build syntactic structure of a sentence. In this way, we can have a more concrete picture of our linguistic competence, competence that is not detached from actual computational processes.

  • PDF

A Word Sense Disambiguation Method with a Semantic Network (의미네트워크를 이용한 단어의미의 모호성 해결방법)

  • DingyulRa
    • Korean Journal of Cognitive Science
    • /
    • v.3 no.2
    • /
    • pp.225-248
    • /
    • 1992
  • In this paper, word sense disambiguation methods utilizing a knowledge base based on a semantic network are introduced. The basic idea is to keep track of a set of paths in the knowledge base which correspond to the inctemental semantic interpretation of a input sentence. These paths are called the semantic paths. when the parser reads a word, the senses of this word which are not involved in any of the semantic paths are removed. Then the removal operation is propagated through the knowledge base to invoke the removal of the senses of other words that have been read before. This removal operation is called recusively as long as senses can be removed. This is called the recursive word sense removal. Concretion of a vague word's concept is one of the important word sense disambiguation methods. We introduce a method called the path adjustment that extends the conctetion operation. How to use semantic association or syntactic processing in coorporation with the above methods is also considered.

PPEditor: Semi-Automatic Annotation Tool for Korean Dependency Structure (PPEditor: 한국어 의존구조 부착을 위한 반자동 말뭉치 구축 도구)

  • Kim Jae-Hoon;Park Eun-Jin
    • The KIPS Transactions:PartB
    • /
    • v.13B no.1 s.104
    • /
    • pp.63-70
    • /
    • 2006
  • In general, a corpus contains lots of linguistic information and is widely used in the field of natural language processing and computational linguistics. The creation of such the corpus, however, is an expensive, labor-intensive and time-consuming work. To alleviate this problem, annotation tools to build corpora with much linguistic information is indispensable. In this paper, we design and implement an annotation tool for establishing a Korean dependency tree-tagged corpus. The most ideal way is to fully automatically create the corpus without annotators' interventions, but as a matter of fact, it is impossible. The proposed tool is semi-automatic like most other annotation tools and is designed to edit errors, which are generated by basic analyzers like part-of-speech tagger and (partial) parser. We also design it to avoid repetitive works while editing the errors and to use it easily and friendly. Using the proposed annotation tool, 10,000 Korean sentences containing over 20 words are annotated with dependency structures. For 2 months, eight annotators have worked every 4 hours a day. We are confident that we can have accurate and consistent annotations as well as reduced labor and time.

Sentence ion : Sentence Revision with Concept ion (문장추상화 : 개념추상화를 도입한 문장교열)

  • Kim, Gon;Yang, Jaegun;Bae, Jaehak;Lee, Jonghyuk
    • The KIPS Transactions:PartB
    • /
    • v.11B no.5
    • /
    • pp.563-572
    • /
    • 2004
  • Sentence ion is a simplification of a sentence preserving its communicative function. It accomplishes sentence revision and concept ion simultaneously. Sentence revision is a method that resolves the discrepancy between human's thoughts and its expressed semantic in sentences. Concept ion is an expression of general ideas acquired from the common elements of concepts. Sentence ion selects the main constituents of given sentences and describes the upper concepts of them with detecting their semantic information. This enables sen fence revision and concept ion simultaneously. In this paper, a syntactic parser LGPI+ and an ontology OfN are utilized for sentence ion. Sentence abstracter SABOT makes use of LGPI+ and OfN. SABOT processes the result of parsing and selects the candidate words for sentence ion. This paper computes the sentence recall of the main sentences and the topic hit ratio of the selected sentences with the text understanding system using sentence ion. The sources are 58 paragraphs in 23 stories. As a result of it, the sentence recall is about .54 ~ 72% and the topic hit ratio is about 76 ~ 86%. This paper verified that sentence ion enables sentence revision that can select the topic sentences of a given text efficiently and concept ion that can improve the depth of text understanding.