• Title/Summary/Keyword: Syntactic Features

Search Result 93, Processing Time 0.026 seconds

An intelligent eddy current signal evaluation system to automate the non-destructive testing of steam generator tubes in nuclear power plant

  • Kang, Soon-Ju;Ryu, Chan-Ho;Choi, In-Seon;Kim, Young-Ill;Kim, kill-Yoo;Hur, Young-Hwan;Choi, Seong-Soo;Choi, Baeng-Jae;Woo, Hee-Gon
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1992.10b
    • /
    • pp.74-78
    • /
    • 1992
  • This paper describes an intelligent system to automatic evaluation of eddy current(EC) signal for Inspection of steam generator(SG) tubes in nuclear power plant. Some features of the intelligent system design in the proposed system are : (1) separation of representation scheme ,or event capturing knowledge in EC signal and for structural inspection knowledge in SG tubes inspection; (2) each representation scheme is implemented in different methods, one is syntactic pattern grammar and the other is rule based production. This intelligent system also includes an data base system and an user interface system to support integration of the hybrid knowledge processing methods. The intelligent system based on the proposed concept is useful in simplifying the knowledge elicitation process of the rule based production system, and in increasing the performance in real time signal inspection application.

  • PDF

The Interlanguage Speech Intelligibility Benefit (ISIB) of English Prosody: The Case of Focal Prominence for Korean Learners of English and Natives

  • Lee, Joo-Kyeong;Han, Jeong-Im;Choi, Tae-Hwan;Lim, Injae
    • Phonetics and Speech Sciences
    • /
    • v.4 no.4
    • /
    • pp.53-68
    • /
    • 2012
  • This study investigated the speech intelligibility of Korean-accented and native English focus speech for Korean and native English listeners. Three different types of focus in English, broad, narrow and contrastive, were naturally induced in semantically optimal dialogues. Seven high and seven low proficiency Korean speakers and seven native speakers participated in recording the stimuli with another native speaker. Fifteen listeners from each of Korean high & low proficiency and native groups judged audio signals of focus sentences. Results showed that Korean listeners were more accurate at identifying the focal prominence for Korean speakers' narrow focus speech than that of native speakers, and this suggests that the interlanguage speech intelligibility benefit-talker (ISIB-T) held true for narrow focus regardless of Korean speakers' and listeners' proficiency. However, Korean listeners did not outperform native listeners for Korean speakers' production of narrow focus, which did not support for the ISIB-listener (L). Broad and contrastive focus speech did not provide evidence for either the ISIB-T or ISIB-L. These findings are explained by the interlanguage shared by Korean speakers and listeners where they have established more L1-like common phonetic features and phonological representations. Once semantically and syntactically interpreted in a higher level processing in Korean narrow focus speech, the narrow focus was phonetically realized in a more intelligible way to Korean listeners due to the interlanguage. This may elicit ISIB. However, Korean speakers did not appear to make complete semantic/syntactic access to either broad or contrastive focus, which might lead to detrimental effects on lower level phonetic outputs in top-down processing. This is, therefore, attributed to the fact that Korean listeners did not take advantage over native listeners for Korean talkers and vice versa.

Functional Expansion of Morphological Analyzer Based on Longest Phrase Matching For Efficient Korean Parsing (효율적인 한국어 파싱을 위한 최장일치 기반의 형태소 분석기 기능 확장)

  • Lee, Hyeon-yoeng;Lee, Jong-seok;Kang, Byeong-do;Yang, Seung-weon
    • Journal of Digital Contents Society
    • /
    • v.17 no.3
    • /
    • pp.203-210
    • /
    • 2016
  • Korean is free of omission of sentence elements and modifying scope, so managing it on morphological analyzer is better than parser. In this paper, we propose functional expansion methods of the morphological analyzer to ease the burden of parsing. This method is a longest phrase matching method. When the series of several morpheme have one syntax category by processing of Unknown-words, Compound verbs, Compound nouns, Numbers and Symbols, our method combines them into a syntactic unit. And then, it is to treat by giving them a semantic features as syntax unit. The proposed morphological analysis method removes unnecessary morphological ambiguities and deceases results of morphological analysis, so improves accuracy of tagger and parser. By empirical results, we found that our method deceases 73.4% of Parsing tree and 52.4% of parsing time on average.

Ranked Web Service Retrieval by Keyword Search (키워드 질의를 이용한 순위화된 웹 서비스 검색 기법)

  • Lee, Kyong-Ha;Lee, Kyu-Chul;Kim, Kyong-Ok
    • The Journal of Society for e-Business Studies
    • /
    • v.13 no.2
    • /
    • pp.213-223
    • /
    • 2008
  • The efficient discovery of services from a large scale collection of services has become an important issue[7, 24]. We studied a syntactic method for Web service discovery, rather than a semantic method. We regarded a service discovery as a retrieval problem on the proprietary XML formats, which were service descriptions in a registry DB. We modeled services and queries as probabilistic values and devised similarity-based retrieval techniques. The benefits of our way are follows. First, our system supports ranked service retrieval by keyword search. Second, we considers both of UDDI data and WSDL definitions of services amid query evaluation time. Last, our technique can be easily implemented on the off-theshelf DBMS and also utilize good features of DBMS maintenance.

  • PDF

A Business Service Identification Techniques Based on XL-BPMN Model (XL-BPMN 모델 기반 비즈니스 서비스 식별 기법)

  • Song, Chee-Yang;Cho, Eun-Sook
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.3
    • /
    • pp.125-138
    • /
    • 2016
  • The service identification in service-oriented developments has been conducted by based on workflow, goals, scenarios, usecases, components, features, and patterns. However, the identification of service by semantic approach at the business value view was not detailed yet. In order to enhance accuracy of identifying business service, this paper proposes a method for identifying business service by analyzing syntax and semantics in XL-BPMN model. The business processes based on business scenario are identified, and they are designed in a XL-BPMN business process model. In this business process model, an unit business service is identified through binding closely related activities by the integrated analysis result of syntax patterns and properties-based semantic similarities between activities. The method through XL-BPMN model at upper business levels can identify the reusable unit business service with high accuracy and modularity. It also can accelerate more service-oriented developments by reusing identified services.

Drone Flight Record Forensic System through DUML Packet Analysis (DUML 패킷 분석을 통한 드론 비행기록 포렌식 시스템)

  • YeoHoon Yoon;Joobeom Yun
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.34 no.1
    • /
    • pp.103-114
    • /
    • 2024
  • In a situation where drone-related crimes continue to rise, research in drone forensics becomes crucial for preventing and responding to incidents involving drones. Conducting forensic analysis on flight record files stored internally is essential for investigating illegal activities. However, analyzing flight record files generated through the exclusive DUML protocol requires a deep understanding of the protocol's structure and characteristics. Additionally, a forensic analysis tool capable of handling cryptographic payloads and analyzing various drone models is imperative. Therefore, this study presents the methods and characteristics of flight record files generated by drones. It also explains the structure of the flight record file and the features of the DUML packet. Ultimately, we conduct forensic analysis based on the presented structure of the DUML packet and propose an extension forensic analysis system that operates more universally than existing tools, performing expanded syntactic analysis.

An Investigation into the Equivalence of Three Pictures for Creative Story Writing: 'Dog Owners', 'Lost Dog', and 'Overslept' (창의적 이야기 작문용 세 그림의 동형 조사: 'Dog Owners,' 'Lost Dog,' 'Overslept')

  • Suh, Heejung;Bae, Jungok
    • Journal of Gifted/Talented Education
    • /
    • v.26 no.4
    • /
    • pp.699-719
    • /
    • 2016
  • Alternate pictures that are proven to be equivalent are in high demand to assess creative thinking and language skills. This study aimed to investigate the equivalence of three pictures ('Dog owners,' 'Lost Dog,' and 'Overslept') recently developed for use in a creative writing task. Middle school students (N=183) wrote a story in English based on one of the three prompts distributed randomly. Four writing features (fluency, syntactic complexity, lexical diversity, and temporality) were analyzed with Coh-Metrix and MANCOVA. The three prompts were largely equivalent in their capacity to detect differences among writers in all the features of writing. The difficulty levels of the three prompts, however, were not necessarily the same. Two prompts, Dog Owners and Lost Dog, were verified as equivalent prompts, and therefore, they are recommended as alternate forms to assess creative language skills in repeated measurements. The Overslept prompt had greater facility in eliciting diverse words and more temporal connectives in composing stories. The differential difficulty shown among the prompts suggests that the validity of using different picture versions in repeated assessment remains questionable unless those versions undergo equivalence verification.

Selective Word Embedding for Sentence Classification by Considering Information Gain and Word Similarity (문장 분류를 위한 정보 이득 및 유사도에 따른 단어 제거와 선택적 단어 임베딩 방안)

  • Lee, Min Seok;Yang, Seok Woo;Lee, Hong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.4
    • /
    • pp.105-122
    • /
    • 2019
  • Dimensionality reduction is one of the methods to handle big data in text mining. For dimensionality reduction, we should consider the density of data, which has a significant influence on the performance of sentence classification. It requires lots of computations for data of higher dimensions. Eventually, it can cause lots of computational cost and overfitting in the model. Thus, the dimension reduction process is necessary to improve the performance of the model. Diverse methods have been proposed from only lessening the noise of data like misspelling or informal text to including semantic and syntactic information. On top of it, the expression and selection of the text features have impacts on the performance of the classifier for sentence classification, which is one of the fields of Natural Language Processing. The common goal of dimension reduction is to find latent space that is representative of raw data from observation space. Existing methods utilize various algorithms for dimensionality reduction, such as feature extraction and feature selection. In addition to these algorithms, word embeddings, learning low-dimensional vector space representations of words, that can capture semantic and syntactic information from data are also utilized. For improving performance, recent studies have suggested methods that the word dictionary is modified according to the positive and negative score of pre-defined words. The basic idea of this study is that similar words have similar vector representations. Once the feature selection algorithm selects the words that are not important, we thought the words that are similar to the selected words also have no impacts on sentence classification. This study proposes two ways to achieve more accurate classification that conduct selective word elimination under specific regulations and construct word embedding based on Word2Vec embedding. To select words having low importance from the text, we use information gain algorithm to measure the importance and cosine similarity to search for similar words. First, we eliminate words that have comparatively low information gain values from the raw text and form word embedding. Second, we select words additionally that are similar to the words that have a low level of information gain values and make word embedding. In the end, these filtered text and word embedding apply to the deep learning models; Convolutional Neural Network and Attention-Based Bidirectional LSTM. This study uses customer reviews on Kindle in Amazon.com, IMDB, and Yelp as datasets, and classify each data using the deep learning models. The reviews got more than five helpful votes, and the ratio of helpful votes was over 70% classified as helpful reviews. Also, Yelp only shows the number of helpful votes. We extracted 100,000 reviews which got more than five helpful votes using a random sampling method among 750,000 reviews. The minimal preprocessing was executed to each dataset, such as removing numbers and special characters from text data. To evaluate the proposed methods, we compared the performances of Word2Vec and GloVe word embeddings, which used all the words. We showed that one of the proposed methods is better than the embeddings with all the words. By removing unimportant words, we can get better performance. However, if we removed too many words, it showed that the performance was lowered. For future research, it is required to consider diverse ways of preprocessing and the in-depth analysis for the co-occurrence of words to measure similarity values among words. Also, we only applied the proposed method with Word2Vec. Other embedding methods such as GloVe, fastText, ELMo can be applied with the proposed methods, and it is possible to identify the possible combinations between word embedding methods and elimination methods.

Time Adverb 'Cengjing (曾經)' and 'Yijing (已經) Tense and Aspect of the Comparative Analysis of the Characteristics of China and South Korea (시간부사 '증경(曾經)', '이경(已經)' 시상(時相) 자질 중한 대조분석)

  • Han, Keung-Shuk
    • Cross-Cultural Studies
    • /
    • v.42
    • /
    • pp.451-474
    • /
    • 2016
  • Analysis of the syntactic structure of the modern Chinese adverbs for time 'Cengjing (曾經)' and 'Yijing (已經)' was performed to examine the tenses and aspects of the terms. The corresponding Korean words were examined and the terms in both languages were compared. The syntactic structures used in China and South Korea were found to be different. We hope the study of the Chinese language will help Korean students. 'Cengjing (曾經)' specific projects with 'aspect' of, 'Past experience aspect', 'Past continuous aspect', 'Past continuous aspect' in the past tense. [ED: unclear, please reword] These correspond to '_었 (았)_', '_었었_' in the Korean language. 'Yijing (已經)' has 'finished phase' of concrete projects, 'Past experience aspect', 'Past continuous aspect', also has a specific project tense, the 'past tense', 'present tense', 'future tense', and so tense. [ED: unclear, please reword] Adjectives can also be modified with a 'change of status'. These correspond to '_었 (았)_', '_고_', '_었었_', '곧' etc. in Korean. 'Cengjing (曾經)' and the dynamic auxiliary 'Guo (過)' were compared to determine whether they have the aspect and tense features. However, 'Guo (過)' can only modify the predicate verb, so it possesses only aspect characteristics. 'Cengjing (曾經)' modifies the range more widely. 'Yijing (已經)' may be modified by the adverb 'Zai (在)' whereas 'Cengjing (曾經)' may not. Additionally, 'Yijing (已經)' can be modified by predicate adjectives and noun predicates, while 'Cengjing (曾經)' cannot.

Sentiment Classification considering Korean Features (한국어 특성을 고려한 감성 분류)

  • Kim, Jung-Ho;Kim, Myung-Kyu;Cha, Myung-Hoon;In, Joo-Ho;Chae, Soo-Hoan
    • Science of Emotion and Sensibility
    • /
    • v.13 no.3
    • /
    • pp.449-458
    • /
    • 2010
  • As occasion demands to obtain efficient information from many documents and reviews on the Internet in many kinds of fields, automatic classification of opinion or thought is required. These automatic classification is called sentiment classification, which can be divided into three steps, such as subjective expression classification to extract subjective sentences from documents, sentiment classification to classify whether the polarity of documents is positive or negative, and strength classification to classify whether the documents have weak polarity or strong polarity. The latest studies in Opinion Mining have used N-gram words, lexical phrase pattern, and syntactic phrase pattern, etc. They have not used single word as feature for classification. Especially, patterns have been used frequently as feature because they are more flexible than N-gram words and are also more deterministic than single word. Theses studies are mainly concerned with English, other studies using patterns for Korean are still at an early stage. Although Korean has a slight difference in the meaning between predicates by the change of endings, which is 'Eomi' in Korean, of declinable words, the earlier studies about Korean opinion classification removed endings from predicates only to extract stems. Finally, this study introduces the earlier studies and methods using pattern for English, uses extracted sentimental patterns from Korean documents, and classifies polarities of these documents. In this paper, it also analyses the influence of the change of endings on performances of opinion classification.

  • PDF