• Title/Summary/Keyword: word context

Search Result 350, Processing Time 0.027 seconds

SG-Drop: Faster Skip-Gram by Dropping Context Words

  • Kim, DongJae;Synn, DoangJoo;Kim, Jong-Kook
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2020.11a
    • /
    • pp.1014-1017
    • /
    • 2020
  • Many natural language processing (NLP) models utilize pre-trained word embeddings to leverage latent information. One of the most successful word embedding model is the Skip-gram (SG). In this paper, we propose a Skipgram drop (SG-Drop) model, which is a variation of the SG model. The SG-Drop model is designed to reduce training time efficiently. Furthermore, the SG-Drop allows controlling training time with its hyperparameter. It could train word embedding faster than reducing training epochs while better preserving the quality.

A Study on Enhancing Emotional Engagement in Learning Situation - Based on Development Case of English Learning Serious Game 'Word Collectrian' (학습 장면에서 감정 개입을 촉진하기 위한 기능성 게임의 활용 - 단어 시각화 기반의 영어 학습용 기능성 게임 '워드 콜렉트리안' 제작 사례를 바탕으로)

  • Lee, Haksu;Doh, Young Yim
    • Journal of Korea Game Society
    • /
    • v.12 no.6
    • /
    • pp.95-106
    • /
    • 2012
  • Emotion is very important feature in educational situation. Because it has high influence to memory, educational achievement, motivation. This study tried to find out possibility of serious game as emotional engagement tool in educational situation. We did our pilot experiment to elementary school students who are english as second language. In this L2 learning situation, we did our basic experiment with English language learning serious game called 'Word Collectrian". Word Collectrian has some features for emotional engagement. It has interaction for dynamic word visualization, providing context video for word usage, putting visualized word on learner's virtual home. According to experimental result, word Collectrian has possibility for educational achievement and emotional engagement effect.

A Study on Pseudo N-gram Language Models for Speech Recognition (음성인식을 위한 의사(疑似) N-gram 언어모델에 관한 연구)

  • 오세진;황철준;김범국;정호열;정현열
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.2 no.3
    • /
    • pp.16-23
    • /
    • 2001
  • In this paper, we propose the pseudo n-gram language models for speech recognition with middle size vocabulary compared to large vocabulary speech recognition using the statistical n-gram language models. The proposed method is that it is very simple method, which has the standard structure of ARPA and set the word probability arbitrary. The first, the 1-gram sets the word occurrence probability 1 (log likelihood is 0.0). The second, the 2-gram also sets the word occurrence probability 1, which can only connect the word start symbol and WORD, WORD and the word end symbol . Finally, the 3-gram also sets the ward occurrence probability 1, which can only connect the word start symbol , WORD and the word end symbol . To verify the effectiveness of the proposed method, the word recognition experiments are carried out. The preliminary experimental results (off-line) show that the word accuracy has average 97.7% for 452 words uttered by 3 male speakers. The on-line word recognition results show that the word accuracy has average 92.5% for 20 words uttered by 20 male speakers about stock name of 1,500 words. Through experiments, we have verified the effectiveness of the pseudo n-gram language modes for speech recognition.

  • PDF

A Study on the Precedent Factors of WOM Intention in the Context of OTT Service: Focusing on Emotional Factors (OTT 서비스 이용자의 구전의도에 영향을 미치는 요인에 관한 연구: 감정요인을 중심으로)

  • Kim, Yoo Jung
    • The Journal of Information Systems
    • /
    • v.32 no.2
    • /
    • pp.63-85
    • /
    • 2023
  • Purpose According to a recent survey, more than 90% of customers make purchase or use decisions for their products and services, primarily based on word-of-mouth(WOM) information (reviews, recommendations, ratings, etc.). However, few studies have explored what factors influence user's WOM intention in the context of OTT service. To address this gap, this study investigates the antecedent factors of WOM intention by adopting stimulus-organism-response (SOR) theory and pleasure arousal dominance (PAD) theory as a theoretical basis. Design/methodology/approach The research model consists of stimulus factors (diversity of original content, aesthetics, interactivity, personalization quality), emotional factors (pleasure, excitement, dominance), and behavioral factor (WOM intention). The research hypothesis was tested using the survey data collected from OTT service users of 305. Findings The result reveals that diversity of original content significantly affects on pleasure and arousal whereas it is not associated with dominance. Findings indicate that aesthetics positively influences pleasure, and interactivity is not associated with arousal. It is also proven that personalization quality has a positive effect on dominance. In addition, pleasure, arousal, and domination are proven to be positively and significantly related to WOM intention.

Association Modeling on Keyword and Abstract Data in Korean Port Research

  • Yoon, Hee-Young;Kwak, Il-Youp
    • Journal of Korea Trade
    • /
    • v.24 no.5
    • /
    • pp.71-86
    • /
    • 2020
  • Purpose - This study investigates research trends by searching for English keywords and abstracts in 1,511 Korean journal articles in the Korea Citation Index from the 2002-2019 period using the term "Port." The study aims to lay the foundation for a more balanced development of port research. Design/methodology - Using abstract and keyword data, we perform frequency analysis and word embedding (Word2vec). A t-SNE plot shows the main keywords extracted using the TextRank algorithm. To analyze which words were used in what context in our two nine-year subperiods (2002-2010 and 2010-2019), we use Scattertext and scaled F-scores. Findings - First, during the 18-year study period, port research has developed through the convergence of diverse academic fields, covering 102 subject areas and 219 journals. Second, our frequency analysis of 4,431 keywords in 1,511 papers shows that the words "Port" (60 times), "Port Competitiveness" (33 times), and "Port Authority" (29 times), among others, are attractive to most researchers. Third, a word embedding analysis identifies the words highly correlated with the top eight keywords and visually shows four different subject clusters in a t-SNE plot. Fourth, we use Scattertext to compare words used in the two research sub-periods. Originality/value - This study is the first to apply abstract and keyword analysis and various text mining techniques to Korean journal articles in port research and thus has important implications. Further in-depth studies should collect a greater variety of textual data and analyze and compare port studies from different countries.

Aspect-Based Sentiment Analysis with Position Embedding Interactive Attention Network

  • Xiang, Yan;Zhang, Jiqun;Zhang, Zhoubin;Yu, Zhengtao;Xian, Yantuan
    • Journal of Information Processing Systems
    • /
    • v.18 no.5
    • /
    • pp.614-627
    • /
    • 2022
  • Aspect-based sentiment analysis is to discover the sentiment polarity towards an aspect from user-generated natural language. So far, most of the methods only use the implicit position information of the aspect in the context, instead of directly utilizing the position relationship between the aspect and the sentiment terms. In fact, neighboring words of the aspect terms should be given more attention than other words in the context. This paper studies the influence of different position embedding methods on the sentimental polarities of given aspects, and proposes a position embedding interactive attention network based on a long short-term memory network. Firstly, it uses the position information of the context simultaneously in the input layer and the attention layer. Secondly, it mines the importance of different context words for the aspect with the interactive attention mechanism. Finally, it generates a valid representation of the aspect and the context for sentiment classification. The model which has been posed was evaluated on the datasets of the Semantic Evaluation 2014. Compared with other baseline models, the accuracy of our model increases by about 2% on the restaurant dataset and 1% on the laptop dataset.

Effects of Conceptual Context on Implicit Memory (의미적 맥락에 대한 처리가 암묵기억에 미치는 영향)

  • 연은경;김민식
    • Korean Journal of Cognitive Science
    • /
    • v.13 no.4
    • /
    • pp.9-21
    • /
    • 2002
  • Four experiments were conducted to examine whether maintaining the same conceptual context across study and test would affect performance on a perceptual implicit memory task. The sense-specific theory of priming (Lewandowsky et al., 1989) predicts greater priming from a match in conceptual context across study and test compared with a condition in which the conceptual context is mismatched, whereas the transfer-appropriate-processing view (e.g., Blaxton, 1989) predicts no difference. In experiment 1 and 2, little or no effect of varying context was observed on a implicit task. In experiment 3 and 4, a process-dissociation procedure (proposed by Jacoby, 1991) was used to separate automatic influences from consciously controlled influence in implicit memory, which was measured by Korean word completion task. The results showed that conceptual context effect was observed in consciously controlled parts of implicit memory. These results suggest that only consciously controlled processing parts of implicit memory is sensitive to conceptual context.

  • PDF

Query Extension of Retrieve System Using Hangul Word Embedding and Apriori (한글 워드임베딩과 아프리오리를 이용한 검색 시스템의 질의어 확장)

  • Shin, Dong-Ha;Kim, Chang-Bok
    • Journal of Advanced Navigation Technology
    • /
    • v.20 no.6
    • /
    • pp.617-624
    • /
    • 2016
  • The hangul word embedding should be performed certainly process for noun extraction. Otherwise, it should be trained words that are not necessary, and it can not be derived efficient embedding results. In this paper, we propose model that can retrieve more efficiently by query language expansion using hangul word embedded, apriori, and text mining. The word embedding and apriori is a step expanding query language by extracting association words according to meaning and context for query language. The hangul text mining is a step of extracting similar answer and responding to the user using noun extraction, TF-IDF, and cosine similarity. The proposed model can improve accuracy of answer by learning the answer of specific domain and expanding high correlation query language. As future research, it needs to extract more correlation query language by analysis of user queries stored in database.

Two Statistical Models for Automatic Word Spacing of Korean Sentences (한글 문장의 자동 띄어쓰기를 위한 두 가지 통계적 모델)

  • 이도길;이상주;임희석;임해창
    • Journal of KIISE:Software and Applications
    • /
    • v.30 no.3_4
    • /
    • pp.358-371
    • /
    • 2003
  • Automatic word spacing is a process of deciding correct boundaries between words in a sentence including spacing errors. It is very important to increase the readability and to communicate the accurate meaning of text to the reader. The previous statistical approaches for automatic word spacing do not consider the previous spacing state, and thus can not help estimating inaccurate probabilities. In this paper, we propose two statistical word spacing models which can solve the problem of the previous statistical approaches. The proposed models are based on the observation that the automatic word spacing is regarded as a classification problem such as the POS tagging. The models can consider broader context and estimate more accurate probabilities by generalizing hidden Markov models. We have experimented the proposed models under a wide range of experimental conditions in order to compare them with the current state of the art, and also provided detailed error analysis of our models. The experimental results show that the proposed models have a syllable-unit accuracy of 98.33% and Eojeol-unit precision of 93.06% by the evaluation method considering compound nouns.

Two-Path Language Modeling Considering Word Order Structure of Korean (한국어의 어순 구조를 고려한 Two-Path 언어모델링)

  • Shin, Joong-Hwi;Park, Jae-Hyun;Lee, Jung-Tae;Rim, Hae-Chang
    • The Journal of the Acoustical Society of Korea
    • /
    • v.27 no.8
    • /
    • pp.435-442
    • /
    • 2008
  • The n-gram model is appropriate for languages, such as English, in which the word-order is grammatically rigid. However, it is not suitable for Korean in which the word-order is relatively free. Previous work proposed a twoply HMM that reflected the characteristics of Korean but failed to reflect word-order structures among words. In this paper, we define a new segment unit which combines two words in order to reflect the characteristic of word-order among adjacent words that appear in verbal morphemes. Moreover, we propose a two-path language model that estimates probabilities depending on the context based on the proposed segment unit. Experimental results show that the proposed two-path language model yields 25.68% perplexity improvement compared to the previous Korean language models and reduces 94.03% perplexity for the prediction of verbal morphemes where words are combined.