• Title/Summary/Keyword: VEC Model

Search Result 150, Processing Time 0.026 seconds

Multi-Emotion Regression Model for Recognizing Inherent Emotions in Speech Data (음성 데이터의 내재된 감정인식을 위한 다중 감정 회귀 모델)

  • Moung Ho Yi;Myung Jin Lim;Ju Hyun Shin
    • Smart Media Journal
    • /
    • v.12 no.9
    • /
    • pp.81-88
    • /
    • 2023
  • Recently, communication through online is increasing due to the spread of non-face-to-face services due to COVID-19. In non-face-to-face situations, the other person's opinions and emotions are recognized through modalities such as text, speech, and images. Currently, research on multimodal emotion recognition that combines various modalities is actively underway. Among them, emotion recognition using speech data is attracting attention as a means of understanding emotions through sound and language information, but most of the time, emotions are recognized using a single speech feature value. However, because a variety of emotions exist in a complex manner in a conversation, a method for recognizing multiple emotions is needed. Therefore, in this paper, we propose a multi-emotion regression model that extracts feature vectors after preprocessing speech data to recognize complex, inherent emotions and takes into account the passage of time.

Major Class Recommendation System based on Deep learning using Network Analysis (네트워크 분석을 활용한 딥러닝 기반 전공과목 추천 시스템)

  • Lee, Jae Kyu;Park, Heesung;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.3
    • /
    • pp.95-112
    • /
    • 2021
  • In university education, the choice of major class plays an important role in students' careers. However, in line with the changes in the industry, the fields of major subjects by department are diversifying and increasing in number in university education. As a result, students have difficulty to choose and take classes according to their career paths. In general, students choose classes based on experiences such as choices of peers or advice from seniors. This has the advantage of being able to take into account the general situation, but it does not reflect individual tendencies and considerations of existing courses, and has a problem that leads to information inequality that is shared only among specific students. In addition, as non-face-to-face classes have recently been conducted and exchanges between students have decreased, even experience-based decisions have not been made as well. Therefore, this study proposes a recommendation system model that can recommend college major classes suitable for individual characteristics based on data rather than experience. The recommendation system recommends information and content (music, movies, books, images, etc.) that a specific user may be interested in. It is already widely used in services where it is important to consider individual tendencies such as YouTube and Facebook, and you can experience it familiarly in providing personalized services in content services such as over-the-top media services (OTT). Classes are also a kind of content consumption in terms of selecting classes suitable for individuals from a set content list. However, unlike other content consumption, it is characterized by a large influence of selection results. For example, in the case of music and movies, it is usually consumed once and the time required to consume content is short. Therefore, the importance of each item is relatively low, and there is no deep concern in selecting. Major classes usually have a long consumption time because they have to be taken for one semester, and each item has a high importance and requires greater caution in choice because it affects many things such as career and graduation requirements depending on the composition of the selected classes. Depending on the unique characteristics of these major classes, the recommendation system in the education field supports decision-making that reflects individual characteristics that are meaningful and cannot be reflected in experience-based decision-making, even though it has a relatively small number of item ranges. This study aims to realize personalized education and enhance students' educational satisfaction by presenting a recommendation model for university major class. In the model study, class history data of undergraduate students at University from 2015 to 2017 were used, and students and their major names were used as metadata. The class history data is implicit feedback data that only indicates whether content is consumed, not reflecting preferences for classes. Therefore, when we derive embedding vectors that characterize students and classes, their expressive power is low. With these issues in mind, this study proposes a Net-NeuMF model that generates vectors of students, classes through network analysis and utilizes them as input values of the model. The model was based on the structure of NeuMF using one-hot vectors, a representative model using data with implicit feedback. The input vectors of the model are generated to represent the characteristic of students and classes through network analysis. To generate a vector representing a student, each student is set to a node and the edge is designed to connect with a weight if the two students take the same class. Similarly, to generate a vector representing the class, each class was set as a node, and the edge connected if any students had taken the classes in common. Thus, we utilize Node2Vec, a representation learning methodology that quantifies the characteristics of each node. For the evaluation of the model, we used four indicators that are mainly utilized by recommendation systems, and experiments were conducted on three different dimensions to analyze the impact of embedding dimensions on the model. The results show better performance on evaluation metrics regardless of dimension than when using one-hot vectors in existing NeuMF structures. Thus, this work contributes to a network of students (users) and classes (items) to increase expressiveness over existing one-hot embeddings, to match the characteristics of each structure that constitutes the model, and to show better performance on various kinds of evaluation metrics compared to existing methodologies.

Backward Path Following Under a Strong Headwind for UAV (강한 맞바람이 발생 했을 때 무인기의 후진경로추종에 관한 연구)

  • Byeon, Gwang-Yeol;Park, Sanghyuk
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.42 no.5
    • /
    • pp.376-382
    • /
    • 2014
  • This paper presents a method to enable a UAV in autonomous flight along a desired path to follow it backwards when a strong headwind prevents the vehicle from proceeding forward. The main purpose of the reverse path following in this study is to return to a mission quickly when the wind becomes weaker. When the nonlinear path following guidance law is used, there are two reference points available in the path following. One of the two points is selected considering a flight direction for calculating a straight-line distance(L) from the vehicle to the point for the path following. An initial heading angle with respect to the wind direction determines whether the reverse path following is feasible or not at the time of the wind is generated. The result of the proposed method based on kinematic model in this study is verified through simulations implemented in Matlab.

An Analysis on the Causality between Production Activity and Electricity Consumption in Manufacturing Sector (제조업 생산활동과 전력소비 간의 인과관계 분석)

  • Lim, Jaekyu;Kim, Jong-Ik
    • Environmental and Resource Economics Review
    • /
    • v.23 no.2
    • /
    • pp.349-364
    • /
    • 2014
  • This study analyzed Granger causality between power consumption and production activity in manufacturing sector, by using error correction model. It found that there exists the connection between power consumption and production activity in manufacturing sector. By reflecting the industrial characteristics, it found not only the bilateral causality (power consumption ${\leftrightarrow}$ production activity) in power non-intensive industry, high value-added industry and low value-added industry, but also one-way causality (power consumption ${\rightarrow}$ production activity) in power-intensive industry. These results imply that power demand management policy focusing on efficiency improvement is necessary primarily to minimize negative impacts on production activity, and also stable power supply system is required to meet the increase of power demand.

Word Embeddings-Based Pseudo Relevance Feedback Using Deep Averaging Networks for Arabic Document Retrieval

  • Farhan, Yasir Hadi;Noah, Shahrul Azman Mohd;Mohd, Masnizah;Atwan, Jaffar
    • Journal of Information Science Theory and Practice
    • /
    • v.9 no.2
    • /
    • pp.1-17
    • /
    • 2021
  • Pseudo relevance feedback (PRF) is a powerful query expansion (QE) technique that prepares queries using the top k pseudorelevant documents and choosing expansion elements. Traditional PRF frameworks have robustly handled vocabulary mismatch corresponding to user queries and pertinent documents; nevertheless, expansion elements are chosen, disregarding similarity to the original query's elements. Word embedding (WE) schemes comprise techniques of significant interest concerning QE, that falls within the information retrieval domain. Deep averaging networks (DANs) defines a framework relying on average word presence passed through multiple linear layers. The complete query is understandably represented using the average vector comprising the query terms. The vector may be employed for determining expansion elements pertinent to the entire query. In this study, we suggest a DANs-based technique that augments PRF frameworks by integrating WE similarities to facilitate Arabic information retrieval. The technique is based on the fundamental that the top pseudo-relevant document set is assessed to determine candidate element distribution and select expansion terms appropriately, considering their similarity to the average vector representing the initial query elements. The Word2Vec model is selected for executing the experiments on a standard Arabic TREC 2001/2002 set. The majority of the evaluations indicate that the PRF implementation in the present study offers a significant performance improvement compared to that of the baseline PRF frameworks.

Semantic Visualization of Dynamic Topic Modeling (다이내믹 토픽 모델링의 의미적 시각화 방법론)

  • Yeon, Jinwook;Boo, Hyunkyung;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.131-154
    • /
    • 2022
  • Recently, researches on unstructured data analysis have been actively conducted with the development of information and communication technology. In particular, topic modeling is a representative technique for discovering core topics from massive text data. In the early stages of topic modeling, most studies focused only on topic discovery. As the topic modeling field matured, studies on the change of the topic according to the change of time began to be carried out. Accordingly, interest in dynamic topic modeling that handle changes in keywords constituting the topic is also increasing. Dynamic topic modeling identifies major topics from the data of the initial period and manages the change and flow of topics in a way that utilizes topic information of the previous period to derive further topics in subsequent periods. However, it is very difficult to understand and interpret the results of dynamic topic modeling. The results of traditional dynamic topic modeling simply reveal changes in keywords and their rankings. However, this information is insufficient to represent how the meaning of the topic has changed. Therefore, in this study, we propose a method to visualize topics by period by reflecting the meaning of keywords in each topic. In addition, we propose a method that can intuitively interpret changes in topics and relationships between or among topics. The detailed method of visualizing topics by period is as follows. In the first step, dynamic topic modeling is implemented to derive the top keywords of each period and their weight from text data. In the second step, we derive vectors of top keywords of each topic from the pre-trained word embedding model. Then, we perform dimension reduction for the extracted vectors. Then, we formulate a semantic vector of each topic by calculating weight sum of keywords in each vector using topic weight of each keyword. In the third step, we visualize the semantic vector of each topic using matplotlib, and analyze the relationship between or among the topics based on the visualized result. The change of topic can be interpreted in the following manners. From the result of dynamic topic modeling, we identify rising top 5 keywords and descending top 5 keywords for each period to show the change of the topic. Existing many topic visualization studies usually visualize keywords of each topic, but our approach proposed in this study differs from previous studies in that it attempts to visualize each topic itself. To evaluate the practical applicability of the proposed methodology, we performed an experiment on 1,847 abstracts of artificial intelligence-related papers. The experiment was performed by dividing abstracts of artificial intelligence-related papers into three periods (2016-2017, 2018-2019, 2020-2021). We selected seven topics based on the consistency score, and utilized the pre-trained word embedding model of Word2vec trained with 'Wikipedia', an Internet encyclopedia. Based on the proposed methodology, we generated a semantic vector for each topic. Through this, by reflecting the meaning of keywords, we visualized and interpreted the themes by period. Through these experiments, we confirmed that the rising and descending of the topic weight of a keyword can be usefully used to interpret the semantic change of the corresponding topic and to grasp the relationship among topics. In this study, to overcome the limitations of dynamic topic modeling results, we used word embedding and dimension reduction techniques to visualize topics by era. The results of this study are meaningful in that they broadened the scope of topic understanding through the visualization of dynamic topic modeling results. In addition, the academic contribution can be acknowledged in that it laid the foundation for follow-up studies using various word embeddings and dimensionality reduction techniques to improve the performance of the proposed methodology.

A Study on the Volatilities of Inbound Tourists Arrivals using the Multivariate BEKK model (다변량 BEKK모형을 이용한 방한 외래 관광객의 변동성에 대한 연구)

  • Kim, Kyung-Soo;Lee, Kyung-Hee
    • Management & Information Systems Review
    • /
    • v.32 no.3
    • /
    • pp.1-23
    • /
    • 2013
  • In this study, we try to investigate the spillover effects of volatility in international tourists arrivals between Korea and US, Japan, China by using the multivariate BEKK model from January 2005 to January 2013. In the results of this study, after the global financial crisis, we found a cointegration relationship and tourist arrivals of Japan were adjusted to recovery in the short term. Also tourists arrivals from China and Japan showed the long-term elasticity. In the conditional mean equation of a BEKK model, there were the spillover effects. And in the conditional variance equation, ARCH(${\epsilon}^2_t$) coefficients showed a strong influence on the arrivals of their own and the spillover effects and the asymmetric effects on the volatility of China and Japan arrivals. In GARCH(${\sigma}^2_t$) coefficients showed the asymmetric effects and the spillover effects of the conditional volatility among source arrivals. Therefore, we examined the asymmetric reaction of one-way or two-way tourist arrivals between source countries and Korea and the spillover effects related to tourists arrivals of source countries to Korea. We has confirmed a causal relationship between some of the tourists arrivals from source countries to korea.

  • PDF

Correct Closure of the Left Atrial Appendage Reduces Stagnant Blood Flow and the Risk of Thrombus Formation: A Proof-of-Concept Experimental Study Using 4D Flow Magnetic Resonance Imaging

  • Min Jae Cha;Don-Gwan An;Minsoo Kang;Hyue Mee Kim;Sang-Wook Kim;Iksung Cho;Joonhwa Hong;Hyewon Choi;Jee-Hyun Cho;Seung Yong Shin;Simon Song
    • Korean Journal of Radiology
    • /
    • v.24 no.7
    • /
    • pp.647-659
    • /
    • 2023
  • Objective: The study was conducted to investigate the effect of correct occlusion of the left atrial appendage (LAA) on intracardiac blood flow and thrombus formation in patients with atrial fibrillation (AF) using four-dimensional (4D) flow magnetic resonance imaging (MRI) and three-dimensional (3D)-printed phantoms. Materials and Methods: Three life-sized 3D-printed left atrium (LA) phantoms, including a pre-occlusion (i.e., before the occlusion procedure) model and correctly and incorrectly occluded post-procedural models, were constructed based on cardiac computed tomography images from an 86-year-old male with long-standing persistent AF. A custom-made closed-loop flow circuit was set up, and pulsatile simulated pulmonary venous flow was delivered by a pump. 4D flow MRI was performed using a 3T scanner, and the images were analyzed using MATLAB-based software (R2020b; Mathworks). Flow metrics associated with blood stasis and thrombogenicity, such as the volume of stasis defined by the velocity threshold ($\left|\vec{V}\right|$ < 3 cm/s), surface-and-time-averaged wall shear stress (WSS), and endothelial cell activation potential (ECAP), were analyzed and compared among the three LA phantom models. Results: Different spatial distributions, orientations, and magnitudes of LA flow were directly visualized within the three LA phantoms using 4D flow MRI. The time-averaged volume and its ratio to the corresponding entire volume of LA flow stasis were consistently reduced in the correctly occluded model (70.82 mL and 39.0%, respectively), followed by the incorrectly occluded (73.17 mL and 39.0%, respectively) and pre-occlusion (79.11 mL and 39.7%, respectively) models. The surfaceand-time-averaged WSS and ECAP were also lowest in the correctly occluded model (0.048 Pa and 4.004 Pa-1, respectively), followed by the incorrectly occluded (0.059 Pa and 4.792 Pa-1, respectively) and pre-occlusion (0.072 Pa and 5.861 Pa-1, respectively) models. Conclusion: These findings suggest that a correctly occluded LAA leads to the greatest reduction in LA flow stasis and thrombogenicity, presenting a tentative procedural goal to maximize clinical benefits in patients with AF.

Topic Based Hierarchical Network Analysis for Entrepreneur Using Text Mining (텍스트 마이닝을 이용한 주제기반의 기업인 네트워크 계층 분석)

  • Lee, Donghun;Kim, Yonghwa;Kim, Kwanho
    • The Journal of Society for e-Business Studies
    • /
    • v.23 no.3
    • /
    • pp.33-49
    • /
    • 2018
  • The importance of convergence activities among business is increasing due to the necessity of designing and developing new products to satisfy various customers' needs. In particular, decision makers such as CEOs are required to participate in networks between entrepreneurs for being connected with valuable convergence partners. Moreover, it is important for entrepreneurs not only to make a large number of network connections, but also to understand the networking relationship with entrepreneurs with similar topic information. However, there is a difficult limit in collecting the topic information that can show the lack of current status of business and the technology and characteristics of entrepreneur in industry sector. In this paper, we solve these problems through the topic extraction method and analyze the business network in three aspects. Specifically, there are C, S, T-Layer models, and each model analyzes amount of entrepreneurs relationship, network centrality, and topic similarity. As a result of experiments using real data, entrepreneur need to activate network by connecting high centrality entrepreneur when the corporate relationship is low. In addition, we confirmed through experiments that there is a need to activate the topic-based network when topic similarity is low between entrepreneurs.

Sentiment Analysis of Korean Reviews Using CNN: Focusing on Morpheme Embedding (CNN을 적용한 한국어 상품평 감성분석: 형태소 임베딩을 중심으로)

  • Park, Hyun-jung;Song, Min-chae;Shin, Kyung-shik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.59-83
    • /
    • 2018
  • With the increasing importance of sentiment analysis to grasp the needs of customers and the public, various types of deep learning models have been actively applied to English texts. In the sentiment analysis of English texts by deep learning, natural language sentences included in training and test datasets are usually converted into sequences of word vectors before being entered into the deep learning models. In this case, word vectors generally refer to vector representations of words obtained through splitting a sentence by space characters. There are several ways to derive word vectors, one of which is Word2Vec used for producing the 300 dimensional Google word vectors from about 100 billion words of Google News data. They have been widely used in the studies of sentiment analysis of reviews from various fields such as restaurants, movies, laptops, cameras, etc. Unlike English, morpheme plays an essential role in sentiment analysis and sentence structure analysis in Korean, which is a typical agglutinative language with developed postpositions and endings. A morpheme can be defined as the smallest meaningful unit of a language, and a word consists of one or more morphemes. For example, for a word '예쁘고', the morphemes are '예쁘(= adjective)' and '고(=connective ending)'. Reflecting the significance of Korean morphemes, it seems reasonable to adopt the morphemes as a basic unit in Korean sentiment analysis. Therefore, in this study, we use 'morpheme vector' as an input to a deep learning model rather than 'word vector' which is mainly used in English text. The morpheme vector refers to a vector representation for the morpheme and can be derived by applying an existent word vector derivation mechanism to the sentences divided into constituent morphemes. By the way, here come some questions as follows. What is the desirable range of POS(Part-Of-Speech) tags when deriving morpheme vectors for improving the classification accuracy of a deep learning model? Is it proper to apply a typical word vector model which primarily relies on the form of words to Korean with a high homonym ratio? Will the text preprocessing such as correcting spelling or spacing errors affect the classification accuracy, especially when drawing morpheme vectors from Korean product reviews with a lot of grammatical mistakes and variations? We seek to find empirical answers to these fundamental issues, which may be encountered first when applying various deep learning models to Korean texts. As a starting point, we summarized these issues as three central research questions as follows. First, which is better effective, to use morpheme vectors from grammatically correct texts of other domain than the analysis target, or to use morpheme vectors from considerably ungrammatical texts of the same domain, as the initial input of a deep learning model? Second, what is an appropriate morpheme vector derivation method for Korean regarding the range of POS tags, homonym, text preprocessing, minimum frequency? Third, can we get a satisfactory level of classification accuracy when applying deep learning to Korean sentiment analysis? As an approach to these research questions, we generate various types of morpheme vectors reflecting the research questions and then compare the classification accuracy through a non-static CNN(Convolutional Neural Network) model taking in the morpheme vectors. As for training and test datasets, Naver Shopping's 17,260 cosmetics product reviews are used. To derive morpheme vectors, we use data from the same domain as the target one and data from other domain; Naver shopping's about 2 million cosmetics product reviews and 520,000 Naver News data arguably corresponding to Google's News data. The six primary sets of morpheme vectors constructed in this study differ in terms of the following three criteria. First, they come from two types of data source; Naver news of high grammatical correctness and Naver shopping's cosmetics product reviews of low grammatical correctness. Second, they are distinguished in the degree of data preprocessing, namely, only splitting sentences or up to additional spelling and spacing corrections after sentence separation. Third, they vary concerning the form of input fed into a word vector model; whether the morphemes themselves are entered into a word vector model or with their POS tags attached. The morpheme vectors further vary depending on the consideration range of POS tags, the minimum frequency of morphemes included, and the random initialization range. All morpheme vectors are derived through CBOW(Continuous Bag-Of-Words) model with the context window 5 and the vector dimension 300. It seems that utilizing the same domain text even with a lower degree of grammatical correctness, performing spelling and spacing corrections as well as sentence splitting, and incorporating morphemes of any POS tags including incomprehensible category lead to the better classification accuracy. The POS tag attachment, which is devised for the high proportion of homonyms in Korean, and the minimum frequency standard for the morpheme to be included seem not to have any definite influence on the classification accuracy.