• Title/Summary/Keyword: Embeddings

Search Result 93, Processing Time 0.035 seconds

CR-M-SpanBERT: Multiple embedding-based DNN coreference resolution using self-attention SpanBERT

  • Joon-young Jung
    • ETRI Journal
    • /
    • v.46 no.1
    • /
    • pp.35-47
    • /
    • 2024
  • This study introduces CR-M-SpanBERT, a coreference resolution (CR) model that utilizes multiple embedding-based span bidirectional encoder representations from transformers, for antecedent recognition in natural language (NL) text. Information extraction studies aimed to extract knowledge from NL text autonomously and cost-effectively. However, the extracted information may not represent knowledge accurately owing to the presence of ambiguous entities. Therefore, we propose a CR model that identifies mentions referring to the same entity in NL text. In the case of CR, it is necessary to understand both the syntax and semantics of the NL text simultaneously. Therefore, multiple embeddings are generated for CR, which can include syntactic and semantic information for each word. We evaluate the effectiveness of CR-M-SpanBERT by comparing it to a model that uses SpanBERT as the language model in CR studies. The results demonstrate that our proposed deep neural network model achieves high-recognition accuracy for extracting antecedents from NL text. Additionally, it requires fewer epochs to achieve an average F1 accuracy greater than 75% compared with the conventional SpanBERT approach.

Alzheimer's disease recognition from spontaneous speech using large language models

  • Jeong-Uk Bang;Seung-Hoon Han;Byung-Ok Kang
    • ETRI Journal
    • /
    • v.46 no.1
    • /
    • pp.96-105
    • /
    • 2024
  • We propose a method to automatically predict Alzheimer's disease from speech data using the ChatGPT large language model. Alzheimer's disease patients often exhibit distinctive characteristics when describing images, such as difficulties in recalling words, grammar errors, repetitive language, and incoherent narratives. For prediction, we initially employ a speech recognition system to transcribe participants' speech into text. We then gather opinions by inputting the transcribed text into ChatGPT as well as a prompt designed to solicit fluency evaluations. Subsequently, we extract embeddings from the speech, text, and opinions by the pretrained models. Finally, we use a classifier consisting of transformer blocks and linear layers to identify participants with this type of dementia. Experiments are conducted using the extensively used ADReSSo dataset. The results yield a maximum accuracy of 87.3% when speech, text, and opinions are used in conjunction. This finding suggests the potential of leveraging evaluation feedback from language models to address challenges in Alzheimer's disease recognition.

Graph Implicit Neural Representations Using Spatial Graph Embeddings (공간적 그래프 임베딩을 활용한 그래프 암시적 신경 표현)

  • Jinho Park;Dongwoo Kim
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2024.01a
    • /
    • pp.23-26
    • /
    • 2024
  • 본 논문에서는 그래프 구조의 데이터에서 각 노드의 신호를 예측하는 연구를 진행하였다. 이를 위해 분석하고자 하는 그래프에 대해 연결 관계를 기반으로 각 노드에 비-유클리드 공간 상에서의 좌표를 부여하여 그래프의 공간적 임베딩을 얻은 뒤, 각 노드의 공간적 임베딩을 입력으로 받고 해당 노드의 신호를 예측하는 그래프 암시적 신경 표현 모델을 제안 하였다. 제안된 모델의 검증을 위해 네트워크형 데이터와 3차원 메시 데이터 두 종류의 그래프 데이터에 대하여 신호 학습, 신호 예측 및 메시 데이터의 초해상도 과정 실험들을 진행하였다. 전반적으로 기존의 그래프 암시적 신경 표현 모델과 비교하였을 때 비슷하거나 더 우수한 성능을 보였으며, 특히 네트워크형 그래프 데이터 신호 예측 실험에서 큰 성능 향상을 보였다.

  • PDF

Selective Word Embedding for Sentence Classification by Considering Information Gain and Word Similarity (문장 분류를 위한 정보 이득 및 유사도에 따른 단어 제거와 선택적 단어 임베딩 방안)

  • Lee, Min Seok;Yang, Seok Woo;Lee, Hong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.4
    • /
    • pp.105-122
    • /
    • 2019
  • Dimensionality reduction is one of the methods to handle big data in text mining. For dimensionality reduction, we should consider the density of data, which has a significant influence on the performance of sentence classification. It requires lots of computations for data of higher dimensions. Eventually, it can cause lots of computational cost and overfitting in the model. Thus, the dimension reduction process is necessary to improve the performance of the model. Diverse methods have been proposed from only lessening the noise of data like misspelling or informal text to including semantic and syntactic information. On top of it, the expression and selection of the text features have impacts on the performance of the classifier for sentence classification, which is one of the fields of Natural Language Processing. The common goal of dimension reduction is to find latent space that is representative of raw data from observation space. Existing methods utilize various algorithms for dimensionality reduction, such as feature extraction and feature selection. In addition to these algorithms, word embeddings, learning low-dimensional vector space representations of words, that can capture semantic and syntactic information from data are also utilized. For improving performance, recent studies have suggested methods that the word dictionary is modified according to the positive and negative score of pre-defined words. The basic idea of this study is that similar words have similar vector representations. Once the feature selection algorithm selects the words that are not important, we thought the words that are similar to the selected words also have no impacts on sentence classification. This study proposes two ways to achieve more accurate classification that conduct selective word elimination under specific regulations and construct word embedding based on Word2Vec embedding. To select words having low importance from the text, we use information gain algorithm to measure the importance and cosine similarity to search for similar words. First, we eliminate words that have comparatively low information gain values from the raw text and form word embedding. Second, we select words additionally that are similar to the words that have a low level of information gain values and make word embedding. In the end, these filtered text and word embedding apply to the deep learning models; Convolutional Neural Network and Attention-Based Bidirectional LSTM. This study uses customer reviews on Kindle in Amazon.com, IMDB, and Yelp as datasets, and classify each data using the deep learning models. The reviews got more than five helpful votes, and the ratio of helpful votes was over 70% classified as helpful reviews. Also, Yelp only shows the number of helpful votes. We extracted 100,000 reviews which got more than five helpful votes using a random sampling method among 750,000 reviews. The minimal preprocessing was executed to each dataset, such as removing numbers and special characters from text data. To evaluate the proposed methods, we compared the performances of Word2Vec and GloVe word embeddings, which used all the words. We showed that one of the proposed methods is better than the embeddings with all the words. By removing unimportant words, we can get better performance. However, if we removed too many words, it showed that the performance was lowered. For future research, it is required to consider diverse ways of preprocessing and the in-depth analysis for the co-occurrence of words to measure similarity values among words. Also, we only applied the proposed method with Word2Vec. Other embedding methods such as GloVe, fastText, ELMo can be applied with the proposed methods, and it is possible to identify the possible combinations between word embedding methods and elimination methods.

Deep recurrent neural networks with word embeddings for Urdu named entity recognition

  • Khan, Wahab;Daud, Ali;Alotaibi, Fahd;Aljohani, Naif;Arafat, Sachi
    • ETRI Journal
    • /
    • v.42 no.1
    • /
    • pp.90-100
    • /
    • 2020
  • Named entity recognition (NER) continues to be an important task in natural language processing because it is featured as a subtask and/or subproblem in information extraction and machine translation. In Urdu language processing, it is a very difficult task. This paper proposes various deep recurrent neural network (DRNN) learning models with word embedding. Experimental results demonstrate that they improve upon current state-of-the-art NER approaches for Urdu. The DRRN models evaluated include forward and bidirectional extensions of the long short-term memory and back propagation through time approaches. The proposed models consider both language-dependent features, such as part-of-speech tags, and language-independent features, such as the "context windows" of words. The effectiveness of the DRNN models with word embedding for NER in Urdu is demonstrated using three datasets. The results reveal that the proposed approach significantly outperforms previous conditional random field and artificial neural network approaches. The best f-measure values achieved on the three benchmark datasets using the proposed deep learning approaches are 81.1%, 79.94%, and 63.21%, respectively.

ROLLING STONES WITH NONCONVEX SIDES II: ALL TIME REGULARITY OF INTERFACE AND SURFACE

  • Lee, Ki-Ahm;Rhee, Eun-Jai
    • Journal of the Korean Mathematical Society
    • /
    • v.49 no.3
    • /
    • pp.585-604
    • /
    • 2012
  • In this paper we consider the evolution of the rolling stone with a rotationally symmetric nonconvex compact initial surface ${\Sigma}_0$ under the Gauss curvature flow. Let $X:S^n{\times}[0,\;{\infty}){\rightarrow}\mathbb{R}^{n+1}$ be the embeddings of the sphere in $\mathbb{R}^{n+1}$ such that $\Sigma(t)=X(S^n,t)$ is the surface at time t and ${\Sigma}(0)={\Sigma}_0$. As a consequence the parabolic equation describing the motion of the hypersurface becomes degenerate on the interface separating the nonconvex part from the strictly convex side, since one of the curvature will be zero on the interface. By expressing the strictly convex part of the surface near the interface as a graph of a function $z=f(r,t)$ and the non-convex part of the surface near the interface as a graph of a function $z={\varphi}(r)$, we show that if at time $t=0$, $g=\frac{1}{n}f^{n-1}_{r}$ vanishes linearly at the interface, the $g(r,t)$ will become smooth up to the interface for long time before focusing.

Recognition of GUI Widgets Utilizing Translational Embeddings based on Relational Learning (트랜슬레이션 임베딩 기반 관계 학습을 이용한 GUI 위젯 인식)

  • Park, Min-Su;Seok, Ho-Sik
    • Journal of IKEEE
    • /
    • v.22 no.3
    • /
    • pp.693-699
    • /
    • 2018
  • CNN based object recognitions have reported splendid results. However, the recognition of mobile apps raises an interesting challenge that recognition performance of similar widgets is not consistent. In order to improve the performance, we propose a noble method utilizing relations between input widgets. The recognition process flows from the Faster R-CNN based recognition to enhancement using a relation recognizer. The relations are represented as vector translation between objects in a relation space. Experiments on 323 apps show that our method significantly enhances the Faster R-CNN only approach.

CNN Architecture Predicting Movie Rating from Audience's Reviews Written in Korean (한국어 관객 평가기반 영화 평점 예측 CNN 구조)

  • Kim, Hyungchan;Oh, Heung-Seon;Kim, Duksu
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.9 no.1
    • /
    • pp.17-24
    • /
    • 2020
  • In this paper, we present a movie rating prediction architecture based on a convolutional neural network (CNN). Our prediction architecture extends TextCNN, a popular CNN-based architecture for sentence classification, in three aspects. First, character embeddings are utilized to cover many variants of words since reviews are short and not well-written linguistically. Second, the attention mechanism (i.e., squeeze-and-excitation) is adopted to focus on important features. Third, a scoring function is proposed to convert the output of an activation function to a review score in a certain range (1-10). We evaluated our prediction architecture on a movie review dataset and achieved a low MSE (e.g., 3.3841) compared with an existing method. It showed the superiority of our movie rating prediction architecture.

Conncetiveity of X-Hypercubes and Its Applications (X-Hypercubes의 연결성과 그 응용)

  • Gwon, Gyeong-Hui
    • The Transactions of the Korea Information Processing Society
    • /
    • v.1 no.1
    • /
    • pp.92-98
    • /
    • 1994
  • The hypercube-like interconncetion network,X-hypercubes,has the same number of nodes and edges as conventional hypercubes.By slightly changing the interconneton way between nodes,however,X-hypercubes reduces the diameter by almost half.Thus the communication delay in X-hypercubes can be expected to be much lower than that in hypercubes. This paper gives a new definition of X-hypercubes establishing clear-cut condition of connection between two nodes.As appliction examples of the new definition,this paper presents simple embeddings of hypercubes in X-hypercubes and vice versa.This means that any programs written for hypercubes can be transported onto X-hypercubes and vice versa with minimal overhead.This paper also present bitonic merge sort for X-hypercubes by simulation that for hypercubes.

  • PDF

Correlation-based Robust Blind Watermarking (상관도 기반의 강인한 블라인드 워터마킹)

  • Joo, Snag-Hyun;Seo, Yong-Seok
    • The KIPS Transactions:PartB
    • /
    • v.10B no.5
    • /
    • pp.479-484
    • /
    • 2003
  • We propose a blind watermarking method that embeds a binary pseudo-random sequence (watermarks), (-1, 1), into wavelet dc components, while most watermarking techniques embed watermarks in the middle frequency range for robustness and fidelity. In our scheme, the watermarks are embedded into particular locations to be selected by a key, where some watermark embeddings are skipped to avoid severe degradation in quality. Our robustness is compared to some results registered to the ChechMark [1] that is one of the most popular benchmarking tools.