• Title/Summary/Keyword: Representations

Search Result 1,314, Processing Time 0.025 seconds

Selective Word Embedding for Sentence Classification by Considering Information Gain and Word Similarity (문장 분류를 위한 정보 이득 및 유사도에 따른 단어 제거와 선택적 단어 임베딩 방안)

  • Lee, Min Seok;Yang, Seok Woo;Lee, Hong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.4
    • /
    • pp.105-122
    • /
    • 2019
  • Dimensionality reduction is one of the methods to handle big data in text mining. For dimensionality reduction, we should consider the density of data, which has a significant influence on the performance of sentence classification. It requires lots of computations for data of higher dimensions. Eventually, it can cause lots of computational cost and overfitting in the model. Thus, the dimension reduction process is necessary to improve the performance of the model. Diverse methods have been proposed from only lessening the noise of data like misspelling or informal text to including semantic and syntactic information. On top of it, the expression and selection of the text features have impacts on the performance of the classifier for sentence classification, which is one of the fields of Natural Language Processing. The common goal of dimension reduction is to find latent space that is representative of raw data from observation space. Existing methods utilize various algorithms for dimensionality reduction, such as feature extraction and feature selection. In addition to these algorithms, word embeddings, learning low-dimensional vector space representations of words, that can capture semantic and syntactic information from data are also utilized. For improving performance, recent studies have suggested methods that the word dictionary is modified according to the positive and negative score of pre-defined words. The basic idea of this study is that similar words have similar vector representations. Once the feature selection algorithm selects the words that are not important, we thought the words that are similar to the selected words also have no impacts on sentence classification. This study proposes two ways to achieve more accurate classification that conduct selective word elimination under specific regulations and construct word embedding based on Word2Vec embedding. To select words having low importance from the text, we use information gain algorithm to measure the importance and cosine similarity to search for similar words. First, we eliminate words that have comparatively low information gain values from the raw text and form word embedding. Second, we select words additionally that are similar to the words that have a low level of information gain values and make word embedding. In the end, these filtered text and word embedding apply to the deep learning models; Convolutional Neural Network and Attention-Based Bidirectional LSTM. This study uses customer reviews on Kindle in Amazon.com, IMDB, and Yelp as datasets, and classify each data using the deep learning models. The reviews got more than five helpful votes, and the ratio of helpful votes was over 70% classified as helpful reviews. Also, Yelp only shows the number of helpful votes. We extracted 100,000 reviews which got more than five helpful votes using a random sampling method among 750,000 reviews. The minimal preprocessing was executed to each dataset, such as removing numbers and special characters from text data. To evaluate the proposed methods, we compared the performances of Word2Vec and GloVe word embeddings, which used all the words. We showed that one of the proposed methods is better than the embeddings with all the words. By removing unimportant words, we can get better performance. However, if we removed too many words, it showed that the performance was lowered. For future research, it is required to consider diverse ways of preprocessing and the in-depth analysis for the co-occurrence of words to measure similarity values among words. Also, we only applied the proposed method with Word2Vec. Other embedding methods such as GloVe, fastText, ELMo can be applied with the proposed methods, and it is possible to identify the possible combinations between word embedding methods and elimination methods.

The Definition of a Catastrophe as Trauma by Visual Media and the Resultant Problems: A Critical Analysis of the "Antimimetic Theory" (시각 미디어에 의한 대재앙적 사건의 트라우마 규정과 그에 따른 문제들 - "반모방 이론"에 대한 비평적인 분석을 통해서)

  • Seoh, Gil-Wan
    • Cross-Cultural Studies
    • /
    • v.43
    • /
    • pp.265-288
    • /
    • 2016
  • This essay begins by discussing the issue of the definition of a catastrophe as a trauma by visual media and the problems that result. We assume a concrete approach towards these problems by examining the process through which 9/11 was defined as a "national trauma" in the exclusion of images of bodies falling from towers, which were some of the most shocking images in the media coverage of 9/11. The choice to exclude images of falling men from American visual media representations of 9/11 goes hand in hand with the tendencies of a contemporary trauma theory. This essay assumes that the representations in the U.S visual media depend on "antimimetic theory," one of the leading contemporary trauma theories, in order to validate its logic, and examines the limitations and problems of the theory. This work aims to examine the issue of the definition of a catastrophe as trauma by visual media on the basis on the "antimimetic theory" and the danger that results. Because the antimimetic theory, which the visual media in the United States uses to define 9/11 as trauma, emphasizes literal and unmediated representation of an external event, it lacks an understanding of the human aspects of the event. There is no way to intervene in the construction and interpretation of the trauma. As a result, the theory discourages active attempts to find a solution to the problems of the people directly connected with the event. Thus, it provides an opening for manipulative intervention of an external power. This essay attempts to provide a critical analysis of the "antimimetic theory" in order to help people who witness catastrophic events through various types of visual media, and to seek an alternative means of experiencing and responding to the trauma, that does not stem from the perspective of specific media outlets or external powers.

The Churchlands' Theory of Representation and the Semantics (처칠랜드의 표상이론과 의미론적 유사성)

  • Park, Je-Youn
    • Korean Journal of Cognitive Science
    • /
    • v.23 no.2
    • /
    • pp.133-164
    • /
    • 2012
  • Paul Churchland(1989) suggests the theory of representation from the results of cognitive biology and connectionist AI studies. According to the theory, our representations of the diverse phenomena in the world can be represented as the positions of phase state spaces with the actions of the neurons or of the assembly of neurons. He insists connectionist AI neural networks can have the semantical category systems to recognize the world. But Fodor and Lepore(1996) don't look the perspective bright. From their points of view, the Churchland's theory of representation stands on the base of Quine's holism, and the network semantics cannot explain how the criteria of semantical content similarity could be possible, and so cannot the theory. This thesis aims to excavate which one is the better between the perspective of the theory and the one of Fodor and Lepore's. From my understandings of state space theory of representation, artificial nets can coordinates the criteria of contents similarity by the learning algorithm. On the basis of these, I can see that Fodor and Lepore's points cannot penetrate the Churchlands' theory. From the view point of the theory, we can see how the future's artificial systems can have the conceptual systems recognizing the world. Therefore we can have the perspectives what cognitive scientists have to focus on.

  • PDF

Segmentation and Visualization of Human Anatomy using Medical Imagery (의료영상을 이용한 인체장기의 분할 및 시각화)

  • Lee, Joon-Ku;Kim, Yang-Mo;Kim, Do-Yeon
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.8 no.1
    • /
    • pp.191-197
    • /
    • 2013
  • Conventional CT and MRI scans produce cross-section slices of body that are viewed sequentially by radiologists who must imagine or extrapolate from these views what the 3 dimensional anatomy should be. By using sophisticated algorithm and high performance computing, these cross-sections may be rendered as direct 3D representations of human anatomy. The 2D medical image analysis forced to use time-consuming, subjective, error-prone manual techniques, such as slice tracing and region painting, for extracting regions of interest. To overcome the drawbacks of 2D medical image analysis, combining with medical image processing, 3D visualization is essential for extracting anatomical structures and making measurements. We used the gray-level thresholding, region growing, contour following, deformable model to segment human organ and used the feature vectors from texture analysis to detect harmful cancer. We used the perspective projection and marching cube algorithm to render the surface from volumetric MR and CT image data. The 3D visualization of human anatomy and segmented human organ provides valuable benefits for radiation treatment planning, surgical planning, surgery simulation, image guided surgery and interventional imaging applications.

An Analysis of the Results of a Mathematics Diagnostic Test taken by Multicultural Koreans in their First or Second Year of Elementary School (다문화가정 학생 대상 언어.인지 진단도구 적용 결과 분석 - 초등학교 1.2학년 수학 -)

  • Cho, Young-Mi;Lee, Og-Young
    • Journal of Educational Research in Mathematics
    • /
    • v.20 no.2
    • /
    • pp.103-119
    • /
    • 2010
  • This study aims to figure out the characteristics of the mathematical ability of multicultural Korean elementary school learners. This was done by analyzing the results of a mathematics diagnostic test given to multicultural Korean first and second year elementary school students. The findings of this study mainly support the following three. First, it was indicated that, regardless of whether the students are multicultural or not, more second-year students had difficulty in understanding mathematics than the first-year students. Specifically, a higher percentage of second-year students were below the reference point (cut-off point) than was the case in the first-year learners, which pattern of the overall Korean students was consistent with that of multicultural Koreans. Second, concerning the sub-fields of mathematics, higher proportion of the students fell below the cut-off point in 'numbers and arithmetics' area than in 'measure and geometry,' which pattern was again the same with the multicultural students. Third, it was implied that, in addition to mathematically more complex questions, linguistically complex sentential representations contributed to increasing the difficulty of the test items. It is suggested that care be taken to enhance linguistic processing and to employ well-defined terms.

  • PDF

Development of Mean Stand Height Module Using Image-Based Point Cloud and FUSION S/W (영상 기반 3차원 점군과 FUSION S/W 기반의 임분고 분석 모듈 개발)

  • KIM, Kyoung-Min
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.19 no.4
    • /
    • pp.169-185
    • /
    • 2016
  • Recently mean stand height has been added as new attribute to forest type maps, but it is often too costly and time consuming to manually measure 9,100,000 points from countrywide stereo aerial photos. In addition, tree heights are frequently measured around tombs and forest edges, which are poor representations of the interior tree stand. This work proposes an estimation of mean stand height using an image-based point cloud, which was extracted from stereo aerial photo with FUSION S/W. Then, a digital terrain model was created by filtering the DSM point cloud and subtracting the DTM from DSM, resulting in nDSM, which represents object heights (buildings, trees, etc.). The RMSE was calculated to compare differences in tree heights between those observed and extracted from the nDSM. The resulting RMSE of average total plot height was 0.96 m. Individual tree heights of the whole study site area were extracted using the USDA Forest Service's FUSION S/W. Finally, mean stand height was produced by averaging individual tree heights in a stand polygon of the forest type map. In order to automate the mean stand height extraction using photogrammetric methods, a module was developed as an ArcGIS add-in toolbox.

A Comparative Analysis of Proportional Expression and Proportional Distribution in Elementary Mathematics Textbooks (비례식과 비례배분에 대한 초등 수학 교과서 비교 분석)

  • Chang, Hyewon;Park, Haemin;Kim, Jusuk;Lim, Miin;Yu, Migyoung;Lee, Hwayoung
    • School Mathematics
    • /
    • v.19 no.2
    • /
    • pp.229-248
    • /
    • 2017
  • This study investigated the factors that should be considered when teaching proportional expression and proportional distribution through literature review. Based on these results, we analyzed and compared Korean and foreign mathematics textbooks on proportional expression and proportional distribution longitudinally and horizontally to search for desirable methods of organizing the unit of proportional expression and proportional distribution in mathematics textbooks. For longitudinal analysis, we took the mathematics textbooks according to the national curriculum since the 5th one. For horizontal analysis, we selected the mathematics textbooks of Japan, Singapore, and China. In each textbook, the contents and the order in relation to proportional expression and proportional distribution, the definitions of terminology, and the contexts and the visual representations for introducing related concepts are selected as the analysis framework. The results of analysis revealed many characteristics and the differences in ways of dealing contents about proportional expression and proportional distribution. Based on these results, we suggested some implications for writing the unit of proportional expression and proportional distribution in elementary mathematics textbooks.

Transformation of Discourse on Uses of Computer Technology in Korean Landscape Architecture - Focused on Journal of the Korean Institute of Landscape Architecture and Environmental & Landscape Architecture of Korea - (한국 조경에서 컴퓨터 테크놀로지의 활용에 관한 담론의 변천 - 『한국조경학회지』와 『환경과조경』을 중심으로 -)

  • Lee, Myeong-Jun
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.48 no.1
    • /
    • pp.15-24
    • /
    • 2020
  • This work examines discourse on uses of computer technology and its transformation during the last thirty years effecting Korea landscape architecture. First, in the 1990s landscape architects begun to use computers as a new technology for landscape architecture, programming computer software utilities suitable for landscape research, planning, and design. The landscape architects, acting as computer programmers, tried to explore various techniques for landscape analysis and design with a piece of software, and various experts within the field of landscape architecture collaborated with each other. However, landscape architects mainly used computer technology as a tool as a substitute for hands-on cases. Since around the 2000s, the discourse on mapping and diagrammatic techniques as a visualization technique for landscaping processes have begun. Also, realistic representations for perspective drawings using graphic software have been increasingly important. The landscape architects, acting as graphic designers, focused on the specific visualization techniques for landscape planning and design. However, computer technology has been mainly used to produce realistic visuals aids for final presentations instead of creative exploration to generate landforms. Additionally, recent landscape architects have been using landscape performance modeling and parametric modeling for landform and landscape furniture design. The landscape architects as spatial designers are actively using computer modeling as creative form-generating tools during the design process.

A Study on Warranty in The Insurance Act 2015 (영국 2015년 보험법 상 담보(워런티)에 관한 연구)

  • SHIN, Gun-Hoon;LEE, Byung-Mun
    • THE INTERNATIONAL COMMERCE & LAW REVIEW
    • /
    • v.73
    • /
    • pp.65-90
    • /
    • 2017
  • The rule of warranty in English insurance law was established in the second part of the $18^{th}$ century by Lord Mansfield, who laid the foundations of the modern English law of insurance contract and developed very different rule of insurance law, especially in the field of warranty. At the time of Lord Mansfield, warranty, that is, the promise given by the assured, played an important role for the insurer to assess the scope of the risk. Legal environments, however, have changed since the age of Lord Mansfield. English and Scottish Commissions proposed very dramatic reform of law in the field of warranty law to reflect the changes of legal environment through the Insurance Act 2016. This article intends to consider the legal implications through the comparative analysis between the new regime of warranty in the Insurance Act 2015 and MIA 1906. The major changes in the Insurance Act 2015 are summarized as following. First, Basis of the contract clauses in non-consumer insurance contracts should be of no effect and representations should not be capable of being converted into warranties by means of a policy term or statement on the proposal form. This requirement should not be capable of being avoided by the use of a contract term and the arrangement of contracting out by parties should be of no effect. Secondly, The existing remedy for breach of warranty, that is, automatic discharge of the insurer's liability, should be removed. Instead, the insurer's libility should be suspended from the point of breach of warranty and reattach if and when a breach of warranty has been remedies. Thirdly, A breach of warranty should genally be regarded as remedied where the insured ceases to be in breach of it. In the other hand, for time-specific warranties which apply at or by an ascertainable time, a breach should be regarded as remedies, if the risk to which the warranty relates later, becomes essentially the same as that originally contemplated by the parties. Fourthly, where a term of an insurance contract relates to a particular kind of loss, or loss at a particular location/time, the breach of that term should only give the remedy in relation to loss of that particular kind of loss, or at a particular location/time. Finally, whether a term of an insurance contrat relates to loss of a particular kind of at a particular location/time should be determined objectively, based on whether compliance with that ther would tend to reduce the risk of the occurrence of that category of loss.

  • PDF

An Analysis on the Understanding of Middle School Students about the Concept of Function Based on Integrated Understanding (통합적 이해의 관점에서 중학교 학생들의 함수 개념 이해 분석)

  • Lee, Young Kyoung;Kim, Eun Sook;Lee, Ha Woo;Cho, Wan Young
    • Communications of Mathematical Education
    • /
    • v.30 no.2
    • /
    • pp.199-223
    • /
    • 2016
  • The purpose of this study is to investigate how first and second graders in middle school take in integrated understanding about the concept of function. The data was collected through the questionnaire conducted by the first and second-year students at A, B middle school in Cheongju. The questionnaire consisted of 14 questions related to the extent of understanding a concept of function, the ability to express function and to translate function. The results are summarized as follows. First, the percentage of correct answer made a difference according to the types of representation. Questions leading students to translate a task into a table or an equation showed quite high correct response rates. However, questions asking students to translate a task into graphs showed high incorrect responses. Second, the result shows that students have the different viewpoints depending on their grades when they have to determine whether the suggested situation belongs to function. The first-year students tended to consider function as the concept of 'definition'. On the other hand, the second-year students emphasized 'equation' of function. Finally, only a few students can distinguish the various situations and representations into the definition of function. This result shows that students didn't get the integrated understanding of the concept of function.