• Title/Summary/Keyword: Tree-Based Network

Search Result 631, Processing Time 0.033 seconds

A Study on the Effect of the Document Summarization Technique on the Fake News Detection Model (문서 요약 기법이 가짜 뉴스 탐지 모형에 미치는 영향에 관한 연구)

  • Shim, Jae-Seung;Won, Ha-Ram;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.201-220
    • /
    • 2019
  • Fake news has emerged as a significant issue over the last few years, igniting discussions and research on how to solve this problem. In particular, studies on automated fact-checking and fake news detection using artificial intelligence and text analysis techniques have drawn attention. Fake news detection research entails a form of document classification; thus, document classification techniques have been widely used in this type of research. However, document summarization techniques have been inconspicuous in this field. At the same time, automatic news summarization services have become popular, and a recent study found that the use of news summarized through abstractive summarization has strengthened the predictive performance of fake news detection models. Therefore, the need to study the integration of document summarization technology in the domestic news data environment has become evident. In order to examine the effect of extractive summarization on the fake news detection model, we first summarized news articles through extractive summarization. Second, we created a summarized news-based detection model. Finally, we compared our model with the full-text-based detection model. The study found that BPN(Back Propagation Neural Network) and SVM(Support Vector Machine) did not exhibit a large difference in performance; however, for DT(Decision Tree), the full-text-based model demonstrated a somewhat better performance. In the case of LR(Logistic Regression), our model exhibited the superior performance. Nonetheless, the results did not show a statistically significant difference between our model and the full-text-based model. Therefore, when the summary is applied, at least the core information of the fake news is preserved, and the LR-based model can confirm the possibility of performance improvement. This study features an experimental application of extractive summarization in fake news detection research by employing various machine-learning algorithms. The study's limitations are, essentially, the relatively small amount of data and the lack of comparison between various summarization technologies. Therefore, an in-depth analysis that applies various analytical techniques to a larger data volume would be helpful in the future.

A Study on Animation Character Face Design System Based on Physiognomic Judgment of Character Study in the Cosmic Dual Forces and the Five Elements Thoughts (음양오행(陰陽五行)사상의 관상학에 기반한 애니메이션 캐릭터 얼굴 설계 시스템 연구)

  • Hong, Soo-Hyeon;Kim, Jae-Ho
    • Journal of Korea Multimedia Society
    • /
    • v.9 no.7
    • /
    • pp.872-893
    • /
    • 2006
  • In this study, I classify the elements of physiognomic judgment of character with regard to form and meaning from a visual perspective based on physiognomic judgment of character study in 'the cosmic dual forces and the Five Elements theory'. Individual characters for each type are designed using graphic data. Based on that, design system of individual characters for each personality type is investigated using Neural Network system. Faces with O-Haeng (Five Elements) shapes are shown to constitute the system with ${\pm}0.3%$ degree of error tolerance for the non-loaming input data. For the shapes of Chinese characters 'tree, fire, soil, gold and water', their MSE(Mean Square Error) are 0.3, 0.3, 0.2, 0.5, 0.2. It seems to be the best regarding the scoring system which ranges from 0 to 5. Therefore, this system might be regarded to produce the most accurate facial shape of character automatically when we input character's personality we desire to make.

  • PDF

Clickstream Big Data Mining for Demographics based Digital Marketing (인구통계특성 기반 디지털 마케팅을 위한 클릭스트림 빅데이터 마이닝)

  • Park, Jiae;Cho, Yoonho
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.3
    • /
    • pp.143-163
    • /
    • 2016
  • The demographics of Internet users are the most basic and important sources for target marketing or personalized advertisements on the digital marketing channels which include email, mobile, and social media. However, it gradually has become difficult to collect the demographics of Internet users because their activities are anonymous in many cases. Although the marketing department is able to get the demographics using online or offline surveys, these approaches are very expensive, long processes, and likely to include false statements. Clickstream data is the recording an Internet user leaves behind while visiting websites. As the user clicks anywhere in the webpage, the activity is logged in semi-structured website log files. Such data allows us to see what pages users visited, how long they stayed there, how often they visited, when they usually visited, which site they prefer, what keywords they used to find the site, whether they purchased any, and so forth. For such a reason, some researchers tried to guess the demographics of Internet users by using their clickstream data. They derived various independent variables likely to be correlated to the demographics. The variables include search keyword, frequency and intensity for time, day and month, variety of websites visited, text information for web pages visited, etc. The demographic attributes to predict are also diverse according to the paper, and cover gender, age, job, location, income, education, marital status, presence of children. A variety of data mining methods, such as LSA, SVM, decision tree, neural network, logistic regression, and k-nearest neighbors, were used for prediction model building. However, this research has not yet identified which data mining method is appropriate to predict each demographic variable. Moreover, it is required to review independent variables studied so far and combine them as needed, and evaluate them for building the best prediction model. The objective of this study is to choose clickstream attributes mostly likely to be correlated to the demographics from the results of previous research, and then to identify which data mining method is fitting to predict each demographic attribute. Among the demographic attributes, this paper focus on predicting gender, age, marital status, residence, and job. And from the results of previous research, 64 clickstream attributes are applied to predict the demographic attributes. The overall process of predictive model building is compose of 4 steps. In the first step, we create user profiles which include 64 clickstream attributes and 5 demographic attributes. The second step performs the dimension reduction of clickstream variables to solve the curse of dimensionality and overfitting problem. We utilize three approaches which are based on decision tree, PCA, and cluster analysis. We build alternative predictive models for each demographic variable in the third step. SVM, neural network, and logistic regression are used for modeling. The last step evaluates the alternative models in view of model accuracy and selects the best model. For the experiments, we used clickstream data which represents 5 demographics and 16,962,705 online activities for 5,000 Internet users. IBM SPSS Modeler 17.0 was used for our prediction process, and the 5-fold cross validation was conducted to enhance the reliability of our experiments. As the experimental results, we can verify that there are a specific data mining method well-suited for each demographic variable. For example, age prediction is best performed when using the decision tree based dimension reduction and neural network whereas the prediction of gender and marital status is the most accurate by applying SVM without dimension reduction. We conclude that the online behaviors of the Internet users, captured from the clickstream data analysis, could be well used to predict their demographics, thereby being utilized to the digital marketing.

Efficient Processing of Aggregate Queries in Wireless Sensor Networks (무선 센서 네트워크에서 효율적인 집계 질의 처리)

  • Kim, Joung-Joon;Shin, In-Su;Lee, Ki-Young;Han, Ki-Joon
    • Spatial Information Research
    • /
    • v.19 no.3
    • /
    • pp.95-106
    • /
    • 2011
  • Recently as efficient processing of aggregate queries for fetching desired data from sensors has been recognized as a crucial part, in-network aggregate query processing techniques are studied intensively in wireless sensor networks. Existing representative in-network aggregate query processing techniques propose routing algorithms and data structures for processing aggregate queries. However, these aggregate query processing techniques have problems such as high energy consumption in sensor nodes, low accuracy of query processing results, and long query processing time. In order to solve these problems and to enhance the efficiency of aggregate query processing in wireless sensor networks, this paper proposes Bucket-based Parallel Aggregation(BPA). BPA divides a query region into several cells according to the distribution of sensor nodes and builds a Quad-tree, and then processes aggregate queries in parallel for each cell region according to routing. And it sends data in duplicate by removing redundant data, which, in turn, enhances the accuracy of query processing results. Also, BPA uses a bucket-based data structure in aggregate query processing, and divides and conquers the bucket data structure adaptively according to the number of data in the bucket. In addition, BPA compresses data in order to reduce the size of data in the bucket and performs data transmission filtering when each sensor node sends data. Finally, in this paper, we prove its superiority through various experiments using sensor data.

An Adaptive Learning System based on Learner's Behavior Preferences (학습자 행위 선호도에 기반한 적응적 학습 시스템)

  • Kim, Yong-Se;Cha, Hyun-Jin;Park, Seon-Hee;Cho, Yun-Jung;Yoon, Tae-Bok;Jung, Young-Mo;Lee, Jee-Hyong
    • 한국HCI학회:학술대회논문집
    • /
    • 2006.02a
    • /
    • pp.519-525
    • /
    • 2006
  • Advances in information and telecommunication technology increasingly reveal the potential of computer supported education. However, most computer supported learning systems until recently did not pay much attention to different characteristics of individual learners. Intelligent learning environments adaptive to learner's preferences and tasks are desired. Each learner has different preferences and needs, so it is very crucial to provide the different styles of learners with different learning environments that are more preferred and more efficient to them. This paper reports a study of the intelligent learning environment where the learner's preferences are diagnosed using learner models, and then user interfaces are customized in an adaptive manner to accommodate the preferences. In this research, the learning user interfaces were designed based on a learning-style model by Felder & Silverman, so that different learner preferences are revealed through user interactions with the system. Then, a learning style modeling is done from learner behavior patterns using Decision Tree and Neural Network approaches. In this way, an intelligent learning system adaptive to learning styles can be built. Further research efforts are being made to accommodate various other kinds of learner characteristics such as emotion and motivation as well as learning mastery in providing adaptive learning support.

  • PDF

An Adaptive Clustering Protocol Based on Position of Base-Station for Sensor Networks (센서 네트워크를 위한 싱크 위치 기반의 적응적 클러스터링 프로토콜)

  • Kook, Joong-Jin;Park, Young-Choong;Park, Byoung-Ha;Hong, Ji-Man
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.12
    • /
    • pp.247-255
    • /
    • 2011
  • Most existing clustering protocols have been aimed to provide balancing the residual energy of each node and maximizing life-time of wireless sensor networks. In this paper, we present the adaptive clustering strategy related to sink position for clustering protocols in wireless sensor networks. This protocol allows networks topology to be adaptive to the change of the sink position by using symmetrical clustering strategy that restricts the growth of clusters based on depth of the tree. In addition, it also guarantees each cluster the equal life-time, which may be extended compared with the existing clustering protocols. We evaluated the performance of our clustering scheme comparing to LEACH and EEUC, and observe that our protocol is observed to outperform existing protocols in terms of energy consumption and longevity of the network.

Workflow Pattern Extraction based on ACTA Formalism (ACTA 형식론에 기반한 워크플로우 패턴추출)

  • Lee Wookey;Bae Joonsoo;Jung Jae-yoon
    • Journal of KIISE:Databases
    • /
    • v.32 no.6
    • /
    • pp.603-615
    • /
    • 2005
  • As recent business environments are changed and become complex, a more efficient and effective business process management are needed. This paper proposes a new approach to the automatic execution of business processes using Event-Condition-Action (ECA) rules that can be automatically triggered by an active database. First of all, we propose the concept of blocks that can classify process flows into several patterns. A block is a minimal unit that can specify the behaviors represented in a process model. An algorithm is developed to detect blocks from a process definition network and transform it into a hierarchical tree model. The behaviors in each block type are modeled using ACTA formalism. This provides a theoretical basis from which ECA rules are identified. The proposed ECA rule-based approach shows that it is possible to execute the workflow using the active capability of database without users' intervention.

A Cluster-Based Top-k Query Processing Algorithm in Wireless Sensor Networks (무선 센서 네트워크에서 클러스터 기반의 Top-k 질의 처리)

  • Yeo, Myung-Ho;Seong, Dong-Ook;Yoo, Jae-Soo
    • Journal of KIISE:Databases
    • /
    • v.36 no.4
    • /
    • pp.306-313
    • /
    • 2009
  • Top-k queries are issued to find out the highest (or lowest) readings in many sensor applications. Many top-k query processing algorithms are proposed to reduce energy consumption; FILA installs a filter at each sensor node and suppress unnecessary sensor updates; PRIM allots priorities to sensor nodes and collects the minimal number of sensor reading according to the priorities. However, if many sensor reading converge into the same range of sensor values, it leads to a problem that many false positives are occurred. In this paper, we propose a cluster-based approach to reduce them effectively. Our proposed algorithm operates in two phases: top-k query processing in the cluster level and top-k query processing in the tree level. False positives are effectively filtered out in each level. Performance evaluations show that our proposed algorithm reduces about 70% false positives and achieves about 105% better performance than the existing top-k algorithms in terms of the network lifetime.

Merging XML Documents Based on Insertion/Deletion Edit Operations (삽입/삭제 편집연산 기반의 XML 문서 병합)

  • Lee, Suk-Kyoon
    • The KIPS Transactions:PartD
    • /
    • v.16D no.4
    • /
    • pp.497-506
    • /
    • 2009
  • The method of effectively merging XML documents becomes necessary, as the use of XML is popular and the collaborative editing is required in the areas such as office documents and scientific documents editing works. As a solution to this problem, in this paper we present a theoretical framework for merging individual editing works by muli-users to a same source document. Different from existing approaches which merge documents themselves when they are merged, we represent editing works with a series of edit operations applied to a source document, which is called a edit script, merge those edit scripts by multi-users, and apply the merged one to the source document so that we can achieve the same effect of merging documents. In order to do this, assuming edit scripts based on insertion and deletion edit operations, we define notions such as static edit scripts, the intervention between edit scripts and the conflict between the ones, then propose the conflict conditions between edit scripts and the method of adjusting edit scripts when merged. This approach is effective in reducing network overhead in distributed environments and also in version management systems because of preserving the semantics of individual editing works.

Landslide Susceptibility Mapping by Comparing GIS-based Spatial Models in the Java, Indonesia (GIS 기반 공간예측모델 비교를 통한 인도네시아 자바지역 산사태 취약지도 제작)

  • Kim, Mi-Kyeong;Kim, Sangpil;Nho, Hyunju;Sohn, Hong-Gyoo
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.37 no.5
    • /
    • pp.927-940
    • /
    • 2017
  • Landslide has been a major disaster in Indonesia, and recent climate change and indiscriminate urban development around the mountains have increased landslide risks. Java Island, Indonesia, where more than half of Indonesia's population lives, is experiencing a great deal of damage due to frequent landslides. However, even in such a dangerous situation, the number of inhabitants residing in the landslide-prone area increases year by year, and it is necessary to develop a technique for analyzing landslide-hazardous and vulnerable areas. In this regard, this study aims to evaluate landslide susceptibility of Java, an island of Indonesia, by using GIS-based spatial prediction models. We constructed the geospatial database such as landslide locations, topography, hydrology, soil type, and land cover over the study area and created spatial prediction models by applying Weight of Evidence (WoE), decision trees algorithm and artificial neural network. The three models showed prediction accuracy of 66.95%, 67.04%, and 69.67%, respectively. The results of the study are expected to be useful for prevention of landslide damage for the future and landslide disaster management policies in Indonesia.