• Title/Summary/Keyword: flow information

Search Result 5,726, Processing Time 0.037 seconds

Semantic Visualization of Dynamic Topic Modeling (다이내믹 토픽 모델링의 의미적 시각화 방법론)

  • Yeon, Jinwook;Boo, Hyunkyung;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.131-154
    • /
    • 2022
  • Recently, researches on unstructured data analysis have been actively conducted with the development of information and communication technology. In particular, topic modeling is a representative technique for discovering core topics from massive text data. In the early stages of topic modeling, most studies focused only on topic discovery. As the topic modeling field matured, studies on the change of the topic according to the change of time began to be carried out. Accordingly, interest in dynamic topic modeling that handle changes in keywords constituting the topic is also increasing. Dynamic topic modeling identifies major topics from the data of the initial period and manages the change and flow of topics in a way that utilizes topic information of the previous period to derive further topics in subsequent periods. However, it is very difficult to understand and interpret the results of dynamic topic modeling. The results of traditional dynamic topic modeling simply reveal changes in keywords and their rankings. However, this information is insufficient to represent how the meaning of the topic has changed. Therefore, in this study, we propose a method to visualize topics by period by reflecting the meaning of keywords in each topic. In addition, we propose a method that can intuitively interpret changes in topics and relationships between or among topics. The detailed method of visualizing topics by period is as follows. In the first step, dynamic topic modeling is implemented to derive the top keywords of each period and their weight from text data. In the second step, we derive vectors of top keywords of each topic from the pre-trained word embedding model. Then, we perform dimension reduction for the extracted vectors. Then, we formulate a semantic vector of each topic by calculating weight sum of keywords in each vector using topic weight of each keyword. In the third step, we visualize the semantic vector of each topic using matplotlib, and analyze the relationship between or among the topics based on the visualized result. The change of topic can be interpreted in the following manners. From the result of dynamic topic modeling, we identify rising top 5 keywords and descending top 5 keywords for each period to show the change of the topic. Existing many topic visualization studies usually visualize keywords of each topic, but our approach proposed in this study differs from previous studies in that it attempts to visualize each topic itself. To evaluate the practical applicability of the proposed methodology, we performed an experiment on 1,847 abstracts of artificial intelligence-related papers. The experiment was performed by dividing abstracts of artificial intelligence-related papers into three periods (2016-2017, 2018-2019, 2020-2021). We selected seven topics based on the consistency score, and utilized the pre-trained word embedding model of Word2vec trained with 'Wikipedia', an Internet encyclopedia. Based on the proposed methodology, we generated a semantic vector for each topic. Through this, by reflecting the meaning of keywords, we visualized and interpreted the themes by period. Through these experiments, we confirmed that the rising and descending of the topic weight of a keyword can be usefully used to interpret the semantic change of the corresponding topic and to grasp the relationship among topics. In this study, to overcome the limitations of dynamic topic modeling results, we used word embedding and dimension reduction techniques to visualize topics by era. The results of this study are meaningful in that they broadened the scope of topic understanding through the visualization of dynamic topic modeling results. In addition, the academic contribution can be acknowledged in that it laid the foundation for follow-up studies using various word embeddings and dimensionality reduction techniques to improve the performance of the proposed methodology.

Analysis of the Impact of Satellite Remote Sensing Information on the Prediction Performance of Ungauged Basin Stream Flow Using Data-driven Models (인공위성 원격 탐사 정보가 자료 기반 모형의 미계측 유역 하천유출 예측성능에 미치는 영향 분석)

  • Seo, Jiyu;Jung, Haeun;Won, Jeongeun;Choi, Sijung;Kim, Sangdan
    • Journal of Wetlands Research
    • /
    • v.26 no.2
    • /
    • pp.147-159
    • /
    • 2024
  • Lack of streamflow observations makes model calibration difficult and limits model performance improvement. Satellite-based remote sensing products offer a new alternative as they can be actively utilized to obtain hydrological data. Recently, several studies have shown that artificial intelligence-based solutions are more appropriate than traditional conceptual and physical models. In this study, a data-driven approach combining various recurrent neural networks and decision tree-based algorithms is proposed, and the utilization of satellite remote sensing information for AI training is investigated. The satellite imagery used in this study is from MODIS and SMAP. The proposed approach is validated using publicly available data from 25 watersheds. Inspired by the traditional regionalization approach, a strategy is adopted to learn one data-driven model by integrating data from all basins, and the potential of the proposed approach is evaluated by using a leave-one-out cross-validation regionalization setting to predict streamflow from different basins with one model. The GRU + Light GBM model was found to be a suitable model combination for target basins and showed good streamflow prediction performance in ungauged basins (The average model efficiency coefficient for predicting daily streamflow in 25 ungauged basins is 0.7187) except for the period when streamflow is very small. The influence of satellite remote sensing information was found to be up to 10%, with the additional application of satellite information having a greater impact on streamflow prediction during low or dry seasons than during wet or normal seasons.

A Study on the Development Trend of Artificial Intelligence Using Text Mining Technique: Focused on Open Source Software Projects on Github (텍스트 마이닝 기법을 활용한 인공지능 기술개발 동향 분석 연구: 깃허브 상의 오픈 소스 소프트웨어 프로젝트를 대상으로)

  • Chong, JiSeon;Kim, Dongsung;Lee, Hong Joo;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.1-19
    • /
    • 2019
  • Artificial intelligence (AI) is one of the main driving forces leading the Fourth Industrial Revolution. The technologies associated with AI have already shown superior abilities that are equal to or better than people in many fields including image and speech recognition. Particularly, many efforts have been actively given to identify the current technology trends and analyze development directions of it, because AI technologies can be utilized in a wide range of fields including medical, financial, manufacturing, service, and education fields. Major platforms that can develop complex AI algorithms for learning, reasoning, and recognition have been open to the public as open source projects. As a result, technologies and services that utilize them have increased rapidly. It has been confirmed as one of the major reasons for the fast development of AI technologies. Additionally, the spread of the technology is greatly in debt to open source software, developed by major global companies, supporting natural language recognition, speech recognition, and image recognition. Therefore, this study aimed to identify the practical trend of AI technology development by analyzing OSS projects associated with AI, which have been developed by the online collaboration of many parties. This study searched and collected a list of major projects related to AI, which were generated from 2000 to July 2018 on Github. This study confirmed the development trends of major technologies in detail by applying text mining technique targeting topic information, which indicates the characteristics of the collected projects and technical fields. The results of the analysis showed that the number of software development projects by year was less than 100 projects per year until 2013. However, it increased to 229 projects in 2014 and 597 projects in 2015. Particularly, the number of open source projects related to AI increased rapidly in 2016 (2,559 OSS projects). It was confirmed that the number of projects initiated in 2017 was 14,213, which is almost four-folds of the number of total projects generated from 2009 to 2016 (3,555 projects). The number of projects initiated from Jan to Jul 2018 was 8,737. The development trend of AI-related technologies was evaluated by dividing the study period into three phases. The appearance frequency of topics indicate the technology trends of AI-related OSS projects. The results showed that the natural language processing technology has continued to be at the top in all years. It implied that OSS had been developed continuously. Until 2015, Python, C ++, and Java, programming languages, were listed as the top ten frequently appeared topics. However, after 2016, programming languages other than Python disappeared from the top ten topics. Instead of them, platforms supporting the development of AI algorithms, such as TensorFlow and Keras, are showing high appearance frequency. Additionally, reinforcement learning algorithms and convolutional neural networks, which have been used in various fields, were frequently appeared topics. The results of topic network analysis showed that the most important topics of degree centrality were similar to those of appearance frequency. The main difference was that visualization and medical imaging topics were found at the top of the list, although they were not in the top of the list from 2009 to 2012. The results indicated that OSS was developed in the medical field in order to utilize the AI technology. Moreover, although the computer vision was in the top 10 of the appearance frequency list from 2013 to 2015, they were not in the top 10 of the degree centrality. The topics at the top of the degree centrality list were similar to those at the top of the appearance frequency list. It was found that the ranks of the composite neural network and reinforcement learning were changed slightly. The trend of technology development was examined using the appearance frequency of topics and degree centrality. The results showed that machine learning revealed the highest frequency and the highest degree centrality in all years. Moreover, it is noteworthy that, although the deep learning topic showed a low frequency and a low degree centrality between 2009 and 2012, their ranks abruptly increased between 2013 and 2015. It was confirmed that in recent years both technologies had high appearance frequency and degree centrality. TensorFlow first appeared during the phase of 2013-2015, and the appearance frequency and degree centrality of it soared between 2016 and 2018 to be at the top of the lists after deep learning, python. Computer vision and reinforcement learning did not show an abrupt increase or decrease, and they had relatively low appearance frequency and degree centrality compared with the above-mentioned topics. Based on these analysis results, it is possible to identify the fields in which AI technologies are actively developed. The results of this study can be used as a baseline dataset for more empirical analysis on future technology trends that can be converged.

A Scalable and Modular Approach to Understanding of Real-time Software: An Architecture-based Software Understanding(ARSU) and the Software Re/reverse-engineering Environment(SRE) (실시간 소프트웨어의 조절적${\cdot}$단위적 이해 방법 : ARSU(Architecture-based Software Understanding)와 SRE(Software Re/reverse-engineering Environment))

  • Lee, Moon-Kun
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.12
    • /
    • pp.3159-3174
    • /
    • 1997
  • This paper reports a research to develop a methodology and a tool for understanding of very large and complex real-time software. The methodology and the tool mostly developed by the author are called the Architecture-based Real-time Software Understanding (ARSU) and the Software Re/reverse-engineering Environment (SRE) respectively. Due to size and complexity, it is commonly very hard to understand the software during reengineering process. However the research facilitates scalable re/reverse-engineering of such real-time software based on the architecture of the software in three-dimensional perspectives: structural, functional, and behavioral views. Firstly, the structural view reveals the overall architecture, specification (outline), and the algorithm (detail) views of the software, based on hierarchically organized parent-chi1d relationship. The basic building block of the architecture is a software Unit (SWU), generated by user-defined criteria. The architecture facilitates navigation of the software in top-down or bottom-up way. It captures the specification and algorithm views at different levels of abstraction. It also shows the functional and the behavioral information at these levels. Secondly, the functional view includes graphs of data/control flow, input/output, definition/use, variable/reference, etc. Each feature of the view contains different kind of functionality of the software. Thirdly, the behavioral view includes state diagrams, interleaved event lists, etc. This view shows the dynamic properties or the software at runtime. Beside these views, there are a number of other documents: capabilities, interfaces, comments, code, etc. One of the most powerful characteristics of this approach is the capability of abstracting and exploding these dimensional information in the architecture through navigation. These capabilities establish the foundation for scalable and modular understanding of the software. This approach allows engineers to extract reusable components from the software during reengineering process.

  • PDF

Geoscientific land management planning in salt-affected areas* (염기화된 지역에서의 지구과학적 토지 관리 계획)

  • Abbott, Simon;Chadwick, David;Street, Greg
    • Geophysics and Geophysical Exploration
    • /
    • v.10 no.1
    • /
    • pp.98-109
    • /
    • 2007
  • Over the last twenty years, farmers in Western Australia have begun to change land management practices to minimise the effects of salinity to agricultural land. A farm plan is often used as a guide to implement changes. Most plans are based on minimal data and an understanding of only surface water flow. Thus farm plans do not effectively address the processes that lead to land salinisation. A project at Broomehill in the south-west of Western Australia applied an approach using a large suite of geospatial data that measured surface and subsurface characteristics of the regolith. In addition, other data were acquired, such as information about the climate and the agricultural history. Fundamental to the approach was the collection of airborne geophysical data over the study area. This included radiometric data reflecting soils, magnetic data reflecting bedrock geology, and SALTMAP electromagnetic data reflecting regolith thickness and conductivity. When interpreted, these datasets added paddock-scale information of geology and hydrogeology to the other datasets, in order to make on-farm and in-paddock decisions relating directly to the mechanisms driving the salinising process. The location and design of surface-water management structures such as grade banks and seepage interceptor banks was significantly influenced by the information derived from the airborne geophysical data. To evaluate the effectiveness ofthis planning., one whole-farm plan has been monitored by the Department of Agriculture and the farmer since 1996. The implemented plan shows a positive cost-benefit ratio, and the farm is now in the top 5% of farms in its regional productivity benchmarking group. The main influence of the airborne geophysical data on the farm plan was on the location of earthworks and revegetation proposals. There had to be a hydrological or hydrogeological justification, based on the site-specific data, for any infrastructure proposal. This approach reduced the spatial density of proposed works compared to other farm plans not guided by site-specific hydrogeological information.

A Dynamic Management Method for FOAF Using RSS and OLAP cube (RSS와 OLAP 큐브를 이용한 FOAF의 동적 관리 기법)

  • Sohn, Jong-Soo;Chung, In-Jeong
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.2
    • /
    • pp.39-60
    • /
    • 2011
  • Since the introduction of web 2.0 technology, social network service has been recognized as the foundation of an important future information technology. The advent of web 2.0 has led to the change of content creators. In the existing web, content creators are service providers, whereas they have changed into service users in the recent web. Users share experiences with other users improving contents quality, thereby it has increased the importance of social network. As a result, diverse forms of social network service have been emerged from relations and experiences of users. Social network is a network to construct and express social relations among people who share interests and activities. Today's social network service has not merely confined itself to showing user interactions, but it has also developed into a level in which content generation and evaluation are interacting with each other. As the volume of contents generated from social network service and the number of connections between users have drastically increased, the social network extraction method becomes more complicated. Consequently the following problems for the social network extraction arise. First problem lies in insufficiency of representational power of object in the social network. Second problem is incapability of expressional power in the diverse connections among users. Third problem is the difficulty of creating dynamic change in the social network due to change in user interests. And lastly, lack of method capable of integrating and processing data efficiently in the heterogeneous distributed computing environment. The first and last problems can be solved by using FOAF, a tool for describing ontology-based user profiles for construction of social network. However, solving second and third problems require a novel technology to reflect dynamic change of user interests and relations. In this paper, we propose a novel method to overcome the above problems of existing social network extraction method by applying FOAF (a tool for describing user profiles) and RSS (a literary web work publishing mechanism) to OLAP system in order to dynamically innovate and manage FOAF. We employed data interoperability which is an important characteristic of FOAF in this paper. Next we used RSS to reflect such changes as time flow and user interests. RSS, a tool for literary web work, provides standard vocabulary for distribution at web sites and contents in the form of RDF/XML. In this paper, we collect personal information and relations of users by utilizing FOAF. We also collect user contents by utilizing RSS. Finally, collected data is inserted into the database by star schema. The system we proposed in this paper generates OLAP cube using data in the database. 'Dynamic FOAF Management Algorithm' processes generated OLAP cube. Dynamic FOAF Management Algorithm consists of two functions: one is find_id_interest() and the other is find_relation (). Find_id_interest() is used to extract user interests during the input period, and find-relation() extracts users matching user interests. Finally, the proposed system reconstructs FOAF by reflecting extracted relationships and interests of users. For the justification of the suggested idea, we showed the implemented result together with its analysis. We used C# language and MS-SQL database, and input FOAF and RSS as data collected from livejournal.com. The implemented result shows that foaf : interest of users has reached an average of 19 percent increase for four weeks. In proportion to the increased foaf : interest change, the number of foaf : knows of users has grown an average of 9 percent for four weeks. As we use FOAF and RSS as basic data which have a wide support in web 2.0 and social network service, we have a definite advantage in utilizing user data distributed in the diverse web sites and services regardless of language and types of computer. By using suggested method in this paper, we can provide better services coping with the rapid change of user interests with the automatic application of FOAF.

Radiation Oncology Digital Image Chart 8nd Digital Radiotherapv Record System at Samsung Medical Center (디지털 화상 병력 시스템과 디지털 방사선치료 기록 시스템의 개발과 사용 경험)

  • Huh Seung Jae;Ahn Yong Chan;Lim Do Hoon;Cho Chung Keun;Kim Dae Yong;Yeo Inhwan;Kim Moon Kyung;Chang Seung Hee;Park Suk Won
    • Radiation Oncology Journal
    • /
    • v.18 no.1
    • /
    • pp.67-72
    • /
    • 2000
  • Background :The authors have developed a Digital image chart(DIC) and digital Radiotherapy Record System (DRRS). We have evaluated the DIC and DRRS for reliability, usefulness, ease of use, and efficiency. Materials and Methods :The basic design of the DIC and DRRS was to build an digital image database of radiation therapy Patient records for a more efficient and timely flow of critical image information throughout the department. This system is a submit of comprehensive radiation oncology management system (C-ROMS) and composed of a picture archiving and communication system (PACS), a radiotherapy information database, and a radiotherapy imaging database. The DIC and DRRS were programmed using Delphi under a Windows 95 environment and is capable of displaying the digital images of patients identification photos, simulation films, radiotherapy setup, diagnostic radiology images, gross lesion Photos, and radiotherapy Planning isodose charts with beam arrangements. Twenty-three clients in the department are connected by Ethernet (10 Mbps) to the central image server (Sun Ultra-sparc 1 workstation). Results :From the introduction of this system in February 1998 through December 1999, we have accumulated a total of 15,732 individual images for 2,556 patients. We can organize radiation therapy in a 'paperless' environment in 120 patients with breast cancer. Using this system, we have succeeded in the prompt, accurate, and simultaneous access to patient care information from multiple locations throughout the department. This coordination has resulted in improved operational efficiency within the department. Conclusion :The authors believe that the DIC and DRRS has contributed to the improvement of radiation oncology department efficacy as well as to time and resource savings by providing necessary visual information throughout the department conveniently and simultaneously. As a result, we can also achieve the 'paperless' and 'filmless' practice of radiation oncology with this system.

  • PDF

A Comparative Study on the Effect of Enterprise SNS on Job Performance - Focused on the Mediation Effect of Communication Level and Moderating Effect of Nationality - (기업용 SNS 이용이 업무성과에 미치는 영향의 국가 간 비교연구 - 커뮤니케이션 수준의 매개효과와 국적의 조절효과를 중심으로 -)

  • Chen, Jing-Yuan;Kwon, Sun-Dong
    • Management & Information Systems Review
    • /
    • v.38 no.4
    • /
    • pp.137-157
    • /
    • 2019
  • Companies are trying to use enterprise SNS for collaboration and speedy decision-making. This study verified the mediating effect of communication between enterprise SNS and job performance, and proved the moderating effect of nationality between enterprise SNS and communication. This study collected survey data of 81 Korean and 81 Chinese from employees who have used enterprise SNS in Korea and China. As results of data analysis, first, enterprise SNS improved job performance through speedy information sharing and error reduction. Second, communication mediated the effect of enterprise SNS on job performance. Third, enterprise SNS increased the level of organizational communication through decreasing the burden of offline face-to-face communication. Compared with Chinese corporate organizations, Korean corporate organizations have high power distances, centralized control, and high superior authority. Therefore, in the off-line communication situation, the subordinate feels the social pressure to follow the command of the superior. Thus communication is one-way and closed. In this Korean organizational situation, corporate SNS can be used as a means to bypass rigid offline communication. In the online communication environment of non face-to-face corporate SNS, anxiety and stress of face-to-face communication can be reduced, so communication between the upper and lower sides can flow more smoothly. The contribution of this paper is that it proved that enterprise SNS promotes communication and improve job performance by reducing the anxiety or stress of offline communication, while according to prior research successful adoption of many types of information systems requires the fit between an organization and its organizational culture.

Development of the forecasting model for import volume by item of major countries based on economic, industrial structural and cultural factors: Focusing on the cultural factors of Korea (경제적, 산업구조적, 문화적 요인을 기반으로 한 주요 국가의 한국 품목별 수입액 예측 모형 개발: 한국의, 한국에 대한 문화적 요인을 중심으로)

  • Jun, Seung-pyo;Seo, Bong-Goon;Park, Do-Hyung
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.4
    • /
    • pp.23-48
    • /
    • 2021
  • The Korean economy has achieved continuous economic growth for the past several decades thanks to the government's export strategy policy. This increase in exports is playing a leading role in driving Korea's economic growth by improving economic efficiency, creating jobs, and promoting technology development. Traditionally, the main factors affecting Korea's exports can be found from two perspectives: economic factors and industrial structural factors. First, economic factors are related to exchange rates and global economic fluctuations. The impact of the exchange rate on Korea's exports depends on the exchange rate level and exchange rate volatility. Global economic fluctuations affect global import demand, which is an absolute factor influencing Korea's exports. Second, industrial structural factors are unique characteristics that occur depending on industries or products, such as slow international division of labor, increased domestic substitution of certain imported goods by China, and changes in overseas production patterns of major export industries. Looking at the most recent studies related to global exchanges, several literatures show the importance of cultural aspects as well as economic and industrial structural factors. Therefore, this study attempted to develop a forecasting model by considering cultural factors along with economic and industrial structural factors in calculating the import volume of each country from Korea. In particular, this study approaches the influence of cultural factors on imports of Korean products from the perspective of PUSH-PULL framework. The PUSH dimension is a perspective that Korea develops and actively promotes its own brand and can be defined as the degree of interest in each country for Korean brands represented by K-POP, K-FOOD, and K-CULTURE. In addition, the PULL dimension is a perspective centered on the cultural and psychological characteristics of the people of each country. This can be defined as how much they are inclined to accept Korean Flow as each country's cultural code represented by the country's governance system, masculinity, risk avoidance, and short-term/long-term orientation. The unique feature of this study is that the proposed final prediction model can be selected based on Design Principles. The design principles we presented are as follows. 1) A model was developed to reflect interest in Korea and cultural characteristics through newly added data sources. 2) It was designed in a practical and convenient way so that the forecast value can be immediately recalled by inputting changes in economic factors, item code and country code. 3) In order to derive theoretically meaningful results, an algorithm was selected that can interpret the relationship between the input and the target variable. This study can suggest meaningful implications from the technical, economic and policy aspects, and is expected to make a meaningful contribution to the export support strategies of small and medium-sized enterprises by using the import forecasting model.

Extraction and Analysis of Ganghwa Tidal Flat Channels Using TanDEM-X DEM (TanDEM-X DEM을 이용한 강화도 갯벌 조류로 추출과 분석)

  • Yun, Ga-Ram;Kim, Lyn;Kim, Nam-Yeong;Kim, Na-Gyeong;Jang, Yun-Yeong;Choi, Yeong-Jin;Lee, Seung-Kuk
    • The Journal of Engineering Geology
    • /
    • v.32 no.3
    • /
    • pp.411-420
    • /
    • 2022
  • Recently, research using remote sensing has been active in various fields such as environment, science, and society. The results of research using remote sensing are not only numerical results, but also play an important role in solving and preventing social and scientific problems. The purpose of this thesis is to tell the correlation between the data provided and each data by using remote sensing technology for the tidal flat environment. The purpose of this study is to obtain high-resolution data using artificial satellites during remote sensing to find out information on tidal flat currents. Tidal flats created by erosion, sedimentation, low tide, and high tide contain information about the tidal flat slope and information about the ecosystem. Therefore, it can be considered as one of the very important studies to analyze the overall tidal flow channel. This paper creates a DEM (Digital Elevation Model) through TanDEM-X, and DEM is used as the most basic data to create a tidal channel. The research area is a tidal flat located in the middle of the west coast of Ganghwado tidal flat. By analyzing the tidal channel created, various information such as the slope direction of Ganghwado tidal flat and the shape of the tidal channel can be grasped. It is expected that the results of this study will increase the importance and necessity of using DEM data for tidal flat research in the future, and that high-quality results can be obtained.