• Title/Summary/Keyword: Open Source Visualization Tool

Search Result 20, Processing Time 0.023 seconds

Y Block Diagram as a New Process Notation in a GPS Manufacture

  • Lee, Jung-Gyu;Jeong, Seung Ryul
    • Journal of Internet Computing and Services
    • /
    • v.20 no.1
    • /
    • pp.125-133
    • /
    • 2019
  • Company A should maintain myriad conversion tools for the purpose of making a geometric compilation of navigation maps. Company A is already using complex compilation tools, which are tailored to geographical areas and various GPS models. However, due to frequent requirement and personnel changes, there is an endless challenge for perfect tool configuration and multiple map consolidation. To solve this problem, Company A launched a process automation project using Graphviz, which is an open source workflow graph visualization software. Before implementation, they had to document their current map compilation processes and then match it with the applicable conversion tool. For effective representation of process controls, a new graphical process notation is designed, i.e. Y Block diagram. The authors will compare Y Block diagram with other process notations and explain why Y Block diagram is more useful for tool based business processes such as digital map generation processes.

A Method for Tool-Chain-driven Quality Control based on Visualization for Small and Medium Scale Software Development Projects (중소규모 SW개발 프로젝트를 위한 시각화 기반의 Tool-Chain 품질관리 방법 제안)

  • Kim, Jung-Bo;Jung, Jin-Young;Kim, Jung-In
    • Journal of Korea Multimedia Society
    • /
    • v.18 no.4
    • /
    • pp.546-556
    • /
    • 2015
  • Since the concept of software engineering was first used in 1968 by NATO Science Committee, a lot of research work and improvements have been made on software development methodology and software quality control, but they still fall short of ensuring successful development of small and medium scale software systems. Under these circumstances, Center for Software Engineering (CSE) at National IT Industry Promotion Agency(NIPA) has been conducting studies on quality control methodologies of software visualization well-suited for small and medium scale software systems, and also working on the systemization and quantification of software quality control. In this paper, we attempt to scope on the software development management of domestic and foreign small and medium-sized enterprises that are lying in the blind spot, compared to large enterprises with well-organized software development systems. In particular, based on software visualization that CSE is pursuing for small and medium-sized developers, we propose a practical quality control methodology well-suited for small and medium scale projects, and a low-cost quality control management tool by combining open-source quality control tools. Our proposal is expected to induce developers' mind change in SI-specialized small and medium-sized software enterprises, increase their profits and improve customer satisfaction through project quality control.

R programming: Language and Environment for Statistical Computing and Data Visualization (R 프로그래밍: 통계 계산과 데이터 시각화를 위한 환경)

  • Lee, D.H.;Ren, Ye
    • Electronics and Telecommunications Trends
    • /
    • v.28 no.1
    • /
    • pp.42-51
    • /
    • 2013
  • The R language is an open source programming language and a software environment for statistical computing and data visualization. The R language is widely used among a lot of statisticians and data scientists to develop statistical software and data analysis. The R language provides a variety of statistical and graphical techniques, including basic descriptive statistics, linear or nonlinear modeling, conventional or advanced statistical tests, time series analysis, clustering, simulation, and others. In this paper, we first introduce the R language and investigate its features as a data analytics tool. As results, we may explore the application possibility of the R language in the field of data analytics.

  • PDF

Research on the Teaching Building-blocks in Elementary Geometry Class using 3D Visualization SW (3D Visualization SW를 활용한 초등학교 쌓기나무 도형교육에 관한 연구)

  • Bae, Hun Joong;Kim, Jong-seong
    • The Journal of the Korea Contents Association
    • /
    • v.17 no.6
    • /
    • pp.71-80
    • /
    • 2017
  • The standards for achievement levels for building blocks in elementary geometry class is to enhance spatial cognitive ability through practices describing shape patterns of building blocks observed from different directions. However, most of building block in the textbook is described from only one perspective. Even worse, some examples in the textbook are almost impossible to observe in the real world. Contrary to this, simulated views by Wings3D has shown that each box may look quite differently from different angles let alone the size of each box. Using Wings3D, it is also very easy to build different types of building blocks with various levels of difficulty in the virtual space. Based on these results, in this study, 3D visualization SW is suggested as a potential pedagogical tool for the elementary geometry class to help kids perceive objects in space more precisely. We have shown that 3D visualization SW such as Wings3D could be a powerful, compact 3D SW for most of subjects which are covered in elementary geometry education. Wings3D has another advantage of economic open source SW fully compatible with school PCs.

Teaching Statistical Graphics using R (R에 의한 통계그래픽스 : 강의 내용 및 방법의 논의)

  • Park, Dong-Ryeon
    • The Korean Journal of Applied Statistics
    • /
    • v.20 no.3
    • /
    • pp.619-634
    • /
    • 2007
  • It is well known that graphical display is critical to data analysis. A lot of research for data visualization has been done, so many effective graphical tools are now available. With the proper use of these graphical tools, we can penetrate the complex structure of data set easily. To enjoy the benefit of the powerful graphical display, the choice of the statistical software is very crucial. R is a popular open source software tool for statistical analysis and graphics, and can provide the very powerful graphics facilities. Moreover, many researchers believe that R is the best software for statistical graphics. In this paper, we would like to discuss what we teach and how we teach in statistical graphics course using R.

On Visualization of Trajectory Data for Traffic Flow Simulation of Urban-scale (도시 스케일의 교통 흐름 시뮬레이션을 위한 궤적 데이터 시각화)

  • Choi, Namshik;Onuean, Athita;Jung, Hanmin
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2018.10a
    • /
    • pp.582-585
    • /
    • 2018
  • As traffic volume increases and road networks become more complicated, identifying for accurate traffic flow and driving smooth traffic flow are a concern of many countries. There are various analytical techniques and studies which desire to study about effective traffic flow. However, the necessary activity is finding the traffic flow pattern through data visualization including location information. In this paper aim to study a real-world urban traffic trajectory and visualize a pattern of traffic flow with a simulation tool. Our experiment is installing the sensor module in 40 taxis and our dataset is generated along 24 hours and unscheduled routes. After pre-processing data, we improved an open source traffic visualize tools to suitable for our experiment. Then we simulate our vehicle trajectory data with a dots animation over a period of time, which allows clearly view a traffic flow simulation and a understand the direction of movement of the vehicle or route pattern. In addition we further propose some novel timelines to show spatial-temporal features to improve an urban environment due to the traffic flow.

  • PDF

A Machine Learning Model Learning and Utilization Education Curriculum for Non-majors (비전공자 대상 머신러닝 모델 학습 및 활용교육 커리큘럼)

  • Kyeong Hur
    • Journal of Practical Engineering Education
    • /
    • v.15 no.1
    • /
    • pp.31-38
    • /
    • 2023
  • In this paper, a basic machine learning model learning and utilization education curriculum for non-majors is proposed, and an education method using Orange machine learning model learning and analysis tools is proposed. Orange is an open-source machine learning and data visualization tool that can create machine learning models by learning data using visual widgets without complex programming. Orange is a platform that is widely used by non-major undergraduates to expert groups. In this paper, a basic machine learning model learning and utilization education curriculum and weekly practice contents for one semester are proposed. In addition, in order to demonstrate the reality of practice contents for machine learning model learning and utilization, we used the Orange tool to learn machine learning models from categorical data samples and numerical data samples, and utilized the models. Thus, use cases for predicting the outcome of the population were proposed. Finally, the educational satisfaction of this curriculum is surveyed and analyzed for non-majors.

Performance Evaluation of Medical Big Data Analysis based on RHadoop (RHadoop 기반 보건의료 빅데이터 분석의 성능 평가)

  • Ryu, Woo-Seok
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.13 no.1
    • /
    • pp.207-212
    • /
    • 2018
  • As a data analysis tool which is becoming popular in the Big Data era, R is rapidly expanding its user range by providing powerful statistical analysis and data visualization functions. Major advantage of R is its functional scalability based on open source, but its scale scalability is limited, resulting in performance degrades in large data processing. RHadoop, one of the extension packages to complement it, can improve data analysis performance as it supports Hadoop platform-based distributed processing of programs written in R. In this paper, we evaluate the validity of RHadoop by evaluating the performance improvement of RHadoop in real medical big data analysis. Performance evaluation of the analysis of the medical history information, which is provided by National Health Insurance Service, using R and RHadoop shows that RHadoop cluster composed of 8 data nodes can improve performance up to 8 times compared with R.

Visualizing the Results of Opinion Mining from Social Media Contents: Case Study of a Noodle Company (소셜미디어 콘텐츠의 오피니언 마이닝결과 시각화: N라면 사례 분석 연구)

  • Kim, Yoosin;Kwon, Do Young;Jeong, Seung Ryul
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.4
    • /
    • pp.89-105
    • /
    • 2014
  • After emergence of Internet, social media with highly interactive Web 2.0 applications has provided very user friendly means for consumers and companies to communicate with each other. Users have routinely published contents involving their opinions and interests in social media such as blogs, forums, chatting rooms, and discussion boards, and the contents are released real-time in the Internet. For that reason, many researchers and marketers regard social media contents as the source of information for business analytics to develop business insights, and many studies have reported results on mining business intelligence from Social media content. In particular, opinion mining and sentiment analysis, as a technique to extract, classify, understand, and assess the opinions implicit in text contents, are frequently applied into social media content analysis because it emphasizes determining sentiment polarity and extracting authors' opinions. A number of frameworks, methods, techniques and tools have been presented by these researchers. However, we have found some weaknesses from their methods which are often technically complicated and are not sufficiently user-friendly for helping business decisions and planning. In this study, we attempted to formulate a more comprehensive and practical approach to conduct opinion mining with visual deliverables. First, we described the entire cycle of practical opinion mining using Social media content from the initial data gathering stage to the final presentation session. Our proposed approach to opinion mining consists of four phases: collecting, qualifying, analyzing, and visualizing. In the first phase, analysts have to choose target social media. Each target media requires different ways for analysts to gain access. There are open-API, searching tools, DB2DB interface, purchasing contents, and so son. Second phase is pre-processing to generate useful materials for meaningful analysis. If we do not remove garbage data, results of social media analysis will not provide meaningful and useful business insights. To clean social media data, natural language processing techniques should be applied. The next step is the opinion mining phase where the cleansed social media content set is to be analyzed. The qualified data set includes not only user-generated contents but also content identification information such as creation date, author name, user id, content id, hit counts, review or reply, favorite, etc. Depending on the purpose of the analysis, researchers or data analysts can select a suitable mining tool. Topic extraction and buzz analysis are usually related to market trends analysis, while sentiment analysis is utilized to conduct reputation analysis. There are also various applications, such as stock prediction, product recommendation, sales forecasting, and so on. The last phase is visualization and presentation of analysis results. The major focus and purpose of this phase are to explain results of analysis and help users to comprehend its meaning. Therefore, to the extent possible, deliverables from this phase should be made simple, clear and easy to understand, rather than complex and flashy. To illustrate our approach, we conducted a case study on a leading Korean instant noodle company. We targeted the leading company, NS Food, with 66.5% of market share; the firm has kept No. 1 position in the Korean "Ramen" business for several decades. We collected a total of 11,869 pieces of contents including blogs, forum contents and news articles. After collecting social media content data, we generated instant noodle business specific language resources for data manipulation and analysis using natural language processing. In addition, we tried to classify contents in more detail categories such as marketing features, environment, reputation, etc. In those phase, we used free ware software programs such as TM, KoNLP, ggplot2 and plyr packages in R project. As the result, we presented several useful visualization outputs like domain specific lexicons, volume and sentiment graphs, topic word cloud, heat maps, valence tree map, and other visualized images to provide vivid, full-colored examples using open library software packages of the R project. Business actors can quickly detect areas by a swift glance that are weak, strong, positive, negative, quiet or loud. Heat map is able to explain movement of sentiment or volume in categories and time matrix which shows density of color on time periods. Valence tree map, one of the most comprehensive and holistic visualization models, should be very helpful for analysts and decision makers to quickly understand the "big picture" business situation with a hierarchical structure since tree-map can present buzz volume and sentiment with a visualized result in a certain period. This case study offers real-world business insights from market sensing which would demonstrate to practical-minded business users how they can use these types of results for timely decision making in response to on-going changes in the market. We believe our approach can provide practical and reliable guide to opinion mining with visualized results that are immediately useful, not just in food industry but in other industries as well.

Twitter Issue Tracking System by Topic Modeling Techniques (토픽 모델링을 이용한 트위터 이슈 트래킹 시스템)

  • Bae, Jung-Hwan;Han, Nam-Gi;Song, Min
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.2
    • /
    • pp.109-122
    • /
    • 2014
  • People are nowadays creating a tremendous amount of data on Social Network Service (SNS). In particular, the incorporation of SNS into mobile devices has resulted in massive amounts of data generation, thereby greatly influencing society. This is an unmatched phenomenon in history, and now we live in the Age of Big Data. SNS Data is defined as a condition of Big Data where the amount of data (volume), data input and output speeds (velocity), and the variety of data types (variety) are satisfied. If someone intends to discover the trend of an issue in SNS Big Data, this information can be used as a new important source for the creation of new values because this information covers the whole of society. In this study, a Twitter Issue Tracking System (TITS) is designed and established to meet the needs of analyzing SNS Big Data. TITS extracts issues from Twitter texts and visualizes them on the web. The proposed system provides the following four functions: (1) Provide the topic keyword set that corresponds to daily ranking; (2) Visualize the daily time series graph of a topic for the duration of a month; (3) Provide the importance of a topic through a treemap based on the score system and frequency; (4) Visualize the daily time-series graph of keywords by searching the keyword; The present study analyzes the Big Data generated by SNS in real time. SNS Big Data analysis requires various natural language processing techniques, including the removal of stop words, and noun extraction for processing various unrefined forms of unstructured data. In addition, such analysis requires the latest big data technology to process rapidly a large amount of real-time data, such as the Hadoop distributed system or NoSQL, which is an alternative to relational database. We built TITS based on Hadoop to optimize the processing of big data because Hadoop is designed to scale up from single node computing to thousands of machines. Furthermore, we use MongoDB, which is classified as a NoSQL database. In addition, MongoDB is an open source platform, document-oriented database that provides high performance, high availability, and automatic scaling. Unlike existing relational database, there are no schema or tables with MongoDB, and its most important goal is that of data accessibility and data processing performance. In the Age of Big Data, the visualization of Big Data is more attractive to the Big Data community because it helps analysts to examine such data easily and clearly. Therefore, TITS uses the d3.js library as a visualization tool. This library is designed for the purpose of creating Data Driven Documents that bind document object model (DOM) and any data; the interaction between data is easy and useful for managing real-time data stream with smooth animation. In addition, TITS uses a bootstrap made of pre-configured plug-in style sheets and JavaScript libraries to build a web system. The TITS Graphical User Interface (GUI) is designed using these libraries, and it is capable of detecting issues on Twitter in an easy and intuitive manner. The proposed work demonstrates the superiority of our issue detection techniques by matching detected issues with corresponding online news articles. The contributions of the present study are threefold. First, we suggest an alternative approach to real-time big data analysis, which has become an extremely important issue. Second, we apply a topic modeling technique that is used in various research areas, including Library and Information Science (LIS). Based on this, we can confirm the utility of storytelling and time series analysis. Third, we develop a web-based system, and make the system available for the real-time discovery of topics. The present study conducted experiments with nearly 150 million tweets in Korea during March 2013.