• Title/Summary/Keyword: Science and Technology Data

Search Result 15,532, Processing Time 0.057 seconds

Data Technology: New Interdisciplinary Science & Technology (데이터 기술: 지식창조를 위한 새로운 융합과학기술)

  • Park, Sung-Hyun
    • Journal of Korean Society for Quality Management
    • /
    • v.38 no.3
    • /
    • pp.294-312
    • /
    • 2010
  • Data Technology (DT) is a new technology which deals with data collection, data analysis, information generation from data, knowledge generation from modelling and future prediction. DT is a newly emerged interdisciplinary science & technology in this 21st century knowledge society. Even though the main body of DT is applied statistics, it also contains management information system (MIS), quality management, process system analysis and so on. Therefore, it is an interdisciplinary science and technology of statistics, management science, industrial engineering, computer science and social science. In this paper, first of all, the definition of DT is given, and then the effects and the basic properties of DT, the differences between IT and DT, the 6 step process for DT application, and a DT example are provided. Finally, the relationship among DT, e-Statistics and Data Mining is explained, and the direction of DT development is proposed.

Hybrid Recommendation Algorithm for User Satisfaction-oriented Privacy Model

  • Sun, Yinggang;Zhang, Hongguo;Zhang, Luogang;Ma, Chao;Huang, Hai;Zhan, Dongyang;Qu, Jiaxing
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.10
    • /
    • pp.3419-3437
    • /
    • 2022
  • Anonymization technology is an important technology for privacy protection in the process of data release. Usually, before publishing data, the data publisher needs to use anonymization technology to anonymize the original data, and then publish the anonymized data. However, for data publishers who do not have or have less anonymized technical knowledge background, how to configure appropriate parameters for data with different characteristics has become a more difficult problem. In response to this problem, this paper adds a historical configuration scheme resource pool on the basis of the traditional anonymization process, and configuration parameters can be automatically recommended through the historical configuration scheme resource pool. On this basis, a privacy model hybrid recommendation algorithm for user satisfaction is formed. The algorithm includes a forward recommendation process and a reverse recommendation process, which can respectively perform data anonymization processing for users with different anonymization technical knowledge backgrounds. The privacy model hybrid recommendation algorithm for user satisfaction described in this paper is suitable for a wider population, providing a simpler, more efficient and automated solution for data anonymization, reducing data processing time and improving the quality of anonymized data, which enhances data protection capabilities.

Development of a distributed high-speed data acquisition and monitoring system based on a special data packet format for HUST RF negative ion source

  • Li, Dong;Yin, Ling;Wang, Sai;Zuo, Chen;Chen, Dezhi
    • Nuclear Engineering and Technology
    • /
    • v.54 no.10
    • /
    • pp.3587-3594
    • /
    • 2022
  • A distributed high-speed data acquisition and monitoring system for the RF negative ion source at Huazhong University of Science and Technology (HUST) is developed, which consists of data acquisition, data forwarding and data processing. Firstly, the data acquisition modules sample physical signals at high speed and upload the sampling data with corresponding absolute-time labels over UDP, which builds the time correlation among different signals. And a special data packet format is proposed for the data upload, which is convenient for packing or parsing a fixed-length packet, especially when the span of the time labels in a packet crosses an absolute second. The data forwarding modules then receive the UDP messages and distribute their data packets to the real-time display module and the data storage modules by PUB/SUB-pattern message queue of ZeroMQ. As for the data storage, a scheme combining the file server and MySQL database is adopted to increase the storage rate and facilitate the data query. The test results show that the loss rate of the data packets is within the range of 0-5% and the storage rate is higher than 20 Mbps, both acceptable for the HUST RF negative ion source.

Improving Utilization of GPS Data for Urban Traffic Applications

  • Nguyen, Duc Hai;Nguyen, Tan Phuc;Doan, Khue;Ta, Ho Thai Hai;Pham, Tran Vu;Huynh, Nam;Le, Thanh Van
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.7 no.1
    • /
    • pp.6-9
    • /
    • 2015
  • The use of Intelligent Transportation System (ITS) is promising to bring better solutions for managing and handling the city traffic. This system combines many fields in advanced technology such as Global Positioning System (GPS), Geographic Information System (GIS) and so on. The basement of applications in ITS is the effective collections and data integration tools. The purpose of our research is to propose solutions which involve the use of GPS time series data collected from GPS devices in order to improve the quality of output traffic data. In this study, GPS data is collected from devices attached to vehicles travelling on routes in Ho Chi Minh City (HCMC). Then, GPS data is stored in database system to serve in many transportation applications. The proposed method combines the data usage level and data coverage to improve the quality of traffic data.

Verification Control Algorithm of Data Integrity Verification in Remote Data sharing

  • Xu, Guangwei;Li, Shan;Lai, Miaolin;Gan, Yanglan;Feng, Xiangyang;Huang, Qiubo;Li, Li;Li, Wei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.2
    • /
    • pp.565-586
    • /
    • 2022
  • Cloud storage's elastic expansibility not only provides flexible services for data owners to store their data remotely, but also reduces storage operation and management costs of their data sharing. The data outsourced remotely in the storage space of cloud service provider also brings data security concerns about data integrity. Data integrity verification has become an important technology for detecting the integrity of remote shared data. However, users without data access rights to verify the data integrity will cause unnecessary overhead to data owner and cloud service provider. Especially malicious users who constantly launch data integrity verification will greatly waste service resources. Since data owner is a consumer purchasing cloud services, he needs to bear both the cost of data storage and that of data verification. This paper proposes a verification control algorithm in data integrity verification for remotely outsourced data. It designs an attribute-based encryption verification control algorithm for multiple verifiers. Moreover, data owner and cloud service provider construct a common access structure together and generate a verification sentinel to verify the authority of verifiers according to the access structure. Finally, since cloud service provider cannot know the access structure and the sentry generation operation, it can only authenticate verifiers with satisfying access policy to verify the data integrity for the corresponding outsourced data. Theoretical analysis and experimental results show that the proposed algorithm achieves fine-grained access control to multiple verifiers for the data integrity verification.

Development of a National Research Data Platform for Sharing and Utilizing Research Data

  • Shin, Youngho;Um, Jungho;Seo, Dongmin;Shin, Sungho
    • Journal of Information Science Theory and Practice
    • /
    • v.10 no.spc
    • /
    • pp.25-38
    • /
    • 2022
  • Research data means data used or created in the course of research or experiments. Research data is very important for validation of research conducted and for use in future research and projects. Recently, convergence research between various fields and international cooperation has been continuously done due to the explosive increase of research data and the increase in the complexity of science and technology. Developed countries are actively promoting open science policies that share research results and processes to create new knowledge and values through convergence research. Communities to promote the sharing and utilization of research data such as RDA (Research Data Alliance) and COAR (Confederation of Open Access Repositories) are active, and various platforms for managing and sharing research data are being developed and used. OpenAIRE (Open Access Infrastructure for Research In Europe), a research data platform in Europe, ARDC (Australian Research Data Commons) in Australia, and IRDB (Institutional Repositories DataBase) in Japan provide research data or research data related services. Korea has been establishing and implementing a research data sharing and utilization strategy to promote the sharing and utilization of research data at the national level, led by the central government. Based on this strategy, KISTI has been building a Korean research data platform (DataON) since 2018, and has been providing research data sharing and utilization services to users since January 2020. This paper reviews the characteristics of DataON and how it is used for research by showing its applications.

History and Trends of Data Education in Korea - KISTI Data Education Based on 2001-2019 Statistics

  • Min, Jaehong;Han, Sunggeun;Ahn, Bu-young
    • Journal of Internet Computing and Services
    • /
    • v.21 no.6
    • /
    • pp.133-139
    • /
    • 2020
  • Big data, artificial intelligence (AI), and machine learning are keywords that represent the Fourth industrial Revolution. In addition, as the development of science and technology, the Korean government, public institutions and industries want professionals who can collect, analyze, utilize and predict data. This means that data analysis and utilization education become more important. Education on data analysis and utilization is increasing with trends in other academy. However, it is true that not many academy run long-term and systematic education. Korea Institute of Science and Technology Information (KISTI) is a data ecosystem hub and one of its performance missions has been providing data utilization and analysis education to meet the needs of industries, institutions and governments since 1966. In this study, KISTI's data education was analyzed using the number of curriculum trainees per year from 2001 to 2019. With this data, the change of interest in education in information and data field was analyzed by reflecting social and historical situations. And we identified the characteristics of KISTI and trainees. It means that the identity, characteristics, infrastructure, and resources of the institution have a greater impact on the trainees' interest of data-use education.In particular, KISTI, as a research institute, conducts research in various fields, including bio, weather, traffic, disaster and so on. And it has various research data in science and technology field. The purpose of this study can provide direction forthe establishment of new curriculum using data that can represent KISTI's strengths and identity. One of the conclusions of this paper would be KISTI's greatest advantages if it could be used in education to analyze and visualize many research data. Finally, through this study, it can expect that KISTI will be able to present a new direction for designing data curricula with quality education that can fulfill its role and responsibilities and highlight its strengths.

Data Framework Design of EDISON 2.0 Digital Platform for Convergence Research

  • Sunggeun Han;Jaegwang Lee;Inho Jeon;Jeongcheol Lee;Hoon Choi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.8
    • /
    • pp.2292-2313
    • /
    • 2023
  • With improving computing performance, various digital platforms are being developed to enable easily utilization of high-performance computing environments. EDISON 1.0 is an online simulation platform widely used in computational science and engineering education. As the research paradigm changes, the demand for developing the EDISON 1.0 platform centered on simulation into the EDISON 2.0 platform centered on data and artificial intelligence is growing. Herein, a data framework, a core module for data-centric research on EDISON 2.0 digital platform, is proposed. The proposed data framework provides the following three functions. First, it provides a data repository suitable for the data lifecycle to increase research reproducibility. Second, it provides a new data model that can integrate, manage, search, and utilize heterogeneous data to support a data-driven interdisciplinary convergence research environment. Finally, it provides an exploratory data analysis (EDA) service and data enrichment using an AI model, both developed to strengthen data reliability and maximize the efficiency and effectiveness of research endeavors. Using the EDISON 2.0 data framework, researchers can conduct interdisciplinary convergence research using heterogeneous data and easily perform data pre-processing through the web-based UI. Further, it presents the opportunity to leverage the derived data obtained through AI technology to gain insights and create new research topics.

LDBAS: Location-aware Data Block Allocation Strategy for HDFS-based Applications in the Cloud

  • Xu, Hua;Liu, Weiqing;Shu, Guansheng;Li, Jing
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.1
    • /
    • pp.204-226
    • /
    • 2018
  • Big data processing applications have been migrated into cloud gradually, due to the advantages of cloud computing. Hadoop Distributed File System (HDFS) is one of the fundamental support systems for big data processing on MapReduce-like frameworks, such as Hadoop and Spark. Since HDFS is not aware of the co-location of virtual machines in the cloud, the default scheme of block allocation in HDFS does not fit well in the cloud environments behaving in two aspects: data reliability loss and performance degradation. In this paper, we present a novel location-aware data block allocation strategy (LDBAS). LDBAS jointly optimizes data reliability and performance for upper-layer applications by allocating data blocks according to the locations and different processing capacities of virtual nodes in the cloud. We apply LDBAS to two stages of data allocation of HDFS in the cloud (the initial data allocation and data recovery), and design the corresponding algorithms. Finally, we implement LDBAS into an actual Hadoop cluster and evaluate the performance with the benchmark suite BigDataBench. The experimental results show that LDBAS can guarantee the designed data reliability while reducing the job execution time of the I/O-intensive applications in Hadoop by 8.9% on average and up to 11.2% compared with the original Hadoop in the cloud.

Analysis of the Present Status and Future Prospects for Smart Agriculture Technologies in South Korea Using National R&D Project Data

  • Lee, Sujin;Park, Jun-Hwan;Kim, EunSun;Jang, Wooseok
    • Journal of Information Science Theory and Practice
    • /
    • v.10 no.spc
    • /
    • pp.112-122
    • /
    • 2022
  • Food security and its sovereignty have become among the most important key issues due to changes in the international situation. Regarding these issues, many countries now give attention to smart agriculture, which would increase production efficiency through a data-based system. The Korean government also has attempted to promote smart agriculture by 1) implementing the agri-food ICT (information and communications technology) policy, and 2) increasing the R&D budget by more than double in recent years. However, its endeavors only centered on large-scale farms which a number of domestic farmers rarely utilized in their farming. To promote smart agriculture more effectively, we diagnosed the government R&D trends of smart agriculture based on NTIS (National Science and Technology Information Service) data. We identified the research trends for each R&D period by analyzing three pieces of information: the regional information, research actor, and topic. Based on these findings, we could suggest systematic R&D directions and implications.