• Title/Summary/Keyword: 데이터 융합 관리

Search Result 901, Processing Time 0.029 seconds

Cost Management Optimization Based on RPA for Management Accounting (관리회계실행을 위한 RPA기반 원가관리 최적화 방안)

  • Kim, Kyung-ihl
    • Journal of Convergence for Information Technology
    • /
    • v.10 no.5
    • /
    • pp.8-15
    • /
    • 2020
  • Due to the advance of artificial intelligence, wide use of RPA(Robotic Process Automation) became inevitable. The purpose of this study is to seek cost management optimization based on RPA which has automatic collection of cost information, timeliness and flexibility. The cost management system based on RPA will be able to optimize and improve the cost management process through the cross-system of cost information recognition and the cloud platform. Following the review of previous researches on the benefit of the RPA-related technology along with the investigation on the problems of current cost management system, this study will suggest a way to adopt RPA to optimize cost management system for the implement of strategic management accounting to support management decision making.

A Study on the Dataset Structure of Digital Twin for Disaster and Safety Management (재난안전관리를 위한 디지털 트윈 데이터셋 구조 연구)

  • Ki-Sook Chung;Woo-Sug Jung
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.23 no.5
    • /
    • pp.89-95
    • /
    • 2023
  • The underground utility tunnel is an urban infrastructure that accommodates and manages important facilities such as water and sewage, electricity, and communication in the city, and is a national facility that needs to be protected from disasters such as fire, earthquake, and flooding. In establishing a disaster safety life cycle management system such as prediction, prevention, preparedness, response, and recovery, a disaster safety management platform for underground utility tunnel is being developed by utilizing digital twin technology in which advanced ICT technology and data are converged. In this paper, the maturity model for the disaster safety digital twin was reviewed, and the datasets necessary for implementing the digital twin at each stage were defined.

Finite Element Model Updating Based on Data Fusion of Acceleration and Angular Velocity (가속도 및 각속도 데이터 융합 기반 유한요소모델 개선)

  • Kim, Hyun-Jun;Cho, Soo-Jin;Sim, Sung-Han
    • Journal of the Korea institute for structural maintenance and inspection
    • /
    • v.19 no.2
    • /
    • pp.60-67
    • /
    • 2015
  • The finite element (FE) model updating is a commonly used approach in civil engineering, enabling damage detection, design verification, and load capacity identification. In the FE model updating, acceleration responses are generally employed to determine modal properties of a structure, which are subsequently used to update the initial FE model. While the acceleration-based model updating has been successful in finding better approximations of the physical systems including material and sectional properties, the boundary conditions have been considered yet to be difficult to accurately estimate as the acceleration responses only correspond to translational degree-of-freedoms (DOF). Recent advancement in the sensor technology has enabled low-cost, high-precision gyroscopes that can be adopted in the FE model updating to provide angular information of a structure. This study proposes a FE model updating strategy based on data fusion of acceleration and angular velocity. The usage of both acceleration and angular velocity gives richer information than the sole use of acceleration, allowing the enhanced performance particularly in determining the boundary conditions. A numerical simulation on a simply supported beam is presented to demonstrate the proposed FE model updating approach.

A Study on Concept and Services Framework of Geo-Spatial Big Data (공간 빅데이터의 개념 및 서비스 프레임워크 구상에 관한 연구)

  • Yu, Seon Cheol;Choi, Won Wook;Shin, Dong Bin;Ahn, Jong Wook
    • Spatial Information Research
    • /
    • v.22 no.6
    • /
    • pp.13-21
    • /
    • 2014
  • This study defines concept and service framework of Geo-Spatial Big Data(GSBD). The major concept of the GSBD is formulated based on the 7V characteristics: the general characteristics of big data with 3V(Volume, Variety, Velocity); Geo-spatial oriented characteristics with 4V(Veracity, Visualization, Versatile, Value). GSBD is the technology to extract meaningful information from Geo-spatial fusion data and support decision making responding with rapidly changing activities by analysing with almost realtime solutions while efficiently collecting, storing and managing structured, semi-structured or unstructured big data. The application area of the GSBD is segmented in terms of technical aspect(store, manage, analyze and service) and public/private area. The service framework for the GSBD composed of modules to manage, contain and monitor GSBD services is suggested. Such additional studies as building specific application service models and formulating service delivery strategies for the GSBD are required based on the services framework.

A Study on the MyData Service Model Based on DID Platform (DID 플랫폼 기반의 마이데이터 서비스 모델 연구)

  • Sohyeon Park;Hyunjun Kim;Kanghyo Lee;Tae Gyun Ha;Kyungbaek Kim
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.05a
    • /
    • pp.268-270
    • /
    • 2023
  • 기존 Web2.0 시대의 플랫폼 기업은 서비스를 통해 생성된 개인 데이터로 다양한 비즈니스를 창출해왔다. 하지만 데이터 제공자인 개인은 해당 수익에서 제외되는 모순된 상황에 놓였다. 이에 개인이 자신의 데이터를 적극 관리·통제하면서 능동적으로 활용할 수 있는 개념인 마이데이터(MyData)가 등장했다. 국내에서는 '20.8월 데이터3법(개인정보보호법, 신용정보법, 정보통신망법)이 통과되면서 신용정보법에 근거해 금융 분야 마이데이터 서비스가 활성화되기 시작했다. 그러나 현존하는 마이데이터 플랫폼은 중앙화된 시스템으로 본래 취지와 다르게 개인의 데이터 소유권과 통제권을 보장하기에 부족하다. 이에 본 논문에서는 기존 마이데이터 플랫폼의 한계점을 분석하고, Web3.0 등 변화하는 환경에서 개인의 데이터 주권을 보장하고, 데이터 가치를 공정하게 분배받을 수 있는 DID 플랫폼 기반의 마이데이터 서비스 모델을 제안한다.

Artificial Intelligence and Blockchain Convergence Trend and Policy Improvement Plan (인공지능과 블록체인 융합 동향 및 정책 개선방안)

  • Yang, Hee-Tae
    • Informatization Policy
    • /
    • v.27 no.2
    • /
    • pp.3-19
    • /
    • 2020
  • Artificial intelligence(AI) and blockchain are developing as the core technology leading the Fourth Industrial Revolution. However, AI is still showing limitations in securing and verifying data and explaining the evidence for the results, and blockchain also has some drawbacks such as excessive energy consumption and lack of flexibility in data management. This study analyzed technological limitations of AI and blockchain and convergence trends to overcome them, and finally suggested ways to improve Korea's related policies. Specifically, in terms of R&D reinforcement, we proposed 1) mid- and long-term AI /blockchain convergence research at the national level and 2) blockchain-based AI data platform development. In terms of creating an innovative ecosystem, we also suggested 3) development of AI/blockchain convergence applications by industry, and 4) Start-up support for developing AI/blockchain convergence business models. Lastly, in terms of improving the legal system, we insisted that 5) widening the application of regulatory sandboxes and 6) improving regulations related to privacy protection is necessary.

An empirical study on data governance: Focusing on structural relationships and effects of components (데이터 거버넌스 실증연구: 구성요소 간 구조적 관계와 영향을 중심으로)

  • Yoon, Kun
    • Informatization Policy
    • /
    • v.30 no.3
    • /
    • pp.29-48
    • /
    • 2023
  • This study aims to investigate empirically the structural relationships among the components of data governance and their impacts on data integration and data-based administration. Through literature review, various definitions, typologies, and case studies of data governance were examined, with the definition of data governance from a public policy perspective developed and applied. The study then analyzed the data from a survey conducted by the Korea Institute of Public Administration on the use of public data policies and confirmed that organizational factors play a mediating role between institutional and technical factors, and that institutional and technical factors have statistically significant positive relationships with data fusion and data-driven administration. Based on these results, interest and investment in the improvement and development of the legal system in data governance from the institutional, technical, and organizational perspective, clarification of means and purposes of data technology, interest in data organizations and human resources, and practical operation can be achieved. Policy implications such as the development of an effective mechanism were presented.

A Study on the Necessity of Smart Factory Application in Electronic Components Assembly Process (전자부품 조립공정에서 스마트팩토리 적용 필요성에 대한 연구)

  • Kim, Tae-Jong;Lee, Dong-Yoon
    • Journal of Convergence for Information Technology
    • /
    • v.11 no.9
    • /
    • pp.138-144
    • /
    • 2021
  • In the electronic component assembly business, when product defects occur, it is important to track incoming raw material defects or work defects, and it is important to improve suppliers or work sites according to the results. The core task of the smart factory is to build an integrated data hub to process storage, management, and analysis in real time, and to manage cluster processes, energy, environment, and safety. In order to improve reliability through accurate analysis and collection of production data by real-time monitoring of production site management for electronic parts-related small and medium-sized enterprises (SMEs), the establishment of a smart factory is essential. This paper was developed to be utilized in the construction by defining the system configuration method, smart factory-related technology and application cases, considering the characteristics of SMEs related to electronic components that want to introduce a smart factory.

Research on Countermeasure of SQL Injection Attack (SQL Injection 공격을 효율적으로 방어하는 대응책 연구)

  • Hong, Sunghyuck
    • Journal of the Korea Convergence Society
    • /
    • v.10 no.10
    • /
    • pp.21-26
    • /
    • 2019
  • At present, it is indispensable to utilize data as an information society. Therefore, the database is used to manage large amounts of data. In real life, most of the data in a database is the personal information of a group of members. Because personal information is sensitive data, the role of the database administrator who manages personal information is important. However, there is a growing number of attacks on databases to use this personal information in a malicious way. SQL Injection is one of the most known and old hacking techniques. SQL Injection attacks are known as an easy technique, but countermeasures are easy, but a lot of efforts are made to avoid SQL attacks on web pages that require a lot of logins, but some sites are still vulnerable to SQL attacks. Therefore, this study suggests effective defense measures through analysis of SQL hacking technology cases and contributes to preventing web hacking and providing a secure information communication environment.

Design of An Authentication System Proper for Hybrid Cloud Computing System (하이브리드 클라우드 컴퓨팅 환경에 적합한 인증시스템 설계)

  • Lee, Geuk;Ji, Jae-Won;Chun, Hyun-Woo;Lee, Kyu-Won
    • Convergence Security Journal
    • /
    • v.11 no.6
    • /
    • pp.31-36
    • /
    • 2011
  • Cloud computing is a system which efficiently utilizes resources. In this paper, we propose 2-factor authentication system combing PKI, ID_PW and location information. The proposed method improve the security of hybrid cloud systems and manage resources more safely.