• Title/Summary/Keyword: 산업 클라우드

Search Result 422, Processing Time 0.024 seconds

Trustworthy AI Framework for Malware Response (악성코드 대응을 위한 신뢰할 수 있는 AI 프레임워크)

  • Shin, Kyounga;Lee, Yunho;Bae, ByeongJu;Lee, Soohang;Hong, Heeju;Choi, Youngjin;Lee, Sangjin
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.32 no.5
    • /
    • pp.1019-1034
    • /
    • 2022
  • Malware attacks become more prevalent in the hyper-connected society of the 4th industrial revolution. To respond to such malware, automation of malware detection using artificial intelligence technology is attracting attention as a new alternative. However, using artificial intelligence without collateral for its reliability poses greater risks and side effects. The EU and the United States are seeking ways to secure the reliability of artificial intelligence, and the government announced a reliable strategy for realizing artificial intelligence in 2021. The government's AI reliability has five attributes: Safety, Explainability, Transparency, Robustness and Fairness. We develop four elements of safety, explainable, transparent, and fairness, excluding robustness in the malware detection model. In particular, we demonstrated stable generalization performance, which is model accuracy, through the verification of external agencies, and developed focusing on explainability including transparency. The artificial intelligence model, of which learning is determined by changing data, requires life cycle management. As a result, demand for the MLops framework is increasing, which integrates data, model development, and service operations. EXE-executable malware and documented malware response services become data collector as well as service operation at the same time, and connect with data pipelines which obtain information for labeling and purification through external APIs. We have facilitated other security service associations or infrastructure scaling using cloud SaaS and standard APIs.

Efficient QoS Policy Implementation Using DSCP Redefinition: Towards Network Load Balancing (DSCP 재정의를 통한 효율적인 QoS 정책 구현: 네트워크 부하 분산을 위해)

  • Hanwoo Lee;Suhwan Kim;Gunwoo Park
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.3
    • /
    • pp.715-720
    • /
    • 2023
  • The military is driving innovative changes such as AI, cloud computing, and drone operation through the Fourth Industrial Revolution. It is expected that such changes will lead to a rapid increase in the demand for information exchange requirements, reaching all lower-ranking soldiers, as networking based on IoT occurs. The flow of such information must ensure efficient information distribution through various infrastructures such as ground networks, stationary satellites, and low-earth orbit small communication satellites, and the demand for information exchange that is distributed through them must be appropriately dispersed. In this study, we redefined the DSCP, which is closely related to QoS (Quality of Service) in information dissemination, into 11 categories and performed research to map each cluster group identified by cluster analysis to the defense "information exchange requirement list" on a one-to-one basis. The purpose of the research is to ensure efficient information dissemination within a multi-layer integrated network (ground network, stationary satellite network, low-earth orbit small communication satellite network) with limited bandwidth by re-establishing QoS policies that prioritize important information exchange requirements so that they are routed in priority. In this paper, we evaluated how well the information exchange requirement lists classified by cluster analysis were assigned to DSCP through M&S, and confirmed that reclassifying DSCP can lead to more efficient information distribution in a network environment with limited bandwidth.

A Study on the Metadata Schema for the Collection of Sensor Data in Weapon Systems (무기체계 CBM+ 적용 및 확대를 위한 무기체계 센서데이터 수집용 메타데이터 스키마 연구)

  • Jinyoung Kim;Hyoung-seop Shim;Jiseong Son;Yun-Young Hwang
    • Journal of Internet Computing and Services
    • /
    • v.24 no.6
    • /
    • pp.161-169
    • /
    • 2023
  • Due to the Fourth Industrial Revolution, innovation in various technologies such as artificial intelligence (AI), big data (Big Data), and cloud (Cloud) is accelerating, and data is considered an important asset. With the innovation of these technologies, various efforts are being made to lead technological innovation in the field of defense science and technology. In Korea, the government also announced the "Defense Innovation 4.0 Plan," which consists of five key points and 16 tasks to foster advanced science and technology forces in March 2023. The plan also includes the establishment of a Condition-Based Maintenance system (CBM+) to improve the operability and availability of weapons systems and reduce defense costs. Condition Based Maintenance (CBM) aims to secure the reliability and availability of the weapon system and analyze changes in equipment's state information to identify them as signs of failure and defects, and CBM+ is a concept that adds Remaining Useful Life prediction technology to the existing CBM concept [1]. In order to establish a CBM+ system for the weapon system, sensors are installed and sensor data are required to obtain condition information of the weapon system. In this paper, we propose a sensor data metadata schema to efficiently and effectively manage sensor data collected from sensors installed in various weapons systems.

Development of Metrics to Measure Reusability Quality of AIaaS

  • Eun-Sook Cho
    • Journal of the Korea Society of Computer and Information
    • /
    • v.28 no.12
    • /
    • pp.147-153
    • /
    • 2023
  • As it spreads to all industries of artificial intelligence technology, AIaaS equipped with artificial intelligence services is emerging. In particular, non-IT companies are suffering from the absence of software experts, difficulties in training big data models, and difficulties in collecting and analyzing various types of data. AIaaS makes it easier and more economical for users to build a system by providing various IT resources necessary for artificial intelligence software development as well as functions necessary for artificial intelligence software in the form of a service. Therefore, the supply and demand for such cloud-based AIaaS services will increase rapidly. However, the quality of services provided by AIaaS becomes an important factor in what is required as the supply and demand for AIaaS increases. However, research on a comprehensive and practical quality evaluation metric to measure this is currently insufficient. Therefore, in this paper, we develop and propose a usability, replacement, scalability, and publicity metric, which are the four metrics necessary for measuring reusability, based on implementation, convenience, efficiency, and accessibility, which are characteristics of AIaaS, for reusability evaluation among the service quality measurement factors of AIaaS. The proposed metrics can be used as a tool to predict how much services provided by AIaaS can be reused for potential users in the future.

Analysis of media trends related to spent nuclear fuel treatment technology using text mining techniques (텍스트마이닝 기법을 활용한 사용후핵연료 건식처리기술 관련 언론 동향 분석)

  • Jeong, Ji-Song;Kim, Ho-Dong
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.2
    • /
    • pp.33-54
    • /
    • 2021
  • With the fourth industrial revolution and the arrival of the New Normal era due to Corona, the importance of Non-contact technologies such as artificial intelligence and big data research has been increasing. Convergent research is being conducted in earnest to keep up with these research trends, but not many studies have been conducted in the area of nuclear research using artificial intelligence and big data-related technologies such as natural language processing and text mining analysis. This study was conducted to confirm the applicability of data science analysis techniques to the field of nuclear research. Furthermore, the study of identifying trends in nuclear spent fuel recognition is critical in terms of being able to determine directions to nuclear industry policies and respond in advance to changes in industrial policies. For those reasons, this study conducted a media trend analysis of pyroprocessing, a spent nuclear fuel treatment technology. We objectively analyze changes in media perception of spent nuclear fuel dry treatment techniques by applying text mining analysis techniques. Text data specializing in Naver's web news articles, including the keywords "Pyroprocessing" and "Sodium Cooled Reactor," were collected through Python code to identify changes in perception over time. The analysis period was set from 2007 to 2020, when the first article was published, and detailed and multi-layered analysis of text data was carried out through analysis methods such as word cloud writing based on frequency analysis, TF-IDF and degree centrality calculation. Analysis of the frequency of the keyword showed that there was a change in media perception of spent nuclear fuel dry treatment technology in the mid-2010s, which was influenced by the Gyeongju earthquake in 2016 and the implementation of the new government's energy conversion policy in 2017. Therefore, trend analysis was conducted based on the corresponding time period, and word frequency analysis, TF-IDF, degree centrality values, and semantic network graphs were derived. Studies show that before the 2010s, media perception of spent nuclear fuel dry treatment technology was diplomatic and positive. However, over time, the frequency of keywords such as "safety", "reexamination", "disposal", and "disassembly" has increased, indicating that the sustainability of spent nuclear fuel dry treatment technology is being seriously considered. It was confirmed that social awareness also changed as spent nuclear fuel dry treatment technology, which was recognized as a political and diplomatic technology, became ambiguous due to changes in domestic policy. This means that domestic policy changes such as nuclear power policy have a greater impact on media perceptions than issues of "spent nuclear fuel processing technology" itself. This seems to be because nuclear policy is a socially more discussed and public-friendly topic than spent nuclear fuel. Therefore, in order to improve social awareness of spent nuclear fuel processing technology, it would be necessary to provide sufficient information about this, and linking it to nuclear policy issues would also be a good idea. In addition, the study highlighted the importance of social science research in nuclear power. It is necessary to apply the social sciences sector widely to the nuclear engineering sector, and considering national policy changes, we could confirm that the nuclear industry would be sustainable. However, this study has limitations that it has applied big data analysis methods only to detailed research areas such as "Pyroprocessing," a spent nuclear fuel dry processing technology. Furthermore, there was no clear basis for the cause of the change in social perception, and only news articles were analyzed to determine social perception. Considering future comments, it is expected that more reliable results will be produced and efficiently used in the field of nuclear policy research if a media trend analysis study on nuclear power is conducted. Recently, the development of uncontact-related technologies such as artificial intelligence and big data research is accelerating in the wake of the recent arrival of the New Normal era caused by corona. Convergence research is being conducted in earnest in various research fields to follow these research trends, but not many studies have been conducted in the nuclear field with artificial intelligence and big data-related technologies such as natural language processing and text mining analysis. The academic significance of this study is that it was possible to confirm the applicability of data science analysis technology in the field of nuclear research. Furthermore, due to the impact of current government energy policies such as nuclear power plant reductions, re-evaluation of spent fuel treatment technology research is undertaken, and key keyword analysis in the field can contribute to future research orientation. It is important to consider the views of others outside, not just the safety technology and engineering integrity of nuclear power, and further reconsider whether it is appropriate to discuss nuclear engineering technology internally. In addition, if multidisciplinary research on nuclear power is carried out, reasonable alternatives can be prepared to maintain the nuclear industry.

The Need and Improvement Direction of New Computer Media Classes in Landscape Architectural Education in University (대학 내 조경전공 교육과정에 있어 새로운 컴퓨터 미디어 수업의 필요와 개선방향)

  • Na, Sungjin
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.49 no.1
    • /
    • pp.54-69
    • /
    • 2021
  • In 2020, civilized society's overall lifestyle showed a distinct change from consumable analog media, such as paper, to digital media with the increased penetration of cloud computing, and from wired media to wireless media. Based on these social changes, this work examines whether the use of computer media in the field of landscape architecture is appropriately applied. This study will give directions for new computer media classes in landscape architectural education in the 4th Industrial Revolution era. Landscape architecture is a field that directly proposes the realization of a positive lifestyle and the creation of a living environment and is closely connected with social change. However, there is no clear evidence that landscape architectural education is making any visible change, while the digital infrastructure of the 4th Industrial Revolution, such as Artificial Intelligence (AI), Big Data, autonomous vehicles, cloud networks, and the Internet of Things, is changing the contemporary society in terms of technology, culture, and economy among other aspects. Therefore, it is necessary to review the current state of the use of computer technology and media in landscape architectural education, and also to examine the alternative direction of the curriculum for the new digital era. First, the basis for discussion was made by studying the trends of computational design in modern landscape architecture. Next, the changes and current status of computer media classes in domestic and overseas landscape education were analyzed based on prior research and curriculum. As a result, the number and the types of computer media classes increased significantly between the study in 1994 and the current situation in 2020 in the foreign landscape department, whereas there were no obvious changes in the domestic landscape department. This shows that the domestic landscape education is passively coping with the changes in the digital era. Lastly, based on the discussions, this study examined alternatives to the new curriculum that landscape architecture department should pursue in a new degital world.

Strategies for Increasing the Value and Sustainability of Archaeological Education in the Post-COVID-19 Era (포스트 코로나 시대 고고유산 교육의 가치와 지속가능성을 위한 전략)

  • KIM, Eunkyung
    • Korean Journal of Heritage: History & Science
    • /
    • v.55 no.2
    • /
    • pp.82-100
    • /
    • 2022
  • With the crisis of the COVID-19 pandemic and the era of the 4th industrial revolution, archaeological heritage education has entered a new phase. This article responds to the trends in the post-COVID-19 era, seeking ways to develop archaeological heritage education and sustainable strategies necessary in the era of the 4th industrial revolution. The program of archaeological heritage education required in the era of the 4th industrial revolution must cultivate creative talent, solve problems, and improve self-efficacy. It should also draw attention to archaeological heritage maker education. Such maker education should be delivered based on constructivism and be designed by setting specific learning goals in consideration of various age-specific characteristics. Moreover, various ICT-based contents applying VR, AR, cloud, and drone imaging technologies should be developed and expanded, and, above all, ontact digital education(real-time virtual learning) should seek ways to revitalize communities capable of interactive communication in non-face-to-face situations. The development of such ancient heritage content needs to add AI functions that consider learners' interests, learning abilities, and learning purposes while producing various convergent contents from the standpoint of "cultural collage." Online archaeological heritage content education should be delivered following prior learning or with supplementary learning in consideration of motivation or field learning to access the real thing in the future. Ultimately, archaeological ontact education will be delivered using cutting-edge technologies that reflect the current trends. In conjunction with this, continuous efforts are needed for constructive learning that enables discovery and question-exploration.

A Study on Trust Transfer in Traditional Fintech of Smart Banking (핀테크 서비스에서 오프라인에서 온라인으로의 신뢰전이에 관한 연구 - 스마트뱅킹을 중심으로 -)

  • Ai, Di;Kwon, Sun-Dong;Lee, Su-Chul;Ko, Mi-Hyun;Lee, Bo-Hyung
    • Management & Information Systems Review
    • /
    • v.36 no.3
    • /
    • pp.167-184
    • /
    • 2017
  • In this study, we investigated the effect of offline banking trust on smart banking trust. As influencing factors of smart banking trust, this study compared offline banking trust, smart banking's system quality, and information quality. For the empirical study, 186 questionnaire data were collected from smart banking users and the data were analyzed using Smart-PLS 2.0. As results, it was verified that there is trust transfer in FinTech service, by the significant effect of offline banking trust on smart banking trust. And it was proved that the effect of offline banking trust on smart banking trust is lower than that of smart banking itself. The contribution of this study can be seen in both academic and industrial aspects. First, it is the contribution of the academic aspect. Previous studies on banking were focused on either offline banking or smart banking. But this study, focus on the relationship between offline banking and online banking, proved that offline banking trust affects smart banking trust. Next, it is the industrial contribution. This study showed that offline banking characteristics of traditional commercial banks affect the trust of emerging smart banking service. This means that the emerging FinTech companies are not advantageous in the competition of trust building compared to traditional commercial banks. Unlike traditional commercial banks, the emerging FinTech is innovating the convenience of customers by arming them with new technologies such as mobile Internet, social network, cloud technology, and big data. However, these FinTech strengths alone can not guarantee sufficient trust needed for financial transactions, because banking customers do not change a habit or an inertia that they already have during using traditional banks. Therefore, emerging FinTech companies should strive to create destructive value that reflects the connection with various Internet services and the strength of online interaction such as social services, which have an advantage over customer contacts. And emerging FinTech companies should strive to build service trust, focused on young people with low resistance to new services.

  • PDF

Determinants Affecting Organizational Open Source Software Switch and the Moderating Effects of Managers' Willingness to Secure SW Competitiveness (조직의 오픈소스 소프트웨어 전환에 영향을 미치는 요인과 관리자의 SW 경쟁력 확보의지의 조절효과)

  • Sanghyun Kim;Hyunsun Park
    • Information Systems Review
    • /
    • v.21 no.4
    • /
    • pp.99-123
    • /
    • 2019
  • The software industry is a high value-added industry in the knowledge information age, and its importance is growing as it not only plays a key role in knowledge creation and utilization, but also secures global competitiveness. Among various SW available in today's business environment, Open Source Software(OSS) is rapidly expanding its activity area by not only leading software development, but also integrating with new information technology. Therefore, the purpose of this research is to empirically examine and analyze the effect of factors on the switching behavior to OSS. To accomplish the study's purpose, we suggest the research model based on "Push-Pull-Mooring" framework. This study empirically examines the two categories of antecedents for switching behavior toward OSS. The survey was conducted to employees at various firms that already switched OSS. A total of 268 responses were collected and analyzed by using the structural equational modeling. The results of this study are as follows; first, continuous maintenance cost, vender dependency, functional indifference, and SW resource inefficiency are significantly related to switch to OSS. Second, network-oriented support, testability and strategic flexibility are significantly related to switch to OSS. Finally, the results show that willingness to secures SW competitiveness has a moderating effect on the relationships between push factors and pull factor with exception of improved knowledge, and switch to OSS. The results of this study will contribute to fields related to OSS both theoretically and practically.

Development of a complex failure prediction system using Hierarchical Attention Network (Hierarchical Attention Network를 이용한 복합 장애 발생 예측 시스템 개발)

  • Park, Youngchan;An, Sangjun;Kim, Mintae;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.127-148
    • /
    • 2020
  • The data center is a physical environment facility for accommodating computer systems and related components, and is an essential foundation technology for next-generation core industries such as big data, smart factories, wearables, and smart homes. In particular, with the growth of cloud computing, the proportional expansion of the data center infrastructure is inevitable. Monitoring the health of these data center facilities is a way to maintain and manage the system and prevent failure. If a failure occurs in some elements of the facility, it may affect not only the relevant equipment but also other connected equipment, and may cause enormous damage. In particular, IT facilities are irregular due to interdependence and it is difficult to know the cause. In the previous study predicting failure in data center, failure was predicted by looking at a single server as a single state without assuming that the devices were mixed. Therefore, in this study, data center failures were classified into failures occurring inside the server (Outage A) and failures occurring outside the server (Outage B), and focused on analyzing complex failures occurring within the server. Server external failures include power, cooling, user errors, etc. Since such failures can be prevented in the early stages of data center facility construction, various solutions are being developed. On the other hand, the cause of the failure occurring in the server is difficult to determine, and adequate prevention has not yet been achieved. In particular, this is the reason why server failures do not occur singularly, cause other server failures, or receive something that causes failures from other servers. In other words, while the existing studies assumed that it was a single server that did not affect the servers and analyzed the failure, in this study, the failure occurred on the assumption that it had an effect between servers. In order to define the complex failure situation in the data center, failure history data for each equipment existing in the data center was used. There are four major failures considered in this study: Network Node Down, Server Down, Windows Activation Services Down, and Database Management System Service Down. The failures that occur for each device are sorted in chronological order, and when a failure occurs in a specific equipment, if a failure occurs in a specific equipment within 5 minutes from the time of occurrence, it is defined that the failure occurs simultaneously. After configuring the sequence for the devices that have failed at the same time, 5 devices that frequently occur simultaneously within the configured sequence were selected, and the case where the selected devices failed at the same time was confirmed through visualization. Since the server resource information collected for failure analysis is in units of time series and has flow, we used Long Short-term Memory (LSTM), a deep learning algorithm that can predict the next state through the previous state. In addition, unlike a single server, the Hierarchical Attention Network deep learning model structure was used in consideration of the fact that the level of multiple failures for each server is different. This algorithm is a method of increasing the prediction accuracy by giving weight to the server as the impact on the failure increases. The study began with defining the type of failure and selecting the analysis target. In the first experiment, the same collected data was assumed as a single server state and a multiple server state, and compared and analyzed. The second experiment improved the prediction accuracy in the case of a complex server by optimizing each server threshold. In the first experiment, which assumed each of a single server and multiple servers, in the case of a single server, it was predicted that three of the five servers did not have a failure even though the actual failure occurred. However, assuming multiple servers, all five servers were predicted to have failed. As a result of the experiment, the hypothesis that there is an effect between servers is proven. As a result of this study, it was confirmed that the prediction performance was superior when the multiple servers were assumed than when the single server was assumed. In particular, applying the Hierarchical Attention Network algorithm, assuming that the effects of each server will be different, played a role in improving the analysis effect. In addition, by applying a different threshold for each server, the prediction accuracy could be improved. This study showed that failures that are difficult to determine the cause can be predicted through historical data, and a model that can predict failures occurring in servers in data centers is presented. It is expected that the occurrence of disability can be prevented in advance using the results of this study.