• Title/Summary/Keyword: Computing ability

Search Result 382, Processing Time 0.028 seconds

Scalable Approach to Failure Analysis of High-Performance Computing Systems

  • Shawky, Doaa
    • ETRI Journal
    • /
    • v.36 no.6
    • /
    • pp.1023-1031
    • /
    • 2014
  • Failure analysis is necessary to clarify the root cause of a failure, predict the next time a failure may occur, and improve the performance and reliability of a system. However, it is not an easy task to analyze and interpret failure data, especially for complex systems. Usually, these data are represented using many attributes, and sometimes they are inconsistent and ambiguous. In this paper, we present a scalable approach for the analysis and interpretation of failure data of high-performance computing systems. The approach employs rough sets theory (RST) for this task. The application of RST to a large publicly available set of failure data highlights the main attributes responsible for the root cause of a failure. In addition, it is used to analyze other failure characteristics, such as time between failures, repair times, workload running on a failed node, and failure category. Experimental results show the scalability of the presented approach and its ability to reveal dependencies among different failure characteristics.

Exploring reward efficacy in traffic management using deep reinforcement learning in intelligent transportation system

  • Paul, Ananya;Mitra, Sulata
    • ETRI Journal
    • /
    • v.44 no.2
    • /
    • pp.194-207
    • /
    • 2022
  • In the last decade, substantial progress has been achieved in intelligent traffic control technologies to overcome consistent difficulties of traffic congestion and its adverse effect on smart cities. Edge computing is one such advanced progress facilitating real-time data transmission among vehicles and roadside units to mitigate congestion. An edge computing-based deep reinforcement learning system is demonstrated in this study that appropriately designs a multiobjective reward function for optimizing different objectives. The system seeks to overcome the challenge of evaluating actions with a simple numerical reward. The selection of reward functions has a significant impact on agents' ability to acquire the ideal behavior for managing multiple traffic signals in a large-scale road network. To ascertain effective reward functions, the agent is trained withusing the proximal policy optimization method in several deep neural network models, including the state-of-the-art transformer network. The system is verified using both hypothetical scenarios and real-world traffic maps. The comprehensive simulation outcomes demonstrate the potency of the suggested reward functions.

The Effect of Physical Computing Programming Education Integrating Artificial Intelligence on Computing Thinking Ability of Elementary School Students (인공지능을 융합한 피지컬 컴퓨팅 프로그래밍 교육이 초등학생의 컴퓨팅 사고력에 미치는 영향)

  • Yungsik Kim;Yoo-Seong Kim
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2024.01a
    • /
    • pp.5-12
    • /
    • 2024
  • 챗GPT 등 인공지능이 일상생활에서 점차 그 활용이 확대되고 있는 시기에 교육현장에서는 인공지능 융합 교육에 대한 관심이 더욱 증대되고 있다. 이에 본 논문에서는 인공지능을 융합한 피지컬 컴퓨팅 프로그래밍 교육 방법을 제안하고 초등학생의 컴퓨팅 사고력에 미치는 영향을 분석하였다. 분석 결과 인공지능을 융합한 피지컬 컴퓨팅 프로그래밍 교육이 인공지능을 융합하지 않은 피지컬 컴퓨팅 프로그래밍 교육에 비해 초등학생의 컴퓨팅 사고력 신장에 더욱 긍정적인 효과를 나타내었음을 통계적으로 확인할 수 있었다. 본 연구의 결과를 바탕으로 보다 다양한 인공지능을 융합한 소프트웨어 교육 방법과 관련된 다양한 논의와 연구가 더욱 활발히 이루어지기를 기대한다.

  • PDF

Performance Evaluation of Scheduling Algorithms according to Communication Cost in the Grid System of Co-allocation Environment (Co-allocation 환경의 그리드 시스템에서 통신비용에 따른 스케줄링 알고리즘의 성능 분석)

  • Kang, Oh-Han;Kang, Sang-Seong;Kim, Jin-Suk
    • The KIPS Transactions:PartA
    • /
    • v.14A no.2
    • /
    • pp.99-106
    • /
    • 2007
  • Grid computing, a mechanism which uses heterogeneous systems that are geographically distributed, draws attention as a new paradigm for the next generation operation of parallel and distributed computing. The importance of grid computing concerning communication cost is very huge because grid computing furnishes uses with integrated virtual computing service, in which a number of computer systems are connected by a high-speed network. Therefore, to reduce the execution time, the scheduling algorithm in grid environment should take communication cost into consideration as well as computing ability of resources. However, most scheduling algorithms have not only ignored the communication cost by assuming that all tasks were dealt in one cluster, but also did not consider the overhead of communication cost when the tasks were processed in a number of clusters. In this paper, the functions of original scheduling algorithms are analyzed. More importantly, the functions of algorithms are compared and analyzed with consideration of communication cost within the co allocation environment, in which a task is performed separately in many clusters.

EPfuzzer: Improving Hybrid Fuzzing with Hardest-to-reach Branch Prioritization

  • Wang, Yunchao;Wu, Zehui;Wei, Qiang;Wang, Qingxian
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.9
    • /
    • pp.3885-3906
    • /
    • 2020
  • Hybrid fuzzing which combines fuzzing and concolic execution, has proved its ability to achieve higher code coverage and therefore find more bugs. However, current hybrid fuzzers usually suffer from inefficiency and poor scalability when applied to complex, real-world program testing. We observed that the performance bottleneck is the inefficient cooperation between the fuzzer and concolic executor and the slow symbolic emulation. In this paper, we propose a novel solution named EPfuzzer to improve hybrid fuzzing. EPfuzzer implements two key ideas: 1) only the hardest-to-reach branch will be prioritized for concolic execution to avoid generating uninteresting inputs; and 2) only input bytes relevant to the target branch to be flipped will be symbolized to reduce the overhead of the symbolic emulation. With these optimizations, EPfuzzer can be efficiently targeted to the hardest-to-reach branch. We evaluated EPfuzzer with three sets of programs: five real-world applications and two popular benchmarks (LAVA-M and the Google Fuzzer Test Suite). The evaluation results showed that EPfuzzer was much more efficient and scalable than the state-of-the-art concolic execution engine (QSYM). EPfuzzer was able to find more bugs and achieve better code coverage. In addition, we discovered seven previously unknown security bugs in five real-world programs and reported them to the vendors.

M2M Architecture: Can It Realize Ubiquitous Computing in Daily life?

  • Babamir, Seyed Morteza
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.6 no.2
    • /
    • pp.566-579
    • /
    • 2012
  • Ubiquitous computing called pervasive one is based on the thought of pervading ability of computation in daily life applications. In other words, it aims to include computation in devices such as electronic equipment and automobiles. This has led to disengagement of computers from desktop form. Accordingly, the notice in ubiquitous computing being taken of a world steeped in remote and wireless computer-based-services. Handheld and wearable programmed devices such as sense and control appliances are such devices. This advancement is rapidly moving domestic tasks and life from device-and-human communication to the device-and-device model. This model called Machine to Machine (M2M) has led to acceleration of developments in sciences such as nano-science, bio-science, and information science. As a result, M2M led to appearance of applications in various fields such as, environment monitoring, agricultural, health care, logistics, and business. Since it is envisaged that M2M communications will play a big role in the future in all wireless applications and will be emerged as a progressive linkage for next-generation communications, this paper aims to consider how much M2M architectures can realize ubiquitous computing in daily life applications. This is carried out after acquainting and initiating readers with M2M architectures and arguments for M2M. Some of the applications was not achievable before but are becoming viable owing to emergence of M2M communications.

Cloud Computing to Improve JavaScript Processing Efficiency of Mobile Applications

  • Kim, Daewon
    • Journal of Information Processing Systems
    • /
    • v.13 no.4
    • /
    • pp.731-751
    • /
    • 2017
  • The burgeoning distribution of smartphone web applications based on various mobile environments is increasingly focusing on the performance of mobile applications implemented by JavaScript and HTML5 (Hyper Text Markup Language 5). If application software has a simple functional processing structure, then the problem is benign. However, browser loads are becoming more burdensome as the amount of JavaScript processing continues to increase. Processing time and capacity of the JavaScript in current mobile browsers are limited. As a solution, the Web Worker is designed to implement multi-threading. However, it cannot guarantee the computing ability as a native application on mobile devices, and is not sufficient to improve processing speed. The method proposed in this research overcomes the limitation of resources as a mobile client and guarantees performance by native application software by providing high computing service. It shifts the JavaScript process of a mobile device on to a cloud-based computer server. A performance evaluation experiment revealed the proposed algorithm to be up to 6 times faster in computing speed compared to the existing mobile browser's JavaScript process, and 3 to 6 times faster than Web Worker. In addition, memory usage was also less than the existing technology.

An Intelligent Residual Resource Monitoring Scheme in Cloud Computing Environments

  • Lim, JongBeom;Yu, HeonChang;Gil, Joon-Min
    • Journal of Information Processing Systems
    • /
    • v.14 no.6
    • /
    • pp.1480-1493
    • /
    • 2018
  • Recently, computational intelligence has received a lot of attention from researchers due to its potential applications to artificial intelligence. In computer science, computational intelligence refers to a machine's ability to learn how to compete various tasks, such as making observations or carrying out experiments. We adopted a computational intelligence solution to monitoring residual resources in cloud computing environments. The proposed residual resource monitoring scheme periodically monitors the cloud-based host machines, so that the post migration performance of a virtual machine is as consistent with the pre-migration performance as possible. To this end, we use a novel similarity measure to find the best target host to migrate a virtual machine to. The design of the proposed residual resource monitoring scheme helps maintain the quality of service and service level agreement during the migration. We carried out a number of experimental evaluations to demonstrate the effectiveness of the proposed residual resource monitoring scheme. Our results show that the proposed scheme intelligently measures the similarities between virtual machines in cloud computing environments without causing performance degradation, whilst preserving the quality of service and service level agreement.

A Multi-level Perception Security Model Using Virtualization

  • Lou, Rui;Jiang, Liehui;Chang, Rui;Wang, Yisen
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.11
    • /
    • pp.5588-5613
    • /
    • 2018
  • Virtualization technology has been widely applied in the area of computer security research that provides a new method for system protection. It has been a hotspot in system security research at present. Virtualization technology brings new risk as well as progress to computer operating system (OS). A multi-level perception security model using virtualization is proposed to deal with the problems of over-simplification of risk models, unreliable assumption of secure virtual machine monitor (VMM) and insufficient integration with virtualization technology in security design. Adopting the enhanced isolation mechanism of address space, the security perception units can be protected from risk environment. Based on parallel perceiving by the secure domain possessing with the same privilege level as VMM, a mechanism is established to ensure the security of VMM. In addition, a special pathway is set up to strengthen the ability of information interaction in the light of making reverse use of the method of covert channel. The evaluation results show that the proposed model is able to obtain the valuable risk information of system while ensuring the integrity of security perception units, and it can effectively identify the abnormal state of target system without significantly increasing the extra overhead.

Efficient Decentralized Sharing Economy Model Based on Blockchain Technology: A Case Study of Najm for Insurance Services Company

  • Atheer Alkhammash;Kawther Saeedi;Fatmah Baothman;Amal Babour
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.9
    • /
    • pp.29-36
    • /
    • 2023
  • Blockchain is an emerging technology that is used to address ownership, centrality, and security issues in different fields. The blockchain technology has converted centralized applications into decentralized and distributed ones. In existing sharing economy applications, there are issues related to low efficiency and high complexity of services. However, blockchain technology can be adopted to overcome these issues by effectively opening up secure information channels of the sharing economy industry and other related parties, encouraging industry integration and improving the ability of sharing economy organizations to readily gain required information. This paper discusses blockchain technology to enhance the development of insurance services by proposing a five-layer decentralized model. The Najm for Insurance Services Company in Saudi Arabia was employed in a case study for applying the proposed model to effectively solve the issue of online underwriting, and to securely and efficiently enhance the verification and validation of transactions. The paper concludes with a review of the lessons learned and provides suggestions for blockchain application development process.