• Title/Summary/Keyword: CPU 시간

Search Result 518, Processing Time 0.03 seconds

Detecting Repackaged Applications using the Information of App Installation in Android Smartphones (안드로이드 스마트폰에서 앱 설치 정보를 이용한 리패키징 앱 탐지 기법)

  • Joun, Young Nam;Ahn, Woo Hyun
    • Convergence Security Journal
    • /
    • v.12 no.4
    • /
    • pp.9-15
    • /
    • 2012
  • In recently years, repackaged malwares are becoming increased rapidly in Android smartphones. The repackaging is a technique to disassemble an app in a market, modify its source code, and then re-assemble the code, so that it is commonly used to make malwares by inserting malicious code in an app. However, it is impossible to collect all the apps in many android markets including too many apps. To solve the problem, we propose RePAD (RePackaged App Detector) scheme that is composed of a client and a remote server. In the smartphone-side, the client extracts the information of an app with low CPU overhead when a user installs the app. The remote server analyzes the information to decide whether the app is repackaged or not. Thus, the scheme reduces the time and cost to decide whether apps are repackaged. For the experiments, the client and server are implemented as an app on Galaxy TAB and PC respectively. We indicated that seven pairs of apps among ones collected in official and unofficial market are repackaged. Furthermore, RePAD only increases the average of CPU overhead of 1.9% and the maximum memory usage of 3.5 MB in Galaxy TAB.

A group based management method of flash memory for enhancing wear-leveling (Wear-leveling 향상을 위한 플래시 메모리의 그룹단위 관리 방법)

  • Jang, Si-Woong;Kim, Young-Ju;Yu, Yun-Sik
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.13 no.2
    • /
    • pp.315-320
    • /
    • 2009
  • Since flash memory can not be overwritten on updating data, new data are updated in new area and old data should be invalidated and erased for garbage collection. With develop of flash memory technology, capacity of flash memory is rapidly increasing. It increases rapidly execution time of CPU to search an entire flash memory of large capacity when choosing the block to erase in garbage collection. To solve the problem that is increasing execution time of CPU, flash memory is partitioned into several groups, the block to erase in garbage collection is searched within the corresponding group. In workload of access locality, we enhanced wear-leveling within group by allocating hot data to hot group and cold data to cold group respectively and enhanced wear-leveling among groups by exchanging periodically hot group and cold group.

Development of a High Performance Web Server Using A Real-Time Compression Architecture (실시간 압축 전송 아키텍쳐를 이용한 고성능 웹서버 구현)

  • Min Byungjo;Hwang June;Kim Hagbae
    • The KIPS Transactions:PartC
    • /
    • v.11C no.6 s.95
    • /
    • pp.781-786
    • /
    • 2004
  • In these days, such services are popularized as E-commerce, E- government, multimedia services, and home networking applications. Most web traffics generated contemporarily basically use the Hyper Text Transfer Protocol(HTTP). Unfortunately, the HTTP is improper for these applications that comprise significant components of the web traffics. In this paper, we introduce a real-time contents compression architecture that maximizes the web service performance as well as reduces the response time. This architecture is built into the linux kernel-based web accelerating module. It guarantees not only the freshness of compressed contents but also the minimum time delay using an server-state adaptive algorithm, which can determine whether the server sends the compressed message considering the consumption of sewer resources when heavy requests reach the web server. Also, We minimize the CPU overhead of the web server by exclusively implementing the compression kernel-thread. The testing results validates that this architecture saves the bandwidth of the web server and that elapsed time improvement is dramatic.

comparison of Numercal Methods for Obtaining 2-D Impurity Profile in Semiconductor (반도체 내에서의 2차원 불순물 분포를 얻기 위한 수치해법의 비교)

  • Yang, Yeong-Il;Gyeong, Jong-Min;O, Hyeong-Cheol
    • Journal of the Korean Institute of Telematics and Electronics
    • /
    • v.22 no.3
    • /
    • pp.95-102
    • /
    • 1985
  • An efficient numerical scheme for assessing the two-dimensional diffusion problem for modelling impurity profile in semiconductor is described. 4 unique combination of ADI (Al-ternating Direction Bmplicit) method and Gauss Elimination has resulted in a reduction of CPU time for most diffusion processes by a factor of 3, compared to other iteration schemes such as SOR (Successive Over-Relaxation) or Stone's iterative method without additional storage re-quirement. Various numerical schemes were compared for 2-D as well as 1-0 diffusion profile in terms of their CPU time while retaining the magnitude of relative error within 0.001%. good agree-ment between 1-D and 2-D simulation profile as well as between 1-D simulation profile and experiment has been obtained.

  • PDF

A VIA-based RDMA Mechanism for High Performance PC Cluster Systems (고성능 PC 클러스터 시스템을 위한 VIA 기반 RDMA 메커니즘 구현)

  • Jung In-Hyung;Chung Sang-Hwa;Park Sejin
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.31 no.11
    • /
    • pp.635-642
    • /
    • 2004
  • The traditional communication protocols such as TCP/IP are not suitable for PC cluster systems because of their high software processing overhead. To eliminate this overhead, industry leaders have defined the Virtual Interface Architecture (VIA). VIA provides two different data transfer mechanisms, a traditional Send/Receive model and the Remote Direct Memory Access (RDMA) model. RDMA is extremely efficient way to reduce software overhead because it can bypass the OS and use the network interface controller (NIC) directly for communication, also bypass the CPU on the remote host. In this paper, we have implemented VIA-based RDMA mechanism in hardware. Compared to the traditional Send/Receive model, the RDMA mechanism improves latency and bandwidth. Our RDMA mechanism can also communicate without using remote CPU cycles. Our experimental results show a minimum latency of 12.5${\mu}\textrm{s}$ and a maximum bandwidth of 95.5MB/s. As a result, our RDMA mechanism allows PC cluster systems to have a high performance communication method.

Optimal Operation of the 3D Water Quality Model for Water Quality Forecast (수질예보를 위한 3차원 모형의 최적 운영 기법)

  • Lee, Seungjae;Kim, Hyeonsik;Sa, Sungoh;Hwang, Hyunsik
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2016.05a
    • /
    • pp.72-72
    • /
    • 2016
  • 최근 발생하고 있는 기후변화로 인하여 하천 및 저수지의 수질문제가 커지고 있다. 특히 여름철 부영양화로 인해 발생하는 녹조현상은 사회적인 문제로 과학적인 수질사고에 대한 예측과 관리가 필요한 실정이다. 수질예보는 정기적으로 하천 및 저수지의 수질을 예측하여 사용자에게 제공하는 분석기법으로 수질현황을 파악하고 수질을 관리하고 의사결정을 하는데 도움을 줄 수 있다. 수질예보에 사용되는 모형은 유역모형, 하천모형, 저수지모형이 있으며, 이중 하천 및 저수지에 주로 적용되고 있는 3차원 수리수질모형의 경우 격자의 개수가 많아 모의시간이 길어지게 되고 이로 인해 일일 수질 예보가 어렵게 된다. 3차원 수리수질모형의 모의속도를 개선하는 방법에는 하드웨어의 성능을 높이는 방법과 병렬화를 이용한 소프트웨어적인 방법이 있다. 이중 하드웨어의 성능을 높이는 방법은 컴퓨터의 사양을 높이는 방법으로 높은 비용이 소요된다. 하지만 병렬화 방법은 컴퓨팅 기술의 발전으로 멀티코어가 대중화가 된 최근에 코드의 적용만으로 모의속도를 향상시킬 수 있다. 본 연구에서 사용된 모형은 서호주대학에서 개발한 3차원 수리 수질모형인 ELCOM-CAEDYM 모형으로 적용된 병렬화 기법은 OpenMP(Open Multi-Processing)방법이다. 기존 직렬 컴퓨팅 방식으로 구성되어 한번에 한 개의 명령어 밖에 처리할 수 없었던 작업방법을 동시에 여러 개의 처리요소를 이용하여 명령을 실행할 수 있게 하는 방식이다. 하지만 CPU의 개수는 제한되어 있으며, Amdahl's law에 따르면 OpenMP방식의 병렬화시 속도개선효과는 95% 병렬화 프로그램에서 최대 CPU 개수의 제한이 없다면 20배 까지 속도향상이 가능하다고 하였다. 본 연구에서는 3차원 수리 수질예측 모형인 ELCOM-CAEDYM에 적용된 병렬화 기법을 적용하는데 있어 최적 CPU사용 개수를 파악 하고자 하였으며, 이를 통해 수질예보시스템을 운영하는데 가장 효율적인 방법을 찾아 적용하고자 하고자 한다.

  • PDF

A study of extended processor trace decoder structure for malicious code detection (악성코드 검출을 위한 확장된 프로세서 트레이스 디코더 구조 연구)

  • Kang, Seungae;Kim, Youngsoo;Kim, Jonghyun;Kim, Hyuncheol
    • Convergence Security Journal
    • /
    • v.18 no.5_1
    • /
    • pp.19-24
    • /
    • 2018
  • For a long time now, general-purpose processors have provided dedicated hardware / software tracing modules to provide developers with tools to fix bugs. A hardware tracer generates its enormous data into a log that is used for both performance analysis and debugging. Processor Trace (PT) is a new hardware-based tracing feature for Intel CPUs that traces branches executing on the CPU, which allows the reconstruction of the control flow of all executed code with minimal labor. Hardware tracer has been integrated into the operating system, which allows tight integration with its profiling and debugging mechanisms. However, in the Windows environment, existing studies related to PT focused on decoding only one flow in sequence. In this paper, we propose an extended PT decoder structure that provides basic data for real-time trace and malicious code detection using the functions provided by PT in Windows environment.

  • PDF

A Study on Efficient User Management System of Combat System

  • Hee-Soo Kim
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.7
    • /
    • pp.191-198
    • /
    • 2024
  • In this paper, we proposes a user management system for efficient operation of the combat system within naval ship. Recently, naval ships have seen performance enhancements through various sensors, features, and continuous system development. This progress in the system has led to an increase in multi-funstion consoles that can manipulate various sensors and features within naval ship, consequently increasing the number of operators for these consoles. Therefore, a user management system that can control and manage multi-function consoles and operators in real-time is necessary for efficient management within naval ship. This paper suggests a user management system that can effectively manage the real-time situation of users accessing multi-function consoles. Additionally, a parallelization method using GPUs to reduce the CPU workload in operating various functions of the combat system is proposed. The proposed user management system has shown a performance improvement where the response time decreased by approximately 82% and the occupancy reduced by approximately 20% compared to the method using CPUs.

The Performance Bottleneck of Subsequence Matching in Time-Series Databases: Observation, Solution, and Performance Evaluation (시계열 데이타베이스에서 서브시퀀스 매칭의 성능 병목 : 관찰, 해결 방안, 성능 평가)

  • 김상욱
    • Journal of KIISE:Databases
    • /
    • v.30 no.4
    • /
    • pp.381-396
    • /
    • 2003
  • Subsequence matching is an operation that finds subsequences whose changing patterns are similar to a given query sequence from time-series databases. This paper points out the performance bottleneck in subsequence matching, and then proposes an effective method that improves the performance of entire subsequence matching significantly by resolving the performance bottleneck. First, we analyze the disk access and CPU processing times required during the index searching and post processing steps through preliminary experiments. Based on their results, we show that the post processing step is the main performance bottleneck in subsequence matching, and them claim that its optimization is a crucial issue overlooked in previous approaches. In order to resolve the performance bottleneck, we propose a simple but quite effective method that processes the post processing step in the optimal way. By rearranging the order of candidate subsequences to be compared with a query sequence, our method completely eliminates the redundancy of disk accesses and CPU processing occurred in the post processing step. We formally prove that our method is optimal and also does not incur any false dismissal. We show the effectiveness of our method by extensive experiments. The results show that our method achieves significant speed-up in the post processing step 3.91 to 9.42 times when using a data set of real-world stock sequences and 4.97 to 5.61 times when using data sets of a large volume of synthetic sequences. Also, the results show that our method reduces the weight of the post processing step in entire subsequence matching from about 90% to less than 70%. This implies that our method successfully resolves th performance bottleneck in subsequence matching. As a result, our method provides excellent performance in entire subsequence matching. The experimental results reveal that it is 3.05 to 5.60 times faster when using a data set of real-world stock sequences and 3.68 to 4.21 times faster when using data sets of a large volume of synthetic sequences compared with the previous one.

Construction of a Sub-catchment Connected Nakdong-gang Flood Analysis System Using Distributed Model (분포형 모형을 이용한 소유역 연계 낙동강 홍수해석시스템 구축)

  • Choi, Yun-Seok;Won, Young-Jin;Kim, Kyung-Tak
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2018.05a
    • /
    • pp.202-202
    • /
    • 2018
  • 본 논문에서는 분포형 강우-유출 모형인 GRM(Grid based Rainfall-runoff Model)(최윤석, 김경탁, 2017)을 이용해서 낙동강 유역을 대상으로 대유역 홍수해석시스템을 구축하고, 유출해석을 위한 실행시간을 평가하였다. 유출모형은 낙동강의 주요 지류와 본류를 소유역으로 구분하여 모형을 구축하고, 각 소유역의 유출해석 결과를 실시간으로 연계할 수 있도록 하여 낙동강 전체 유역의 유출모형을 구축하였다. 이와 같이 하나의 대유역을 다수의 소유역시스템으로 분할하여 모형을 구축할 경우, 유출해석시스템 구성이 복잡해지는 단점이 있으나, 소유역별로 각기 다른 자료를 이용하여 다양한 해상도로 유출해석을 할 수 있으므로, 소유역별 특성에 맞는 유출모형 구축이 가능한 장점이 있다. 또한 각 소유역시스템은 별도의 프로세스로 계산이 진행되므로, 대유역을 고해상도로 해석하는 경우에도 계산시간을 단축할 수 있다. 본 연구에서는 낙동강 유역을 20개(본류 구간 3개, 1차 지류 13개, 댐상류 4개)의 소유역으로 분할하여 계산 시간을 검토하였으며, 최종적으로 21개(본류 구간 3개, 1차 지류 13개, 댐상류 5개)의 소유역으로 분할하여 유출해석시스템을 구축하였다. 댐 상류 유역은 댐하류와 유량전달이 없이 독립적으로 모의되고, 댐과 연결된 하류 유역은 관측 방류량을 상류단 하천의 경계조건으로 적용한다. 지류 유역은 본류 구간과 연결되고, 지류의 계산 유량은 본류와의 연결지점에 유량조건으로 실시간으로 입력된다. 이때 본류와 지류의 유량 연계는 데이터베이스를 매개로 하였다. 유출해석시스템의 성능을 평가하기 위해서 Microsoft 클라우드 서비스인 Azure를 이용하였다. 낙동강 유역을 20개 소유역으로 구성한 경우에서의 유출해석시스템의 속도 평가 결과 Azure virtual machine instance DS15 v2(OS : Windows Server 2012 R2, CPU : 2.4 GHz Intel $Xeon^{(R)}$ E5-2673 v3 20 cores)에서 1.5분이 소요 되었다. 계산시간 평가시 GRM은 'IsParallel=false' 옵션을 적용하였으며, 모의 기간은 24시간을 기준으로 하였다. 연구결과 분포형 모형을 이용한 대유역 유출해석시스템 구축이 가능했으며, 계산시간도 충분히 단축할 수 있었다. 또한 추가적인 CPU와 병렬계산을 적용할 경우, 계산시간은 더 단축될 수 있으며, 이러한 기법들은 분포형 모형을 이용한 대유역 유출해석시스템 구축시 유용하게 활용될 수 있을 것으로 판단된다.

  • PDF