• 제목/요약/키워드: computer files

검색결과 548건 처리시간 0.031초

Detecting Android Malware Based on Analyzing Abnormal Behaviors of APK File

  • Xuan, Cho Do
    • International Journal of Computer Science & Network Security
    • /
    • 제21권6호
    • /
    • pp.17-22
    • /
    • 2021
  • The attack trend on end-users via mobile devices is increasing in both the danger level and the number of attacks. Especially, mobile devices using the Android operating system are being recognized as increasingly being exploited and attacked strongly. In addition, one of the recent attack methods on the Android operating system is to take advantage of Android Package Kit (APK) files. Therefore, the problem of early detecting and warning attacks on mobile devices using the Android operating system through the APK file is very necessary today. This paper proposes to use the method of analyzing abnormal behavior of APK files and use it as a basis to conclude about signs of malware attacking the Android operating system. In order to achieve this purpose, we propose 2 main tasks: i) analyzing and extracting abnormal behavior of APK files; ii) detecting malware in APK files based on behavior analysis techniques using machine learning or deep learning algorithms. The difference between our research and other related studies is that instead of focusing on analyzing and extracting typical features of APK files, we will try to analyze and enumerate all the features of the APK file as the basis for classifying malicious APK files and clean APK files.

Hash Based Equality Analysis of Video Files with Steganography of Identifier Information

  • Lee, Wan Yeon;Choi, Yun-Seok
    • 한국컴퓨터정보학회논문지
    • /
    • 제27권7호
    • /
    • pp.17-25
    • /
    • 2022
  • 해시 함수는 입력 크기와 상관없는 고정된 작은 출력 크기로 인해서, 최근 들어 동영상 파일들의 해시 출력 값들을 빠르게 비교하는 동일성 검사에 널리 사용되고 있다. 그러나 해시 함수는 상이한 입력들이 동일한 출력 값을 도출하는 해시 충돌 가능성이 존재하여, 상이한 동영상 파일들을 동일한 파일로 오인할 수 있는 문제가 존재한다. 본 논문에서는 구별자 정보와 이중 해시를 이용하여 상이한 동영상 파일들은 항상 다르게 판단하는 동일성 검사 기법을 제안하였다. 제안된 기법은 먼저 원본 동영상 파일에서 구별자 정보를 추출하고, 이 구별자 정보를 원본 동영상에 은닉하여 추가하는 확장 파일을 생성한다. 그리고 구별자 정보, 원본 파일의 해시 출력 값, 확장 파일의 해시 출력 값을 사용하여 파일의 동일성 검사를 수행한다. 제안된 기법의 성능 평가를 위해서 실제 프로그램으로 구현하였고, 해시 충돌 없이 동영상 파일의 동일성 검사를 잘 수행하고 또한 인위적 해시 충돌 공격에도 뛰어난 내성을 가짐을 확인하였다.

Access efficiency of small sized files in Big Data using various Techniques on Hadoop Distributed File System platform

  • Alange, Neeta;Mathur, Anjali
    • International Journal of Computer Science & Network Security
    • /
    • 제21권7호
    • /
    • pp.359-364
    • /
    • 2021
  • In recent years Hadoop usage has been increasing day by day. The need of development of the technology and its specified outcomes are eagerly waiting across globe to adopt speedy access of data. Need of computers and its dependency is increasing day by day. Big data is exponentially growing as the entire world is working in online mode. Large amount of data has been produced which is very difficult to handle and process within a short time. In present situation industries are widely using the Hadoop framework to store, process and produce at the specified time with huge amount of data that has been put on the server. Processing of this huge amount of data having small files & its storage optimization is a big problem. HDFS, Sequence files, HAR, NHAR various techniques have been already proposed. In this paper we have discussed about various existing techniques which are developed for accessing and storing small files efficiently. Out of the various techniques we have specifically tried to implement the HDFS- HAR, NHAR techniques.

Computer Security에 관한 소고 - 사고범죄예방을 중심으로 - (A Study on Computer Security and Controls)

  • 이종철
    • 산업경영시스템학회지
    • /
    • 제4권4호
    • /
    • pp.25-34
    • /
    • 1981
  • Recently there has been a marked increase in concern for security in computerized operations. The purposes of computer security controls are to protect against the unauthorized access to and modification of data processing resources, unauthorised access to and modification of data files and software, and the misuse of authorized activities. The controls relate to the physical security of the data processing department and of the areas within the data processing department : to the security of the data files, programs, and system software : and to the human interaction with the data files, programs, and system software. The controls that will be discussed in this paper include : I. Risk on the computer use. II. Methods of risk counter measure. III. Role of system auditing.

  • PDF

Semantic Similarity-Based Contributable Task Identification for New Participating Developers

  • Kim, Jungil;Choi, Geunho;Lee, Eunjoo
    • Journal of information and communication convergence engineering
    • /
    • 제16권4호
    • /
    • pp.228-234
    • /
    • 2018
  • In software development, the quality of a product often depends on whether its developers can rapidly find and contribute to the proper tasks. Currently, the word data of projects to which newcomers have previously contributed are mainly utilized to find appropriate source files in an ongoing project. However, because of the vocabulary gap between software projects, the accuracy of source file identification based on information retrieval is not guaranteed. In this paper, we propose a novel source file identification method to reduce the vocabulary gap between software projects. The proposed method employs DBPedia Spotlight to identify proper source files based on semantic similarity between source files of software projects. In an experiment based on the Spring Framework project, we evaluate the accuracy of the proposed method in the identification of contributable source files. The experimental results show that the proposed approach can achieve better accuracy than the existing method based on comparison of word vocabularies.

A Pragmatic Framework for Predicting Change Prone Files Using Machine Learning Techniques with Java-based Software

  • Loveleen Kaur;Ashutosh Mishra
    • Asia pacific journal of information systems
    • /
    • 제30권3호
    • /
    • pp.457-496
    • /
    • 2020
  • This study aims to extensively analyze the performance of various Machine Learning (ML) techniques for predicting version to version change-proneness of source code Java files. 17 object-oriented metrics have been utilized in this work for predicting change-prone files using 31 ML techniques and the framework proposed has been implemented on various consecutive releases of two Java-based software projects available as plug-ins. 10-fold and inter-release validation methods have been employed to validate the models and statistical tests provide supplementary information regarding the reliability and significance of the results. The results of experiments conducted in this article indicate that the ML techniques perform differently under the different validation settings. The results also confirm the proficiency of the selected ML techniques in lieu of developing change-proneness prediction models which could aid the software engineers in the initial stages of software development for classifying change-prone Java files of a software, in turn aiding in the trend estimation of change-proneness over future versions.

하둡 프레임워크 기반 분산시스템 내의 작은 파일들을 효율적으로 처리하기 위한 방법의 설계 (The Design of Method for Efficient Processing of Small Files in the Distributed System based on Hadoop Framework)

  • 김승현;김영근;김원중
    • 한국전자통신학회논문지
    • /
    • 제10권10호
    • /
    • pp.1115-1122
    • /
    • 2015
  • 하둡 프레임워크는 매우 큰 크기의 파일을 처리하기에 적합하도록 설계되었다. 반면 작은 크기의 파일을 처리할 경우, 분산 시스템의 자원 낭비와 분석 성능 저하가 발생하며 이는 작은 파일의 개수가 많을수록 현저하게 나타난다. 이 문제는 파일의 크기가 작기 때문에 발생하므로, 연관성 있는 작은 파일들의 병합을 통해 해결할 수 있다. 그러나 기존의 작은 파일 병합 방법들은 부차적인 한계점을 지니고 있다. 따라서 본 연구는 기존의 병합 방법의 문제점에 대하여 살펴보고, 작은 파일들의 효율적 처리를 위한 병합 방법을 설계하였다.

Tailoring Operations based on Relational Algebra for XES-based Workflow Event Logs

  • Yun, Jaeyoung;Ahn, Hyun;Kim, Kwanghoon Pio
    • 인터넷정보학회논문지
    • /
    • 제20권6호
    • /
    • pp.21-28
    • /
    • 2019
  • Process mining is state-of-the-art technology in the workflow field. Recently, process mining becomes more important because of the fact that it shows the status of the actual behavior of the workflow model. However, as the process mining get focused and developed, the material of the process mining - workflow event log - also grows fast. Thus, the process mining algorithms cannot operate with some data because it is too large. To solve this problem, there should be a lightweight process mining algorithm, or the event log must be divided and processed partly. In this paper, we suggest a set of operations that control and edit XES based event logs for process mining. They are designed based on relational algebra, which is used in database management systems. We designed three operations for tailoring XES event logs. Select operation is an operation that gets specific attributes and excludes others. Thus, the output file has the same structure and contents of the original file, but each element has only the attributes user selected. Union operation makes two input XES files into one XES file. Two input files must be from the same process. As a result, the contents of the two files are integrated into one file. The final operation is a slice. It divides anXES file into several files by the number of traces. We will show the design methods and details below.

Sim-Hadoop : 신뢰성 있고 효율적인 N-body 시뮬레이션을 위한 Hadoop 분산 파일 시스템과 병렬 I / O (Sim-Hadoop : Leveraging Hadoop Distributed File System and Parallel I/O for Reliable and Efficient N-body Simulations)

  • 아마드;이승룡;정태충
    • 한국정보처리학회:학술대회논문집
    • /
    • 한국정보처리학회 2013년도 춘계학술발표대회
    • /
    • pp.476-477
    • /
    • 2013
  • Gadget-2 is a scientific simulation code has been used for many different types of simulations like, Colliding Galaxies, Cluster Formation and the popular Millennium Simulation. The code is parallelized with Message Passing Interface (MPI) and is written in C language. There is also a Java adaptation of the original code written using MPJ Express called Java Gadget. Java Gadget writes a lot of checkpoint data which may or may not use the HDF-5 file format. Since, HDF-5 is MPI-IO compliant, we can use our MPJ-IO library to perform parallel reading and writing of the checkpoint files and improve I/O performance. Additionally, to add reliability to the code execution, we propose the usage of Hadoop Distributed File System (HDFS) for writing the intermediate (checkpoint files) and final data (output files). The current code writes and reads the input, output and checkpoint files sequentially which can easily become bottleneck for large scale simulations. In this paper, we propose Sim-Hadoop, a framework to leverage HDFS and MPJ-IO for improving the I/O performance of Java Gadget code.

DATA WAREHOUSE PROTOTYPE SYSTEM TO SUPPORT THE MASTER PLAN OF URBAN REGENERATION PROJECT

  • Dong-Hyun Cho;Sang-Geum Lee;Hyung-Jin Park;Kyo-Jin Koo
    • 국제학술발표논문집
    • /
    • The 3th International Conference on Construction Engineering and Project Management
    • /
    • pp.361-368
    • /
    • 2009
  • In the Urban Regeneration Project, the scale becomes huge and the composition of participating organizations is complex. It is necessary to establish the master plan fast and effectively harmonize conflict options and interests in the early stage of the project. However, the master planner, who is tasked to establish the master plan, and the master planning team, are difficult to establish the master plan fast while effectively considering various and complex information related to regeneration. The Urban Regeneration Project is composed of various types of computer files like HWP, EXL, PDF, DWG and others. When the master plan is established, it is necessary to link the information items and physical files of urban generation in order to effectively save the files in storage for easy search and utilization. This study developed a prototype system to support participating subjects of the Urban Regeneration Projects. It will effectively utilize the information items in the physical computer files using information blocks and Metadata. The developed system provides the base to effectively manage large amount of documents in the early stage of the Urban Regeneration Project and make effective decision-making using the concept of data warehouse.

  • PDF