• Title/Summary/Keyword: computer files

Search Result 548, Processing Time 0.032 seconds

SplitScreen: Enabling Efficient, Distributed Malware Detection

  • Cha, Sang-Kil;Moraru, Iulian;Jang, Ji-Yong;Truelove, John;Brumley, David;Andersen, David G.
    • Journal of Communications and Networks
    • /
    • v.13 no.2
    • /
    • pp.187-200
    • /
    • 2011
  • We present the design and implementation of a novel anti-malware system called SplitScreen. SplitScreen performs an additional screening step prior to the signature matching phase found in existing approaches. The screening step filters out most non-infected files (90%) and also identifiesmalware signatures that are not of interest (99%). The screening step significantly improves end-to-end performance because safe files are quickly identified and are not processed further, and malware files can subsequently be scanned using only the signatures that are necessary. Our approach naturally leads to a network-based anti-malware solution in which clients only receive signatures they needed, not every malware signature ever created as with current approaches. We have implemented SplitScreen as an extension to ClamAV, the most popular open source anti-malware software. For the current number of signatures, our implementation is $2{\times}$ faster and requires $2{\times}$ less memory than the original ClamAV. These gaps widen as the number of signatures grows.

NVST DATA ARCHIVING SYSTEM BASED ON FASTBIT NOSQL DATABASE

  • Liu, Ying-Bo;Wang, Feng;Ji, Kai-Fan;Deng, Hui;Dai, Wei;Liang, Bo
    • Journal of The Korean Astronomical Society
    • /
    • v.47 no.3
    • /
    • pp.115-122
    • /
    • 2014
  • The New Vacuum Solar Telescope (NVST) is a 1-meter vacuum solar telescope that aims to observe the fine structures of active regions on the Sun. The main tasks of the NVST are high resolution imaging and spectral observations, including the measurements of the solar magnetic field. The NVST has been collecting more than 20 million FITS files since it began routine observations in 2012 and produces maximum observational records of 120 thousand files in a day. Given the large amount of files, the effective archiving and retrieval of files becomes a critical and urgent problem. In this study, we implement a new data archiving system for the NVST based on the Fastbit Not Only Structured Query Language (NoSQL) database. Comparing to the relational database (i.e., MySQL; My Structured Query Language), the Fastbit database manifests distinctive advantages on indexing and querying performance. In a large scale database of 40 million records, the multi-field combined query response time of Fastbit database is about 15 times faster and fully meets the requirements of the NVST. Our slestudy brings a new idea for massive astronomical data archiving and would contribute to the design of data management systems for other astronomical telescopes.

A Rapid Locating Protocol of Corrupted Data for Cloud Data Storage

  • Xu, Guangwei;Yang, Yanbin;Yan, Cairong;Gan, Yanglan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.10
    • /
    • pp.4703-4723
    • /
    • 2016
  • The verification of data integrity is an urgent topic in remote data storage environments with the wide deployment of cloud data storage services. Many traditional verification algorithms focus on the block-oriented verification to resolve the dispute of dynamic data integrity between the data owners and the storage service providers. However, these algorithms scarcely pay attention to the data verification charge and the users' verification experience. The users more concern about the availability of accessed files rather than data blocks. Moreover, the data verification charge limits the number of checked data in each verification. Therefore, we propose a mixed verification protocol to verify the data integrity, which rapidly locates the corrupted files by the file-oriented verification, and then identifies the corrupted blocks in these files by the block-oriented verification. Theoretical analysis and simulation results demonstrate that the protocol reduces the cost of the metadata computation and transmission relative to the traditional block-oriented verification at the expense of little cost of additional file-oriented metadata computation and storage at the data owner. Both the opportunity of data extracted and the scope of suspicious data are optimized to improve the verification efficiency under the same verification cost.

Distributed database replicator without locking base relations

  • Lee, Wookey;Kang, Sukho;Park, Jooseok
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 1996.10a
    • /
    • pp.93-95
    • /
    • 1996
  • A replication server is considered to be one of the most effective tools to cope with the problems that may be caused by the complex data replications in distributed database systems. In the distributed environment, locking a table is inevitable and it is the main reason to coerce the system practically. This paper presents an Asynchronous Replicator Scheme (ARS) that basically utilizes the system log as files named differential files to refresh the distributed data files with complicated queries, and that it prevents (normally, huge) base tables from being locked. We take join operations as the complicated queries, not only because the join operation covers almost all the operations, but also because it is one of the most time-consuming and data intensive operations in query processings.

  • PDF

A Study on Data Allocation Problems of Distributed Databases (분산 데이타 베이스 설계시의 자료 배정문제에 관한 연구)

  • Sin, Gi-Tae;Park, Jin-Woo
    • Asia pacific journal of information systems
    • /
    • v.1 no.1
    • /
    • pp.49-62
    • /
    • 1991
  • This paper examines the problems of database partitioning and file allocation in a fixed topology distributed computer network. The design objective is to make files as collections of attributes and to allocate these files to network nodes so that a minimum total transmission cost is achieved subject to storage capacity constraints. A mathematical model for solving the problem is formulated and, the resulting optimization problem is shown to fall in a class of NP-complete problems. A new heuristic algorithm is developed which uses the idea of allocating attributes according to the transaction requirements at each computer node and then making files using the allocated attributes. Numerical results indicate that the heuristic algorithm yields practicable low cost solutions in comparison with the existing methods which deal with the file allocation problems and database partitioning problems independently.

  • PDF

Design and Implementation of Damaged Video File Recovery Tool using Container Format Structure

  • Choi, Yun-Seok;Lee, Wan Yeon
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.11 no.3
    • /
    • pp.64-70
    • /
    • 2019
  • Video files of video devices such as black box and CCTV may be damaged due to repetitive file read / write and physical environment factors. Even though there are available parts of video information, it may happen that playback can't be performed due to damage of some information. To playback the remaining video information normally, it is necessary to recover damaged areas of the files. For this, it is necessary to accurately check the damage range of the files. In this paper, we propose the design and implementation of a tool which detects damaged areas of a video file and recovers the usable area of the file to playback. The proposed tool can analyze and recover without additional information by analyzing common information of video container format and can check detailed damaged ranges with chunks. It is possible to perform recovery just only with the target file and reference file without any other information such as codec specification.

Design and Implementation of a ML-based Detection System for Malicious Script Hidden Corrupted Digital Files (머신러닝 기반 손상된 디지털 파일 내부 은닉 악성 스크립트 판별 시스템 설계 및 구현)

  • Hyung-Woo Lee;Sangwon Na
    • Journal of Internet of Things and Convergence
    • /
    • v.9 no.6
    • /
    • pp.1-9
    • /
    • 2023
  • Malware files containing concealed malicious scripts have recently been identified within MS Office documents frequently. In response, this paper describes the design and implementation of a system that automatically detects malicious digital files using machine learning techniques. The system is proficient in identifying malicious scripts within MS Office files that exploit the OLE VBA macro functionality, detecting malicious scripts embedded within the CDH/LFH/ECDR internal field values through OOXML structure analysis, and recognizing abnormal CDH/LFH information introduced within the OOXML structure, which is not conventionally referenced. Furthermore, this paper presents a mechanism for utilizing the VirusTotal malicious script detection feature to autonomously determine instances of malicious tampering within MS Office files. This leads to the design and implementation of a machine learning-based integrated software. Experimental results confirm the software's capacity to autonomously assess MS Office file's integrity and provide enhanced detection performance for arbitrary MS Office files when employing the optimal machine learning model.

A Study on the Improving Performance of Massively Small File Using the Reuse JVM in MapReduce (MapReduce에서 Reuse JVM을 이용한 대규모 스몰파일 처리성능 향상 방법에 관한 연구)

  • Choi, Chul Woong;Kim, Jeong In;Kim, Pan Koo
    • Journal of Korea Multimedia Society
    • /
    • v.18 no.9
    • /
    • pp.1098-1104
    • /
    • 2015
  • With the widespread use of smartphones and IoT (Internet of Things), data are being generated on a large scale, and there is increased for the analysis of such data. Hence, distributed processing systems have gained much attention. Hadoop, which is a distributed processing system, saves the metadata of stored files in name nodes; in this case, the main problems are as follows: the memory becomes insufficient; load occurs because of massive small files; scheduling and file processing time increases because of the increased number of small files. In this paper, we propose a solution to address the increase in processing time because of massive small files, and thus improve the processing performance, using the Reuse JVM method provided by Hadoop. Through environment setting, the Reuse JVM method modifies the JVM produced conventionally for every task, so that multiple tasks are reused sequentially in one JVM. As a final outcome, the Reuse JVM method showed the best processing performance when used together with CombineFileInputFormat.

Isonumber based Iso-Key Interchange Protocol for Network Communication

  • Dani, Mamta S.;Meshram, Akshaykumar;Pohane, Rupesh;Meshram, Rupali R.
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.2
    • /
    • pp.209-213
    • /
    • 2022
  • Key exchange protocol (KEP) is an essential setup to secure authenticates transmission among two or more users in cyberspace. Digital files protected and transmitted by the encryption of the files over public channels, a single key communal concerning the channel parties and utilized for both to encrypt the files as well as decrypt the files. If entirely done, this impedes unauthorized third parties from imposing a key optimal on the authorized parties. In this article, we have suggested a new KEP term as isokey interchange protocol based on generalization of modern mathematics term as isomathematics by utilizing isonumbers for corresponding isounits over the Block Upper Triangular Isomatrices (BUTI) which is secure, feasible and extensible. We also were utilizing arithmetic operations like Isoaddition, isosubtraction, isomultiplication and isodivision from isomathematics to build iso-key interchange protocol for network communication. The execution of our protocol is for two isointegers corresponding two elements of the group of isomatrices and cryptographic performance of products eachother. We demonstrate the protection of suggested isokey interchange protocol against Brute force attacks, Menezes et al. algorithm and Climent et al. algorithm.

Implementation and Verification of Multi-level Convolutional Neural Network Algorithm for Identifying Unauthorized Image Files in the Military (국방분야 비인가 이미지 파일 탐지를 위한 다중 레벨 컨볼루션 신경망 알고리즘의 구현 및 검증)

  • Kim, Youngsoo
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.8
    • /
    • pp.858-863
    • /
    • 2018
  • In this paper, we propose and implement a multi-level convolutional neural network (CNN) algorithm to identify the sexually explicit and lewdness of various image files, and verify its effectiveness by using unauthorized image files generated in the actual military. The proposed algorithm increases the accuracy by applying the convolutional artificial neural network step by step to minimize classification error between similar categories. Experimental data have categorized 20,005 images in the real field into 6 authorization categories and 11 non-authorization categories. Experimental results show that the overall detection rate is 99.51% for the image files. In particular, the excellence of the proposed algorithm is verified through reducing the identification error rate between similar categories by 64.87% compared with the general CNN algorithm.