• Title/Summary/Keyword: data recover

Search Result 495, Processing Time 0.025 seconds

Doubly-Selective Channel Estimation for OFDM Systems Using a Pilot-Embedded Training Scheme

  • Wang, Li-Dong;Lim, Dong-Min
    • Journal of electromagnetic engineering and science
    • /
    • v.6 no.4
    • /
    • pp.203-208
    • /
    • 2006
  • Channel estimation and data detection for OFDM systems over time- and frequency-selective channels are investigated. Relying on the complex exponential basis expansion channel model, a pilot-embedded channel estimation scheme with low computational complexity and spectral efficiency is proposed. A periodic pilot sequence is superimposed at a low power on information bearing sequence at the transmitter before modulation and transmission. The channel state information(CSI) can be estimated using the first-order statistics of the received data. In order to enhance the performance of channel estimation, we recover the transmitted data which can be exploited to estimate CSI iteratively. Simulation results show that the proposed method is suitable for doubly-selective channel estimation for the OFDM systems and the performance of the proposed method can be better than that of the Wiener filter method under some conditions. Through simulations, we also analyze the factors which can affect the system performances.

File Signature's Automatic Calculation Algorithm Proposal for Digital Forensic

  • Jang, Eun-Jin;Shin, Seung-Jung
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.13 no.3
    • /
    • pp.118-123
    • /
    • 2021
  • Recently, digital crime is becoming more intelligent, and efficient digital forensic techniques are required to collect evidence for this. In the case of important files related to crime, a specific person may intentionally delete the file. In such a situation, data recovery is a very important procedure that can prove criminal charges. Although there are various methods to recover deleted files, we focuses on the recovery technique using HxD editor. When recovering a deleted file using the HxD editor, check the file structure and access the file data area through calculation. However, there is a possibility that errors such as arithmetic errors may occur when a file approach through calculation is used. Therefore, in this paper, we propose an algorithm that automatically calculates the header and footer of a file after checking the file signature in the root directory for efficient file recovery. If the algorithm proposed in this paper is used, it is expected that the error rate of arithmetic errors in the file recovery process can be reduced.

The Method of Recovery for Deleted Record of Realm Database (Realm 데이터베이스의 삭제된 레코드 복구 기법)

  • Kim, Junki;Han, Jaehyeok;Choi, Jong-Hyun;Lee, Sangjin
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.28 no.3
    • /
    • pp.625-633
    • /
    • 2018
  • Realm is an open source database developed to replace SQLite, which is commonly used in mobile devices. The data stored in the database must be checked during the digital forensic analysis process for mobile devices because it can help to understand the behavior of the user and whether the mobile device is operating or not. In addition, since the user can intentionally use anti-forensic techniques such as deleting data stored in the database, research on how to recover deleted records is needed. In this paper, we propose a method to recover records that have not been overwritten after deletion based on the analysis of the structure and record and deletion process of the Realm database file.

A File Recovery Technique for Digital Forensics on NAND Flash Memory (NAND 플래시 메모리에서 디지털 포렌식을 위한 파일 복구기법)

  • Shin, Myung-Sub;Park, Dong-Joo
    • Journal of KIISE:Databases
    • /
    • v.37 no.6
    • /
    • pp.292-299
    • /
    • 2010
  • Recently, as flash memory is used as digital storage devices, necessity for digital forensics is growing in a flash memory area for digital evidence analysis. For this purpose, it is important to recover crashed files stored on flash memory efficiently. However, it is inefficient to apply the hard disk based file recovery techniques to flash memory, since hard disk and flash memory have different characteristics, especially flash memory being unable to in-place update. In this paper, we propose a flash-aware file recovery technique for digital forensics. First, we propose an efficient search technique to find all crashed files. This uses meta-data maintained by FTL(Flash Translation Layer) which is responsible for write operation in flash memory. Second, we advise an efficient recovery technique to recover a crashed file which uses data location information of the mapping table in FTL. Through diverse experiments, we show that our file recovery technique outperforms the hard disk based technique.

The Research on the Recovery Techniques of Deleted Files in the XFS Filesystem (XFS 파일 시스템 내의 삭제된 파일 복구 기법 연구)

  • Ahn, Jae-Hyoung;Park, Jung-Heum;Lee, Sang-Jin
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.24 no.5
    • /
    • pp.885-896
    • /
    • 2014
  • The files in computer storages can be deleted due to unexpected failures or accidents. Some malicious users often delete data by himself for anti-forensics. If deleted files are associated with crimes or important documents in business, they should be recovered and the recovery tool is necessary. The recovery methods and tools for some filesystems such as NTFS, FAT, and EXT have been developed actively. However, there has not been any researches for recovering deleted files in XFS filesystem applied to NAS or CCTV. In addition, since the current related tools are based on the traditional signature detection methods, they have low recovery rates. Therefore, this paper suggests the recovery methods for deleted files based on metadata and signature detection in XFS filesystem, and verifies the results by conducting experiment in real environment.

Sanitization of Open-Source Based Deduplicated Filesystem (오픈 소스 중복 제거 파일시스템에서의 완전 삭제)

  • Cho, Hyeonwoong;Kim, SeulGi;Kwon, Taekyoung
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.26 no.5
    • /
    • pp.1141-1149
    • /
    • 2016
  • Deduplicated filesystem can reduce usage of storage. However, it be able to recover deleted block. We studied sanitization of deduplicated filesystem, LessFS which is based on FUSE(Filesystem in USErspace). First, we show a vulnerability recover deleted data in the deduplicated filesystem. We implement sanitization of deduplicated filesystem considering the part of fingerprint DB with data blocks. It takes 60~70 times compared to without sanitization. Which means access time to fingerprint DB and overhead derived from increase of number of chunk have a critical impact on sanitization time. But in case of more than 65,536 Byte of chunksize, it is faster than normal filesystem without deduplication.

Dentifying and Clustering the Flood Impacted Areas for Strategic Information Provision (전략적 정보제공을 위한 침수영향구역 클러스터링)

  • Park, Eun Mi;Bilal, Muhammad
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.20 no.6
    • /
    • pp.100-109
    • /
    • 2021
  • Flooding usually brings in disruptions and aggravated congestions to the roadway network. Hence, right information should be provided to road users to avoid the flood-impacted areas and for city officials to recover the network. However, the information about individual link congestion may not be conveyed to roadway users and city officials because too many links are congested at the same time. Therefore, more significant information may be desired, especially in a disastrous situation. This information may include 1) which places to avoid during flooding 2) which places are feasible to drive avoiding flooding. Hence, this paper aims to develop a framework to identify the flood-impacted areas in a roadway network and their criticality. Various impacted clusters and their spatiotemporal properties were identified with field data. From this data, roadway users can reroute their trips, and city officials can take the right actions to recover the affected areas. The information resulting from the developed framework would be significant enough for roadway users and city officials to cope with flooding.

A Fast and Secure Scheme for Data Outsourcing in the Cloud

  • Liu, Yanjun;Wu, Hsiao-Ling;Chang, Chin-Chen
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.8
    • /
    • pp.2708-2721
    • /
    • 2014
  • Data outsourcing in the cloud (DOC) is a promising solution for data management at the present time, but it could result in the disclosure of outsourced data to unauthorized users. Therefore, protecting the confidentiality of such data has become a very challenging issue. The conventional way to achieve data confidentiality is to encrypt the data via asymmetric or symmetric encryptions before outsourcing. However, this is computationally inefficient because encryption/decryption operations are time-consuming. In recent years, a few DOC schemes based on secret sharing have emerged due to their low computational complexity. However, Dautrich and Ravishankar pointed out that most of them are insecure against certain kinds of collusion attacks. In this paper, we proposed a novel DOC scheme based on Shamir's secret sharing to overcome the security issues of these schemes. Our scheme can allow an authorized data user to recover all data files in a specified subset at once rather than one file at a time as required by other schemes that are based on secret sharing. Our thorough analyses showed that our proposed scheme is secure and that its performance is satisfactory.

Design of clock/data recovery circuit for optical communication receiver (광통신 수신기용 클럭/데이타 복구회로 설계)

  • Lee, Jung-Bong;Kim, Sung-Hwan;Choi, Pyung
    • Journal of the Korean Institute of Telematics and Electronics A
    • /
    • v.33A no.11
    • /
    • pp.1-9
    • /
    • 1996
  • In the following paper, new architectural algorithm of clock and data recovery circuit is proposed for 622.08 Mbps optical communication receiver. New algorithm makes use of charge pump PLL using voltage controlled ring oscillator and extracts 8-channel 77.76 MHz clock signals, which are delayed by i/8 (i=1,2, ...8), to convert and recover 8-channel parallel data from 662.08 Mbps MRZ serial data. This circuit includes clock genration block to produce clock signals continuously even if input data doesn't exist. And synchronization of data and clock is doen by the method which compares 1/2 bit delayed onput data and decided dta by extracted clock signals. Thus, we can stabilize frequency and phase of clock signal even if input data is distorted or doesn't exist and simplify receiver architecture compared to traditional receiver's. Also it is possible ot realize clock extraction, data decision and conversion simulataneously. Verification of this algorithm is executed by DESIGN CENTER (version 6.1) using test models which are modelized by analog behavior modeling and digital circuit model, modified to process input frequency sufficiently, in SPICE.

  • PDF

Secure and Efficient Privacy-Preserving Identity-Based Batch Public Auditing with Proxy Processing

  • Zhao, Jining;Xu, Chunxiang;Chen, Kefei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.2
    • /
    • pp.1043-1063
    • /
    • 2019
  • With delegating proxy to process data before outsourcing, data owners in restricted access could enjoy flexible and powerful cloud storage service for productivity, but still confront with data integrity breach. Identity-based data auditing as a critical technology, could address this security concern efficiently and eliminate complicated owners' public key certificates management issue. Recently, Yu et al. proposed an Identity-Based Public Auditing for Dynamic Outsourced Data with Proxy Processing (https://doi.org/10.3837/tiis.2017.10.019). It aims to offer identity-based, privacy-preserving and batch auditing for multiple owners' data on different clouds, while allowing proxy processing. In this article, we first demonstrate this scheme is insecure in the sense that malicious cloud could pass integrity auditing without original data. Additionally, clouds and owners are able to recover proxy's private key and thus impersonate it to forge tags for any data. Secondly, we propose an improved scheme with provable security in the random oracle model, to achieve desirable secure identity based privacy-preserving batch public auditing with proxy processing. Thirdly, based on theoretical analysis and performance simulation, our scheme shows better efficiency over existing identity-based auditing scheme with proxy processing on single owner and single cloud effort, which will benefit secure big data storage if extrapolating in real application.