• Title/Summary/Keyword: Duplicate Verification

Search Result 14, Processing Time 0.02 seconds

Verification Algorithm for the Duplicate Verification Data with Multiple Verifiers and Multiple Verification Challenges

  • Xu, Guangwei;Lai, Miaolin;Feng, Xiangyang;Huang, Qiubo;Luo, Xin;Li, Li;Li, Shan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.2
    • /
    • pp.558-579
    • /
    • 2021
  • The cloud storage provides flexible data storage services for data owners to remotely outsource their data, and reduces data storage operations and management costs for data owners. These outsourced data bring data security concerns to the data owner due to malicious deletion or corruption by the cloud service provider. Data integrity verification is an important way to check outsourced data integrity. However, the existing data verification schemes only consider the case that a verifier launches multiple data verification challenges, and neglect the verification overhead of multiple data verification challenges launched by multiple verifiers at a similar time. In this case, the duplicate data in multiple challenges are verified repeatedly so that verification resources are consumed in vain. We propose a duplicate data verification algorithm based on multiple verifiers and multiple challenges to reduce the verification overhead. The algorithm dynamically schedules the multiple verifiers' challenges based on verification time and the frequent itemsets of duplicate verification data in challenge sets by applying FP-Growth algorithm, and computes the batch proofs of frequent itemsets. Then the challenges are split into two parts, i.e., duplicate data and unique data according to the results of data extraction. Finally, the proofs of duplicate data and unique data are computed and combined to generate a complete proof of every original challenge. Theoretical analysis and experiment evaluation show that the algorithm reduces the verification cost and ensures the correctness of the data integrity verification by flexible batch data verification.

A Study on Duplication Verification of Public Library Catalog Data: Focusing on the Case of G Library in Busan (공공도서관 목록데이터의 중복검증에 관한 연구 - 부산 지역 G도서관 사례를 중심으로 -)

  • Min-geon Song;Soo-Sang Lee
    • Journal of Korean Library and Information Science Society
    • /
    • v.55 no.1
    • /
    • pp.1-26
    • /
    • 2024
  • The purpose of this study is to derive an integration plan for bibliographic records by applying a duplicate verification algorithm to the item-based catalog in public libraries. To this, G Library, which was opened recently in Busan, was selected. After collecting OPAC data from G Library through web crawling, multipart monographs of Korean Literature (KDC 800) were selected and KERIS duplicate verification algorithm was applied. After two rounds of data correction based on the verification results, the duplicate verification rate increased by a total of 2.74% from 95.53% to 98.27%. Even after data correction, 24 books that were judged to be similar or inconsistent were identified as data from other published editions after receiving separate ISBN such as revised versions or hard copies. Through this, it was confirmed that the duplicate verification rate could be improved through catalog data correction work, and the possibility of using the KERIS duplicate verification algorithm as a tool to convert duplicate item-based records from public libraries into manifestation-based records was confirmed.

A Post-Verification Method of Near-Duplicate Image Detection using SIFT Descriptor Binarization (SIFT 기술자 이진화를 이용한 근-복사 이미지 검출 후-검증 방법)

  • Lee, Yu Jin;Nang, Jongho
    • Journal of KIISE
    • /
    • v.42 no.6
    • /
    • pp.699-706
    • /
    • 2015
  • In recent years, as near-duplicate image has been increasing explosively by the spread of Internet and image-editing technology that allows easy access to image contents, related research has been done briskly. However, BoF (Bag-of-Feature), the most frequently used method for near-duplicate image detection, can cause problems that distinguish the same features from different features or the different features from same features in the quantization process of approximating a high-level local features to low-level. Therefore, a post-verification method for BoF is required to overcome the limitation of vector quantization. In this paper, we proposed and analyzed the performance of a post-verification method for BoF, which converts SIFT (Scale Invariant Feature Transform) descriptors into 128 bits binary codes and compares binary distance regarding of a short ranked list by BoF using the codes. Through an experiment using 1500 original images, it was shown that the near-duplicate detection accuracy was improved by approximately 4% over the previous BoF method.

Improved Facial Component Detection Using Variable Parameter and Verification (가변 변수와 검증을 이용한 개선된 얼굴 요소 검출)

  • Oh, Jeong-su
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.3
    • /
    • pp.378-383
    • /
    • 2020
  • Viola & Jones' object detection algorithm is a very good algorithm for the face component(FC) detection, but there are still problems such as duplicate detection, false detection and non-detection due to parameter setting. This paper proposes an improved FC detection algorithm that applies the variable parameter to reduce non-detection and the verification to reduce duplicate detection and false detection to the Viola & Jones' algorithm. The proposed algorithm reduces the non-detection by changing the parameter value of the Viola & Jones' algorithm until the potential valid FCs are detected, and eliminates the duplicate detection and the false detection by using the verification that evaluates size, position, and uniqueness of the detected FCs. Simulation results show that the proposed algorithm includes valid FCs in the detected objects and then detects only the valid FCs by removing invalid FCs from them.

Improved Face Detection Algorithm Using Face Verification (얼굴 검증을 이용한 개선된 얼굴 검출)

  • Oh, Jeong-su
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.22 no.10
    • /
    • pp.1334-1339
    • /
    • 2018
  • Viola & Jones's face detection algorithm is a typical face detection algorithm and shows excellent face detection performance. However, the Viola & Jones's algorithm in images including many faces generates undetected faces and wrong detected faces, such as false faces and duplicate detected faces, due to face diversity. This paper proposes an improved face detection algorithm using a face verification algorithm that eliminates the false detected faces generated from the Viola & Jones's algorithm. The proposed face verification algorithm verifies whether the detected face is valid by evaluating its size, its skin color in the designated area, its edges generated from eyes and mouth, and its duplicate detection. In the face verification experiment of 658 face images detected by the Viola & Jones's algorithm, the proposed face verification algorithm shows that all the face images created in the real person are verified.

A study on the application of blockchain technology to prevent duplicate supply and demand of similar welfare services (복지서비스 유사사업의 중복수급 방지를 위한 블록체인 기술 적용 연구)

  • Min, Youn A
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.20 no.6
    • /
    • pp.151-156
    • /
    • 2020
  • Various institutions provide various welfare-related services, and accordingly, the quantity and quality of welfare-related services are improved. Along with the improvement of welfare services, the rate of fraudulent and redundant supply and demand due to inappropriate applicants is also increasing. In particular, in the case of similarly promoted projects by various institutions, confirmation of duplicate applications may be insufficient due to system inconsistency between institutions and delays in qualification verification, which may result in duplicate supply and demand. In this paper, in order to prevent the redundant supply and demand related to welfare service similar projects in various institutions, the problem of data management and sharing between institutions was reviewed and a method of applying blockchain technology was proposed step by step. Through the proposal of this paper, transparent data management of recipients is possible, and through this, trust-based welfare benefit management will be possible.

MAC Layer Based Certificate Authentication for Multiple Certification Authority in MANET

  • Sekhar, J. Chandra;Prasad, Ramineni Sivarama
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.3 no.5
    • /
    • pp.298-305
    • /
    • 2014
  • In this study, a novel Randomly Shifted Certification Authority Authentication protocol was used in ad hoc networks to provide authentication by considering the MAC layer characteristics. The nodes achieve authentication through the use of public key certificates issued by a CA, which assures the certificate's ownership. As a part of providing key management, the active CA node transfers the image of the stored public keys to other idle CA nodes. Finally the current active CA randomly selects the ID of the available idle CA and shifts the CA ownership by transferring it. Revoking is done if any counterfeit or duplicate non CA node ID is found. Authentication and integrity is provided by preventing MAC control packets, and Enhanced Hash Message Authentication Code (EHMAC) can be used. Here EHMAC with various outputs is introduced in all control packets. When a node transmits a packet to a node with EHMAC, verification is conducted and the node replies with the transmitter address and EHMAC in the acknowledgement.

Implementation of the Stone Classification with AI Algorithm Based on VGGNet Neural Networks (VGGNet을 활용한 석재분류 인공지능 알고리즘 구현)

  • Choi, Kyung Nam
    • Smart Media Journal
    • /
    • v.10 no.1
    • /
    • pp.32-38
    • /
    • 2021
  • Image classification through deep learning on the image from photographs has been a very active research field for the past several years. In this paper, we propose a method of automatically discriminating stone images from domestic source through deep learning, which is to use Python's hash library to scan 300×300 pixel photo images of granites such as Hwangdeungseok, Goheungseok, and Pocheonseok, performing data preprocessing to create learning images by examining duplicate images for each stone, removing duplicate images with the same hash value as a result of the inspection, and deep learning by stone. In addition, to utilize VGGNet, the size of the images for each stone is resized to 224×224 pixels, learned in VGG16 where the ratio of training and verification data for learning is 80% versus 20%. After training of deep learning, the loss function graph and the accuracy graph were generated, and the prediction results of the deep learning model were output for the three kinds of stone images.

Study on the efficient consensus process of PBFT

  • Min, Youn-A
    • Journal of the Korea Society of Computer and Information
    • /
    • v.25 no.4
    • /
    • pp.47-53
    • /
    • 2020
  • Blockchain is a distributed shared ledger that transparently manages information through verification and agreement between nodes connected to a distributed network. Recently, cases of data management among authorized agencies based on private blockchain are increasing. In this paper, we investigated the application cases and technical processes of PBFT, the representative consensus algorithm of private blockchain, and proposed a modified PBFT algorithm that enables efficient consensus by simplifying duplicate verification and consensus processes that occur during PBFT processing. The algorithm proposed in this paper goes through the process of selecting a delegation node through an authoritative node and can increase the safety of the delegation node selection process by considering an efficient re-election algorithm for candidate nodes. By utilizing this research, it is possible to reduce the burden on the network communication cost of the consensus process and effectively process the final consensus process between nodes.

A Strong Address Generation and Verification Scheme using Hash Functions in the IPv6 Environments (IPv6 환경에서 해쉬 함수 기반 강건한 주소 생성 및 검증 기법)

  • Gyeong, Gye-Hyeon;Ko, Kwang-Sun;Eom, Young-Ik
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.17 no.1
    • /
    • pp.115-119
    • /
    • 2007
  • The IPv6 protocol provides the method to automatically generate an address of a node without additional operations of administrators, Before the generated address is used, the duplicate address detection (DAD) mechanism is required in order to verify the address. However, during the process of verification of the address, it is possible for a malicious node to send a message with the address which is identical with the generated address, so the address can be considered as previously used one; although the node properly generates an address, the address cannot be used. In this paper, we present a strong scheme to perform the DAD mechanism based on hash functions in IPv6 networks. Using this scheme, many nodes, which frequently join or separate from wireless networks in public domains like airports, terminals, and conference rooms, can effectively generate and verify an address more than the secure neighbor discovery (SEND) mechanism.