• Title/Summary/Keyword: multiple token

Search Result 36, Processing Time 0.022 seconds

A Self-Timed Ring based Lightweight TRNG with Feedback Structure (피드백 구조를 갖는 Self-Timed Ring 기반의 경량 TRNG)

  • Choe, Jun-Yeong;Shin, Kyung-Wook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.2
    • /
    • pp.268-275
    • /
    • 2020
  • A lightweight hardware design of self-timed ring based true random number generator (TRNG) suitable for information security applications is described. To reduce hardware complexity of TRNG, an entropy extractor with feedback structure was proposed, which minimizes the number of ring stages. The number of ring stages of the FSTR-TRNG was determined to be a multiple of eleven, taking into account operating clock frequency and entropy extraction circuit, and the ratio of tokens to bubbles was determined to operate in evenly-spaced mode. The hardware operation of FSTR-TRNG was verified by FPGA implementation. A set of statistical randomness tests defined by NIST 800-22 were performed by extracting 20 million bits of binary sequences generated by FSTR-TRNG, and all of the fifteen test items were found to meet the criteria. The FSTR-TRNG occupied 46 slices of Spartan-6 FPGA device, and it was implemented with about 2,500 gate equivalents (GEs) when synthesized in 180 nm CMOS standard cell library.

A Study on Suggestions for Activating Smart Contract - Focusing on Software Export Business (스마트계약의 활성화 방안에 관한 연구 - 소프트웨어 수출사업을 중심으로)

  • Whayoon Song
    • Korea Trade Review
    • /
    • v.47 no.1
    • /
    • pp.163-180
    • /
    • 2022
  • The purpose of this study is to examine the extent to which smart contracts can be applied to the software export business and to find out the legislative issues to activate smart contracts. A smart contract is a computer program that automatically executes a contract when conditions are fulfilled. Smart contracts can play a pivotal role in the field that requires immediate execution of contract or in a highly standardized field with multiple parties involved. In the software export business, it is desirable to apply the smart contract partially rather than applying the smart contract to the entire process because various parties are involved and the process is very complicated. The business model of exporting packaged software, a completed software that is mainly licensed for use, rather than the business model of exporting customized software is suitable for using smart contracts because the project for implementing customized software is mainly focused in the development stage. When smart contracts are used in processes such as contract signing, payment, and project management, work efficiency can be increased. In addition, smart contracts can be used when conditions can be quantified, such as error penalties, in areas that previously required contracts with third parties such as banks, guarantors. In order for smart contracts to be actively used in practice, legal reviews on various issues are necessary including the legality of a smart contract and the validity as an electronic document of NFT (non-fungible token) certificate. Also, for the system stability preventing hacking, etc, the periodic verification or inspection by a third party is essential. To activate smart contracts in international transactions the international treaty regarding smart contracts is also necessary.

Authentication Mechanism Using Three-Dimensional Optical Memory (3차원 광메모리를 이용한 인증 기법)

  • Park, CheolYong;Ryou, JaeCheol
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.26 no.6
    • /
    • pp.1361-1373
    • /
    • 2016
  • Recently the need for user authentication with increasing, there are a variety of mechanisms, such as password, graphic authentication, token, biometrics and multiple authentication. in particular, the data of the 2-dimensional(2D) factors such as password, graphic authentication, biometrics is used because of the convenience. The stored information is problematic in that additional data recording needs to be performed whenever authentication data is updated. Furthermore, this storage method is problematic in that the time it takes to perform authentication increases because the time it takes to compare storage data with authentication data increases in proportion to an increase in the amount of the storage data. Accordingly, authentication through the rapid comparison of storage data with authentication data is a very important factor in data recording and authentication technology using memory. Using the three-dimensional(3D) optical memory by variously changing the recoding elements during recoding of data constitutes the way that multiple recoding different data storage. This enables high-density recoding in this way, and by applying the possible parallel processing at the time of recording and restoring method, provided that it is possible to quickly record and restore the data. In addition, each time to solve problems that require additional data recorded by a combination of the stored data record in the old data using a combination of the authentication. The proposed mechanism is proposed an authentication method using scheme after the recoding data in 3D optical memory to apply the conditions corresponding to the recoding condition when restoring the recorded data and through the experiment it was confirmed possible application as an authentication mechanism.

A Study of Autonomous Intelligent Load Management System Based on Queueing Model (큐잉모델에 기초한 자율 지능 부하 관리 시스템 연구)

  • Lee, Seung-Chul;Hong, Chang-Ho;Kim, Kyung-Dong;Lee, In-Yong;Park, Chan-Eom
    • Journal of the Korean Institute of Illuminating and Electrical Installation Engineers
    • /
    • v.22 no.2
    • /
    • pp.134-141
    • /
    • 2008
  • This paper presents an innovative load management technique that can effectively lower the summer peak load by adjusting the aircondition loads through smoothe coordinations between utility companies and large customers. An intelligent hierarchical load management system composed of a Central Intelligent Load Management System(CIMS) and multiple Local Intelligent Management Systems(LIMS) is also proposed to implement the reposed technique. Upon receiving a load curtailment request from the utilities, CIMS issues tokens, which can be used by each LIMS as a right to turn on the airconditioner. CIMS creates and maintains a queue for fair allocation of the tokens among the LIMS demanding tokens. By adjusting the number tokens and queue management Policies, desired load factors can be achieved conveniently. The Markov Birth and Death Process and the Balance Equations are employed in estimating various queue performances. The proposed technique is tested using a summer load data of a large apartment complex and proved to be quite effective in load management while minimizing the customer inconveniences.

An Aggregate Fairness Marker without Per Flow Management for Fairness Improvement of Assured Service in DiffServ (DiffServ 방식의 Assured Service 에서 플로별 관리 없이 Fairness향상을 위한 Aggregate Fairness Marker)

  • Park, Ji-Hoon;Hur, Kyeong;Eom, Doo-Seop
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.7B
    • /
    • pp.613-627
    • /
    • 2004
  • In this paper, we propose an Aggregate Fairness Maker (ARM) required for an Edge router to improve fairness of throughput among the flows of Assured Service in DiffServ with different round trip time (RTT) and we propose a user flow Three Color Marker (uf-TCM) as a flow marker that marks packets from the flow as green, yellow, or red. A yellow packet is the packet that consumes loss token in uf-TCM as well as that is demoted green packet in AM due to disobey the aggregate traffic profile. The proposed AFH promotes yellow packet to green packet or demotes green packet to yellow packet through the fair method without per-flow management, and it improves the feirness of throughput among the flows as well as link utilization. A yellow packet and a red packet have the same drop precedence at Core Router in our scheme. So we can use the RIO buffer management scheme. We evaluated the performance of our proposed AFM and the REDP Marker that was proposed to improve fairness without per-flow management. Simulation results show that, compared with the REDP marker, proposed AFM can improve performance of throughput fairness among the flows with different RTT and link utilization under the over-provisioning, exact-provisioning, and under-provisioning network environments at Multiple DiffServ domains as well as at Single DiffServ domain.

Multi-Vector Document Embedding Using Semantic Decomposition of Complex Documents (복합 문서의 의미적 분해를 통한 다중 벡터 문서 임베딩 방법론)

  • Park, Jongin;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.19-41
    • /
    • 2019
  • According to the rapidly increasing demand for text data analysis, research and investment in text mining are being actively conducted not only in academia but also in various industries. Text mining is generally conducted in two steps. In the first step, the text of the collected document is tokenized and structured to convert the original document into a computer-readable form. In the second step, tasks such as document classification, clustering, and topic modeling are conducted according to the purpose of analysis. Until recently, text mining-related studies have been focused on the application of the second steps, such as document classification, clustering, and topic modeling. However, with the discovery that the text structuring process substantially influences the quality of the analysis results, various embedding methods have actively been studied to improve the quality of analysis results by preserving the meaning of words and documents in the process of representing text data as vectors. Unlike structured data, which can be directly applied to a variety of operations and traditional analysis techniques, Unstructured text should be preceded by a structuring task that transforms the original document into a form that the computer can understand before analysis. It is called "Embedding" that arbitrary objects are mapped to a specific dimension space while maintaining algebraic properties for structuring the text data. Recently, attempts have been made to embed not only words but also sentences, paragraphs, and entire documents in various aspects. Particularly, with the demand for analysis of document embedding increases rapidly, many algorithms have been developed to support it. Among them, doc2Vec which extends word2Vec and embeds each document into one vector is most widely used. However, the traditional document embedding method represented by doc2Vec generates a vector for each document using the whole corpus included in the document. This causes a limit that the document vector is affected by not only core words but also miscellaneous words. Additionally, the traditional document embedding schemes usually map each document into a single corresponding vector. Therefore, it is difficult to represent a complex document with multiple subjects into a single vector accurately using the traditional approach. In this paper, we propose a new multi-vector document embedding method to overcome these limitations of the traditional document embedding methods. This study targets documents that explicitly separate body content and keywords. In the case of a document without keywords, this method can be applied after extract keywords through various analysis methods. However, since this is not the core subject of the proposed method, we introduce the process of applying the proposed method to documents that predefine keywords in the text. The proposed method consists of (1) Parsing, (2) Word Embedding, (3) Keyword Vector Extraction, (4) Keyword Clustering, and (5) Multiple-Vector Generation. The specific process is as follows. all text in a document is tokenized and each token is represented as a vector having N-dimensional real value through word embedding. After that, to overcome the limitations of the traditional document embedding method that is affected by not only the core word but also the miscellaneous words, vectors corresponding to the keywords of each document are extracted and make up sets of keyword vector for each document. Next, clustering is conducted on a set of keywords for each document to identify multiple subjects included in the document. Finally, a Multi-vector is generated from vectors of keywords constituting each cluster. The experiments for 3.147 academic papers revealed that the single vector-based traditional approach cannot properly map complex documents because of interference among subjects in each vector. With the proposed multi-vector based method, we ascertained that complex documents can be vectorized more accurately by eliminating the interference among subjects.