• Title/Summary/Keyword: boolean function

Search Result 84, Processing Time 0.028 seconds

The design of digital circuit for chaotic composition map (혼돈합성맵의 디지털회로설계)

  • Park, Kwang-Hyeon;Seo, Yong-Won
    • Journal of Advanced Navigation Technology
    • /
    • v.17 no.6
    • /
    • pp.652-657
    • /
    • 2013
  • In this paper the design methode of a separated composition state machine based on the compositive map with two chaotic maps together and the result of that is proposed. The digital circuits of chaotic composition map for the use of chaotic binary stream generator are designed in this work. The discretized truth table of chaotic composition function which is composed of two chaotic functions - the saw tooth function and skewed logistic function - is made out, and also simplefied Boolean algebras of digital circuits are obtained as a mathematical model. Consequently, the digital circuits of the map for chaotic composition function are presented in this paper.

Analysis of 74181 Arithmetic Logic Units (74184 Arithmetic Logic Units의 분석)

  • Lee, Jae-Seok;Chung, Tae-Sang
    • Proceedings of the KIEE Conference
    • /
    • 2000.11d
    • /
    • pp.778-780
    • /
    • 2000
  • The 74181 is arithmetic logic units(ALU)/function generator. This circuit performs 16 binary arithmetic operations on two 4-bit words. And a full carry look-ahead scheme is made available in this device. The 74181 can also be utilized as a comparator. This circuit has been also designed to provide 16 possible functions of two Boolean variables without the use of external circuitry. This paper analyzes the function of the logic and the implementation adopted in the design of 74181. The understanding of the logic characteristics of this chip enables us to improve future applications.

  • PDF

Terminal-Pair Availability Evaluation in Irreducible Structure Communication Networks (비가략(非可略) 구조를 갖는 네트워크의 단점간 가용도 평가)

  • Lee, Jun-Hyuk;Kim, Kyoung-Mok;Oh, Young-Hwan
    • Journal of Applied Reliability
    • /
    • v.8 no.4
    • /
    • pp.181-192
    • /
    • 2008
  • In this paper, we proposed effective algorithm that do Availability of communication network account using total probability formula. Use path tracing method about PSTN linked with MSC in WCDMA considering occasion that do not consider mobility of terminal and Availability did account. First, express minimum path series gathering by structure function after network that do modeling by stochastic graph for communication network saves all minimal path sets between particular two abutments and we gain Availability value in angular variable after express structure function by Boolean operation form and Availability between two abutments did account.

  • PDF

Analysis of hash functions based on cellular automata (셀룰러 오토마타 기반 해쉬 함수 분석)

  • Jeong Kitae;Lee Jesang;Chang Donghoon;Sung Jaechul;Lee Sangjin
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.14 no.6
    • /
    • pp.111-123
    • /
    • 2004
  • A hash function is a function that takes bit strings of arbitrary length to bit string of fixed length. A cellular automata is a finite state machine and has the property of generating pseudorandom numbers efficiently by combinational logics of neighbour cells. In [1] and [7], hash functions based on cellular automata which can be implemented efficiently in hardware were proposed. In this paper, we show that we can find collisions of these hash functions with probability 0.46875 and 0.5 respectively.

온라인 목록 검색 행태에 관한 연구-LINNET 시스템의 Transaction log 분석을 중심으로-

  • 윤구호;심병규
    • Journal of Korean Library and Information Science Society
    • /
    • v.21
    • /
    • pp.253-289
    • /
    • 1994
  • The purpose of this study is about the search pattern of LINNET (Library Information Network System) OPAC users by transaction log, maintained by POSTECH(Pohang University of Science and Technology) Central Library, to provide feedback information of OPAC system design. The results of this study are as follows. First, for the period of this analysis, there were totally 11, 218 log-ins, 40, 627 transaction logs and 3.62 retrievals per a log-in. Title keyword was the most frequently used, but accession number, bibliographic control number or call number was very infrequently used. Second, 47.02% of OPAC, searches resulted in zero retrievals. Bibliographic control number was the least successful search. User displayed 2.01% full information and 64.27% local information per full information. Third, special or advanced retrieval features are very infrequently used. Only 22.67% of the searches used right truncation and 0.71% used the qualifier. Only 1 boolean operator was used in every 22 retrievals. The most frequently used operator is 'and (&)' with title keywords. But 'bibliographical control number (N) and accessionnumber (R) are not used at all with any operators. The causes of search failure are as follows. 1. The item was not used in the database. (15, 764 times : 79.42%). 2. The wrong search key was used. (3, 761 times : 18.95%) 3. The senseless string (garbage) was entered. (324 times : 1.63%) On the basis of these results, some recommendations are suggested to improve the search success rate as follows. First, a n.0, ppropriate user education and online help function let users retrieve LINNET OPAC more efficiently. Second, several corrections of retrieval software will decrease the search failure rate. Third, system offers right truncation by default to every search term. This methods will increase success rate but should considered carefully. By a n.0, pplying this method, the number of hit can be overnumbered, and system overhead can be occurred. Fourth, system offers special boolean operator by default to every keyword retrieval when user enters more than two words at a time. Fifth, system assists searchers to overcome the wrong typing of selecting key by automatic korean/english mode change.

  • PDF

Analysis on Vulnerability of Masked SEED Algorithm (마스킹 기법이 적용된 SEED 알고리즘에 대한 취약점 분석)

  • Kim, TaeWon;Chang, Nam Su
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.25 no.4
    • /
    • pp.739-747
    • /
    • 2015
  • Masking technique that is most widely known as countermeasure against power analysis attack prevents leakage for sensitive information during the implementations of cryptography algorithm. it have been studied extensively until now applied on block cipher algorithms. Masking countermeasure have been applied to international standard SEED algorithm. Masked SEED algorithm proposed by Cho et al, not only protects against first order power analysis attacks but also efficient by reducing the execution of Arithmetic to Boolean converting function. In this paper, we analyze the vulnerability of Cho's algorithm against first order power analysis attacks. We targeted additional pre-computation to improve the efficiency in order to recover the random mask value being exploited in first order power analysis attacks. We describe weakness by considering both theoretical and practical aspects and are expecting to apply on every device equipped with cho's algorithm using the proposed attack method.

3-bit Up/Down Counter based on Magnetic-Tunnel-Junction Elements (Magnetic-Tunnel-Junction 소자를 이용한 3비트 업/다운 카운터)

  • Lee, Seung-Yeon;Kim, Ji-Hyun;Lee, Gam-Young;Yang, Hee-Jung;Lee, Seung-Jun;Shin, Hyung-Soon
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.44 no.1
    • /
    • pp.1-7
    • /
    • 2007
  • An MTJ element not only computes Boolean function but also stores the output result in itself. We can make the most use of magneto-logic's merits by employing the magneto-logic in substitution for the sequential logic as well as the combinational logic. This unique feature opens a new horizon for potential application of MTJ as a universal logic element. Magneto-logic circuits using MTJ elements are more integrative and non-volatile. This paper presents novel 3-bit magneto-logic up/down counters and presents simulation results based on the HSPICE macro-model of MTJ that we have developed.

Image Encryption using the chaos function and elementary matrix operations (혼돈함수와 기본 행렬 연산을 이용한 영상의 암호화)

  • Kim Tae-Sik
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.11 no.1
    • /
    • pp.29-37
    • /
    • 2006
  • Due to the spread of mobile communication with the development of computer network, nowadays various types of multimedia data play an important role in many areas such as entertainments, culture contents, e-commerce or medical science. But for the real application of these data, the security in the course of saving or transferring them through the public network should be assured. In this sense, many encryption algorithm have been developed and utilized. Nonetheless, most of them have focused on the text data. So they may not be suitable to the multimedia application because of their large size and real time constraint. In this paper, a chaotic map has been employed to create a symmetric stream type of encryption scheme which may be applied to the digital images with a large amounts of data. Then an efficient algebraic encryption algorithm based on the elementary operations of the Boolean matrix and image data characteristics.

  • PDF

A Study on WT-Algorithm for Effective Reduction of Association Rules (효율적인 연관규칙 감축을 위한 WT-알고리즘에 관한 연구)

  • Park, Jin-Hee;Pi, Su-Young
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.20 no.5
    • /
    • pp.61-69
    • /
    • 2015
  • We are in overload status of information not just in a flood of information due to the data pouring from various kinds of mobile devices, online and Social Network Service(SNS) every day. While there are many existing information already created, lots of new information has been created from moment to moment. Linkage analysis has the shortcoming in that it is difficult to find the information we want since the number of rules increases geometrically as the number of item increases with the method of finding out frequent item set where the frequency of item is bigger than minimum support in this information. In this regard, this thesis proposes WT-algorithm that represents the transaction data set as Boolean variable item and grants weight to each item by making algorithm with Quine-McKluskey used to simplify the logical function. The proposed algorithm can improve efficiency of data mining by reducing the unnecessary rules due to the advantage of simplification regardless of number of items.

Rough Set Analysis for Stock Market Timing (러프집합분석을 이용한 매매시점 결정)

  • Huh, Jin-Nyung;Kim, Kyoung-Jae;Han, In-Goo
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.3
    • /
    • pp.77-97
    • /
    • 2010
  • Market timing is an investment strategy which is used for obtaining excessive return from financial market. In general, detection of market timing means determining when to buy and sell to get excess return from trading. In many market timing systems, trading rules have been used as an engine to generate signals for trade. On the other hand, some researchers proposed the rough set analysis as a proper tool for market timing because it does not generate a signal for trade when the pattern of the market is uncertain by using the control function. The data for the rough set analysis should be discretized of numeric value because the rough set only accepts categorical data for analysis. Discretization searches for proper "cuts" for numeric data that determine intervals. All values that lie within each interval are transformed into same value. In general, there are four methods for data discretization in rough set analysis including equal frequency scaling, expert's knowledge-based discretization, minimum entropy scaling, and na$\ddot{i}$ve and Boolean reasoning-based discretization. Equal frequency scaling fixes a number of intervals and examines the histogram of each variable, then determines cuts so that approximately the same number of samples fall into each of the intervals. Expert's knowledge-based discretization determines cuts according to knowledge of domain experts through literature review or interview with experts. Minimum entropy scaling implements the algorithm based on recursively partitioning the value set of each variable so that a local measure of entropy is optimized. Na$\ddot{i}$ve and Booleanreasoning-based discretization searches categorical values by using Na$\ddot{i}$ve scaling the data, then finds the optimized dicretization thresholds through Boolean reasoning. Although the rough set analysis is promising for market timing, there is little research on the impact of the various data discretization methods on performance from trading using the rough set analysis. In this study, we compare stock market timing models using rough set analysis with various data discretization methods. The research data used in this study are the KOSPI 200 from May 1996 to October 1998. KOSPI 200 is the underlying index of the KOSPI 200 futures which is the first derivative instrument in the Korean stock market. The KOSPI 200 is a market value weighted index which consists of 200 stocks selected by criteria on liquidity and their status in corresponding industry including manufacturing, construction, communication, electricity and gas, distribution and services, and financing. The total number of samples is 660 trading days. In addition, this study uses popular technical indicators as independent variables. The experimental results show that the most profitable method for the training sample is the na$\ddot{i}$ve and Boolean reasoning but the expert's knowledge-based discretization is the most profitable method for the validation sample. In addition, the expert's knowledge-based discretization produced robust performance for both of training and validation sample. We also compared rough set analysis and decision tree. This study experimented C4.5 for the comparison purpose. The results show that rough set analysis with expert's knowledge-based discretization produced more profitable rules than C4.5.