• Title/Summary/Keyword: memory constraint

Search Result 70, Processing Time 0.021 seconds

A Design of Parameterized Viterbi Decoder for Multi-standard Applications (다중 표준용 파라미터화된 비터비 복호기 IP 설계)

  • Park, Sang-Deok;Jeon, Heung-Woo;Shin, Kyung-Wook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.12 no.6
    • /
    • pp.1056-1063
    • /
    • 2008
  • This paper describes an efficient design of a multi-standard Viterbi decoder that supports multiple constraint lengths and code rates. The Viterbi decoder is parameterized for the code rates 1/2, 1/3 and constraint lengths 7,9, thus it has four operation nodes. In order to achieve low hardware complexity and low power, an efficient architecture based on hardware sharing techniques is devised. Also, the optimization of ACCS (Accumulate-Subtract) circuit for the one-point trace-back algorithm reduces its area by about 35% compared to the full parallel ACCS circuit. The parameterized Viterbi decoder core has 79,818 gates and 25,600 bits memory, and the estimated throughput is about 105 Mbps at 70 MHz clock frequency. Also, the simulation results for BER (Bit Error Rate) performance show that the Viterbi decoder has BER of $10^{-4}$ at $E_b/N_o$ of 3.6 dB when it operates with code rate 1/3 and constraints 7.

Location Generalization Method of Moving Object using $R^*$-Tree and Grid ($R^*$-Tree와 Grid를 이용한 이동 객체의 위치 일반화 기법)

  • Ko, Hyun;Kim, Kwang-Jong;Lee, Yon-Sik
    • Journal of the Korea Society of Computer and Information
    • /
    • v.12 no.2 s.46
    • /
    • pp.231-242
    • /
    • 2007
  • The existing pattern mining methods[1,2,3,4,5,6,11,12,13] do not use location generalization method on the set of location history data of moving object, but even so they simply do extract only frequent patterns which have no spatio-temporal constraint in moving patterns on specific space. Therefore, it is difficult for those methods to apply to frequent pattern mining which has spatio-temporal constraint such as optimal moving or scheduling paths among the specific points. And also, those methods are required more large memory space due to using pattern tree on memory for reducing repeated scan database. Therefore, more effective pattern mining technique is required for solving these problems. In this paper, in order to develop more effective pattern mining technique, we propose new location generalization method that converts data of detailed level into meaningful spatial information for reducing the processing time for pattern mining of a massive history data set of moving object and space saving. The proposed method can lead the efficient spatial moving pattern mining of moving object using by creating moving sequences through generalizing the location attributes of moving object into 2D spatial area based on $R^*$-Tree and Area Grid Hash Table(AGHT) in preprocessing stage of pattern mining.

  • PDF

The effect of syntatic and pragmatic Constraints on Sentential Representaition and Memory Accessibility (통사적 제약과 화용적 제약이 문장의 표상과 기억접근에 미치는 효과)

  • Kim, Sung-Il;Lee, Jae-Ho
    • Korean Journal of Cognitive Science
    • /
    • v.6 no.2
    • /
    • pp.97-116
    • /
    • 1995
  • This study was conducted to investigate how syntaction and pragmatic constraints influence the sentential representation and memory accessibility. In order to seperate the syntactic constraints from the pragmatic constraint from the pragmatic constraints,the syntactic role of constituent in the sentence (subject or object) and the order of mention(first or second) were manipulted.After each sentence was presented by RSVP procedure,the probe recognition time was measured to investigate memory accessibility.In Experiment 1,in which SOA interval was 255ms,it was found that the subject of a sentece were more accessible than the object and participants first in a sentence were more accessible than participants mentioned later.However, in Experiment 2,in which SOA interval was 1540ms,it was found that participants mentioned first in a sentence were more accessible than participants mentioned later while there was no significant difference between the subject and object of a sentece.These results suggest that the syntactic and pragmatic constraints have an independent effect on the initial senential representation at the early stage of constructing representation,but as time passes only the pragmatic constraints influence sentential representation.These results also support a theoretical position which assumes that sentential representation is constructed through the process of convergent statisfaction of multiple constraints.

  • PDF

Linguistic Productivity and Chomskyan Grammar: A Critique (언어창조성과 춈스키 문법 비판)

  • Bong-rae Seok
    • Lingua Humanitatis
    • /
    • v.1 no.1
    • /
    • pp.235-251
    • /
    • 2001
  • According to Chomskyan grammar, humans can generate and understand an unbounded number of grammatical sentences. Against the background of pure and idealized linguistic competence, this linguistic productivity is argued and understood. In actual utterances, however, there are many limitations of productivity but they are said to come from the general constraints on performances such as capacity of short term memory or attention. In this paper I discuss a problem raised against idealized productivity. I argue that linguistic productivity idealizes our linguistic competence too much. By separating idealized competence from the various constraints of performance, Chomskyan theorists can argue for unlimited productivity. However, the absolute distinction between grammar (pure competence) and parser (actual psychological processes) makes little sense when we explain the low acceptability(intelligibility) of center embedded sentences. Usually, the problem of center embedded sentence is explained in terms of memory shortage or other performance constraints. To explain the low acceptability, however, we need to assume specialized memory structure because the low acceptability occurs only with a specific type of syntactic pattern. 1 argue that this special memory structure should not be considered as a general performance constraint. It is a domain specific (specifically linguistic) constraints and an intrinsic part of human language processing. Recent development of Chomskyan grammar, i.e., minimalist approach seems to close the gap between pure competence and this type of specialized constraints. Chomsky's earlier approach of generative grammar focuses on end result of the generative derivation. However, economy principle (of minimalist approach) focuses on actual derivational processes. By having less mathematical or less idealized grammar, we can come closer to the actual computational processes that build syntactic structure of a sentence. In this way, we can have a more concrete picture of our linguistic competence, competence that is not detached from actual computational processes.

  • PDF

A Construction of Pointer-based Model for Main Memory Database Systems (주기억장치 데이터베이스를 위한 포인터 기반 모델의 구축)

  • Bae, Myung-Nam;Choi, Wan
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.28 no.4B
    • /
    • pp.323-338
    • /
    • 2003
  • The main memory database systems (MMDBMS) efficiently supports various database applications that require high performance since it employs main memory rather than disk as a primary storage. Recently, it has been increased needs that have the fast data processing as well as the efficient modeling of application requiring for a complicated structure, and conformity to applications that need the strict dta consistency. In MMDBMS, because all the data is located in the main memory, it can support the usable expression methods of data satisfying their needs without performance overhead. The method has the operation to manipulate the data and the constraint such as referential integrity in more detail. The data model consists of this methods is an essential component to decide the expression power of DBMS. In this paper, we discuss about various requests to provide the communication services and propose the data model that support it. The mainly discussed issues are 1) definition of the relationship between tables using the pointer, 2) navigation of the data using the relationship, 3) support of the referential integrity for pointer, 4) support of the uniform processing time for the join, 5) support of the object-oriented concepts, and 6) sharing of an index on multi-tables. We discuss the pointer-based data model that designed to include these issues to efficiently support complication environments.

Study of the semi-segregation algorithms of the incompressible Navier-Stokes equations using P2P1 finite element formulation (P2P1 유한요소 공식을 이용한 비압축성 Navier-Stokes 방정식의 반-분리 해법에 관한 연구)

  • Cho, Myung-H.;Choi, Hyoung-G.;Yoo, Jung-Y.;Park, Jae-I.
    • 유체기계공업학회:학술대회논문집
    • /
    • 2006.08a
    • /
    • pp.349-352
    • /
    • 2006
  • The conventional segregated finite element formulation produces a small and simple matrix at each step than in an integrated formulation. And the memory and cost requirements of computations are significantly reduced because the pressure equation for the mass conservation of the Navier-Stokes equations is constructed only once if the mesh is fixed. However, segregated finite element formulation solves Poisson equation of elliptic type so that it always needs a pressure boundary condition along a boundary even when physical information on pressure is not provided. On the other hand, the conventional integrated finite element formulation in which the governing equations are simultaneously treated has an advantage over a segregated formulation in the sense that it can give a more robust convergence behavior because all variables are implicitly combined. Further it needs a very small number of iterations to achieve convergence. However, the saddle-paint-type matrix (SPTM) in the integrated formulation is assembled and preconditioned every time step, so that it needs a large memory and computing time. Therefore, we newly proposed the P2PI semi-segregation formulation. In order to utilize the fact that the pressure equation is assembled and preconditioned only once in the segregated finite element formulation, a fixed symmetric SPTM has been obtained for the continuity constraint of the present semi-segregation finite element formulation. The momentum equation in the semi-segregation finite element formulation will be separated from the continuity equation so that the saddle-point-type matrix is assembled and preconditioned only once during the whole computation as long as the mesh does not change. For a comparison of the CPU time, accuracy and condition number between the two methods, they have been applied to the well-known benchmark problem. It is shown that the newly proposed semi-segregation finite element formulation performs better than the conventional integrated finite element formulation in terms of the computation time.

  • PDF

ABox Realization Reasoning in Distributed In-Memory System (분산 메모리 환경에서의 ABox 실체화 추론)

  • Lee, Wan-Gon;Park, Young-Tack
    • Journal of KIISE
    • /
    • v.42 no.7
    • /
    • pp.852-859
    • /
    • 2015
  • As the amount of knowledge information significantly increases, a lot of progress has been made in the studies focusing on how to reason large scale ontology effectively at the level of RDFS or OWL. These reasoning methods are divided into TBox classifications and ABox realizations. A TBox classification mainly deals with integrity and dependencies in schema, whereas an ABox realization mainly handles a variety of issues in instances. Therefore, the ABox realization is very important in practical applications. In this paper, we propose a realization method for analyzing the constraint of the specified class, so that the reasoning system automatically infers the classes to which instances belong. Unlike conventional methods that take advantage of the object oriented language based distributed file system, we propose a large scale ontology reasoning method using spark, which is a functional programming-based in-memory system. To verify the effectiveness of the proposed method, we used instances created from the Wine ontology by W3C(120 to 600 million triples). The proposed system processed the largest 600 million triples and generated 951 million triples in 51 minutes (696 K triple / sec) in our largest experiment.

A Low-Power Texture Mapping Technique for Mobile 3D Graphics (모바일 3D 그래픽스를 위한 저전력 텍스쳐 맵핑 기법)

  • Kim, Hyun-Hee;Kim, Ji-Hong
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.2
    • /
    • pp.45-57
    • /
    • 2009
  • ETexture mapping is a technique used for adding reality to an image in 3D graphics. However. this technique becomes the bottleneck of the 3D graphics pipeline because it requires large processing power and high memory bandwidth. For reducing memory latency in texture mapping, texture cache is used. As portable devices become smaller and they have power constraint, it is important to reduce the area and the power consumption of the texture cache. In this paper we propose using a small texture cache to reduce the area and the power consumption of the texture cache. Furthermore, we propose techniques to keep a performance comparable to large texture caches by using prefetch techniques and a victim cache. Simulation results show the proposed small texture cache can reduce the area and the power consumption up to 70% and 60%, respectively, by using $1{\sim}2K$ bytes texture cache compared to the conventional 16K bytes cache while keeping the performance.

Memory Organization for a Fuzzy Controller.

  • Jee, K.D.S.;Poluzzi, R.;Russo, B.
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 1993.06a
    • /
    • pp.1041-1043
    • /
    • 1993
  • Fuzzy logic based Control Theory has gained much interest in the industrial world, thanks to its ability to formalize and solve in a very natural way many problems that are very difficult to quantify at an analytical level. This paper shows a solution for treating membership function inside hardware circuits. The proposed hardware structure optimizes the memoried size by using particular form of the vectorial representation. The process of memorizing fuzzy sets, i.e. their membership function, has always been one of the more problematic issues for the hardware implementation, due to the quite large memory space that is needed. To simplify such an implementation, it is commonly [1,2,8,9,10,11] used to limit the membership functions either to those having triangular or trapezoidal shape, or pre-definite shape. These kinds of functions are able to cover a large spectrum of applications with a limited usage of memory, since they can be memorized by specifying very few parameters ( ight, base, critical points, etc.). This however results in a loss of computational power due to computation on the medium points. A solution to this problem is obtained by discretizing the universe of discourse U, i.e. by fixing a finite number of points and memorizing the value of the membership functions on such points [3,10,14,15]. Such a solution provides a satisfying computational speed, a very high precision of definitions and gives the users the opportunity to choose membership functions of any shape. However, a significant memory waste can as well be registered. It is indeed possible that for each of the given fuzzy sets many elements of the universe of discourse have a membership value equal to zero. It has also been noticed that almost in all cases common points among fuzzy sets, i.e. points with non null membership values are very few. More specifically, in many applications, for each element u of U, there exists at most three fuzzy sets for which the membership value is ot null [3,5,6,7,12,13]. Our proposal is based on such hypotheses. Moreover, we use a technique that even though it does not restrict the shapes of membership functions, it reduces strongly the computational time for the membership values and optimizes the function memorization. In figure 1 it is represented a term set whose characteristics are common for fuzzy controllers and to which we will refer in the following. The above term set has a universe of discourse with 128 elements (so to have a good resolution), 8 fuzzy sets that describe the term set, 32 levels of discretization for the membership values. Clearly, the number of bits necessary for the given specifications are 5 for 32 truth levels, 3 for 8 membership functions and 7 for 128 levels of resolution. The memory depth is given by the dimension of the universe of the discourse (128 in our case) and it will be represented by the memory rows. The length of a world of memory is defined by: Length = nem (dm(m)+dm(fm) Where: fm is the maximum number of non null values in every element of the universe of the discourse, dm(m) is the dimension of the values of the membership function m, dm(fm) is the dimension of the word to represent the index of the highest membership function. In our case then Length=24. The memory dimension is therefore 128*24 bits. If we had chosen to memorize all values of the membership functions we would have needed to memorize on each memory row the membership value of each element. Fuzzy sets word dimension is 8*5 bits. Therefore, the dimension of the memory would have been 128*40 bits. Coherently with our hypothesis, in fig. 1 each element of universe of the discourse has a non null membership value on at most three fuzzy sets. Focusing on the elements 32,64,96 of the universe of discourse, they will be memorized as follows: The computation of the rule weights is done by comparing those bits that represent the index of the membership function, with the word of the program memor . The output bus of the Program Memory (μCOD), is given as input a comparator (Combinatory Net). If the index is equal to the bus value then one of the non null weight derives from the rule and it is produced as output, otherwise the output is zero (fig. 2). It is clear, that the memory dimension of the antecedent is in this way reduced since only non null values are memorized. Moreover, the time performance of the system is equivalent to the performance of a system using vectorial memorization of all weights. The dimensioning of the word is influenced by some parameters of the input variable. The most important parameter is the maximum number membership functions (nfm) having a non null value in each element of the universe of discourse. From our study in the field of fuzzy system, we see that typically nfm 3 and there are at most 16 membership function. At any rate, such a value can be increased up to the physical dimensional limit of the antecedent memory. A less important role n the optimization process of the word dimension is played by the number of membership functions defined for each linguistic term. The table below shows the request word dimension as a function of such parameters and compares our proposed method with the method of vectorial memorization[10]. Summing up, the characteristics of our method are: Users are not restricted to membership functions with specific shapes. The number of the fuzzy sets and the resolution of the vertical axis have a very small influence in increasing memory space. Weight computations are done by combinatorial network and therefore the time performance of the system is equivalent to the one of the vectorial method. The number of non null membership values on any element of the universe of discourse is limited. Such a constraint is usually non very restrictive since many controllers obtain a good precision with only three non null weights. The method here briefly described has been adopted by our group in the design of an optimized version of the coprocessor described in [10].

  • PDF

Mining Frequent Sequential Patterns over Sequence Data Streams with a Gap-Constraint (순차 데이터 스트림에서 발생 간격 제한 조건을 활용한 빈발 순차 패턴 탐색)

  • Chang, Joong-Hyuk
    • Journal of the Korea Society of Computer and Information
    • /
    • v.15 no.9
    • /
    • pp.35-46
    • /
    • 2010
  • Sequential pattern mining is one of the essential data mining tasks, and it is widely used to analyze data generated in various application fields such as web-based applications, E-commerce, bioinformatics, and USN environments. Recently data generated in the application fields has been taking the form of continuous data streams rather than finite stored data sets. Considering the changes in the form of data, many researches have been actively performed to efficiently find sequential patterns over data streams. However, conventional researches focus on reducing processing time and memory usage in mining sequential patterns over a target data stream, so that a research on mining more interesting and useful sequential patterns that efficiently reflect the characteristics of the data stream has been attracting no attention. This paper proposes a mining method of sequential patterns over data streams with a gap constraint, which can help to find more interesting sequential patterns over the data streams. First, meanings of the gap for a sequential pattern and gap-constrained sequential patterns are defined, and subsequently a mining method for finding gap-constrained sequential patterns over a data stream is proposed.