• Title/Summary/Keyword: Information processing

Search Result 42,378, Processing Time 0.07 seconds

Time Synchronization Algorithm using the Clock Drift Rate and Reference Signals Between Two Sensor Nodes (클럭 표류율과 기준 신호를 이용한 두 센서 노드간 시간 동기 알고리즘)

  • Kim, Hyoun-Soo;Jeon, Joong-Nam
    • The KIPS Transactions:PartC
    • /
    • v.16C no.1
    • /
    • pp.51-56
    • /
    • 2009
  • Time synchronization algorithm in wireless sensor networks is essential to various applications such as object tracking, data encryption, duplicate detection, and precise TDMA scheduling. This paper describes CDRS that is a time synchronization algorithm using the Clock Drift rate and Reference Signals between two sensor nodes. CDRS is composed of two steps. At first step, the time correction is calculated using offset and the clock drift rate between the two nodes based on the LTS method. Two nodes become a synchronized state and the time variance can be compensated by the clock drift rate. At second step, the synchronization node transmits reference signals periodically. This reference signals are used to calculate the time difference between nodes. When this value exceeds the maximum error tolerance, the first step is performed again for resynchronization. The simulation results on the performance analysis show that the time accuracy of the proposed algorithm is improved, and the energy consumption is reduced 2.5 times compared to the time synchronization algorithm with only LTS, because CDRS reduces the number of message about 50% compared to LTS and reference signals do not use the data space for timestamp.

An Effective Service Sharing Scheme on Distributed OSGi Frameworks (분산되어 있는 OSGi 프레임워크에서 효과적인 서비스 공유 방안)

  • Yun, Ki-Hyun;Kim, Eun-Hoe;Choi, Jae-Young
    • The KIPS Transactions:PartA
    • /
    • v.16A no.4
    • /
    • pp.227-234
    • /
    • 2009
  • OSGi technology has applied to Home Network, but now it is loaded into even heterogeneous devices in various domains. Therefore, it is necessary to cooperate with each other framework for offering effective services in distributed OSGi frameworks. However until now, OSGi specification doesn't provide any methodthat can share the services registered on remote OSGi frameworks. In order to solve this problem, there have been several researches that used existing distributed middleware technologies such as JXTA and Web Services. However these technologies have some weakness, that is, they consume lots of computing resources and need additional process to transform the services. A middleware called R-OSGi uses only OSGi technology for sharing remote OSGi services, but R-OSGi may have a communication bottleneck and SPOF (Single Point of Failure) problem, because it has a central service registry. In this paper we present RSP (Remote Service Provider), which is a P2P-based effective service sharing scheme on distributed OSGi framework. RSP doesn't need to install additional software nor have the additional transformation process of the service representation, because it uses only OSGi technology. In addition it doesn't have any bottleneck problem and improves scalability by providing the service discovery mechanism using P2P. RSP can also access remote services transparently and it can guarantee reliability by sending an immediate notice about changes of the remote services.

Cache Invalidation Schemes based on Time Guarantee for Improving Access Time in Mobile Ad hoc Networks (모바일 애드혹 네트워크에서 캐쉬 접근 시간 향상을 위한 시간보증 기반의 캐쉬무효화 기법)

  • Choi, Jae-Ho;Oh, Jae-Oh;Lee, Myong-Soo;Lee, Sang-Keun
    • The KIPS Transactions:PartC
    • /
    • v.16C no.1
    • /
    • pp.65-72
    • /
    • 2009
  • Due to the popularity of mobile devices and advances in wireless communication technologies, a mobile ad hoc network has received a lot of attention. In the existing data replication management research, the use of a replica has been shown to be an efficient technique for improving data accessibility. However, to use a replica in ad hoc networks, the data consistency between the original data and the replica should be guaranteed. In the traditional research, a mobile node should check an original data whether the data is updated or not. However, It may be costly or sometimes impossible to check the original data. In the case of the time constraint applications, the checking cost can cause more serious problem. In this paper, we propose the time-guarantee based cache invalidation schemes for time constraint applications and the threshold based compensation method to enhance the time-guarantee based scheme. The proposed schemes can remove the "rollback" problem. Simulation results show that our schemes outperform the previous ones in terms of access time with little loss of data currency.

A Multi-Middleware Bridge for Dynamic Extensibility and Load Balancing in Home Network Environments (홈 네트워크 환경에서의 동적 확장성과 부하분산을 위한 다중 미들웨어 브리지)

  • Kim, Youn-Woo;Jang, Hyun-Su;Song, Chang-Hwan;Eom, Young-Ik
    • The KIPS Transactions:PartA
    • /
    • v.16A no.4
    • /
    • pp.263-272
    • /
    • 2009
  • For implementing the ubiquitous computing environments with smart home infrastructures, various research on the home network have been performed by several research institutes and companies. Due to the various home network middleware that are developed recently, the standardization of the home network middleware is being delayed and it calls for the middleware bridge which solves the interoperability problem among the heterogeneous middlewares. Now the research on the scheme for interoperability and the development of the various bridges are in progress, such as one-to-one bridge supporting interoperability between two middlewares and one-to-many bridge supporting interoperability among the multi-middlewares. However, existing systems and schemes does not consider the dynamic extensibility and performance that is particularly needed in the smart home environments. The middleware bridge should provide bridge extensibility with zero-configuration for non-expert users. It should also provide the load balancing scheme for efficient and proper traffic distribution. In this paper, we propose a Multi-Middleware Bridge(MMB) for dynamic extensibility and load balancing in home network environments. MMB provides bridge scalability and load balancing through the distributed system structure. We also verify the features such as interoperability, bridge extensibility, and the performance of the load balancing algorithm.

Performance Improvement of an Energy Efficient Cluster Management Based on Autonomous Learning (자율학습기반의 에너지 효율적인 클러스터 관리에서의 성능 개선)

  • Cho, Sungchul;Chung, Kyusik
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.4 no.11
    • /
    • pp.369-382
    • /
    • 2015
  • Energy aware server clusters aim to reduce power consumption at maximum while keeping QoS(quality of service) compared to energy non-aware server clusters. They adjust the power mode of each server in a fixed or variable time interval to activate only the minimum number of servers needed to handle current user requests. Previous studies on energy aware server cluster put efforts to reduce power consumption or heat dissipation, but they do not consider energy efficiency well. In this paper, we propose an energy efficient cluster management method to improve not only performance per watt but also QoS of the existing server power mode control method based on autonomous learning. Our proposed method is to adjust server power mode based on a hybrid approach of autonomous learning method with multi level thresholds and power consumption prediction method. Autonomous learning method with multi level thresholds is applied under normal load situation whereas power consumption prediction method is applied under abnormal load situation. The decision on whether current load is normal or abnormal depends on the ratio of the number of current user requests over the average number of user requests during recent past few minutes. Also, a dynamic shutdown method is additionally applied to shorten the time delay to make servers off. We performed experiments with a cluster of 16 servers using three different kinds of load patterns. The multi-threshold based learning method with prediction and dynamic shutdown shows the best result in terms of normalized QoS and performance per watt (valid responses). For banking load pattern, real load pattern, and virtual load pattern, the numbers of good response per watt in the proposed method increase by 1.66%, 2.9% and 3.84%, respectively, whereas QoS in the proposed method increase by 0.45%, 1.33% and 8.82%, respectively, compared to those in the existing autonomous learning method with single level threshold.

Derivation of Data Quality Attributes and their Priorities Based on Customer Requirements (고객의 요구사항에 기반한 데이터품질 평가속성 및 우선순위 도출)

  • Jang, Kyoung-Ae;Kim, Ja-Hee;Kim, Woo Je
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.4 no.12
    • /
    • pp.549-560
    • /
    • 2015
  • There is a wide variety of data quality attributes such as the ones proposed by the ISO/IEC organization and also by many other domestic and international institutions. However, it takes considerable time and costs to apply those criteria and guidelines to real environment. Therefore, it needs to define data quality evaluation attributes which are easily applicable and are not influenced by organizational environment limitations. The purpose of this paper is to derive data quality attributes and order of their priorities based on customer requirements for managing the process systematically and evaluating the data quantitatively. This study identifies the customer cognitive constructs of data quality attributes using the RGT(Repertory Grid Technique) based on a Korean quality standard model (DQC-M). Also the correlation analysis on the identified constructs is conducted, and the evaluation attributes is prioritized and ranked using the AHP. As the results of this paper, the consistent system, the accurate data, the efficient environment, the flexible management, and the continuous improvement are derived at the first level of the data quality evaluation attributes. Also, Control Compliance(13%), Regulatory Compliance(10%), Requirement Completeness(9.6%), Accuracy(8.4%), and Traceability(6.8%) are ranked on the top 5 of the 19 attributes in the second level.

Log-Structured B-Tree for NAND Flash Memory (NAND 플래시 메모리를 위한 로그 기반의 B-트리)

  • Kim, Bo-Kyeong;Joo, Young-Do;Lee, Dong-Ho
    • The KIPS Transactions:PartD
    • /
    • v.15D no.6
    • /
    • pp.755-766
    • /
    • 2008
  • Recently, NAND flash memory is becoming into the spotlight as a next-generation storage device because of its small size, fast speed, low power consumption, and etc. compared to the hard disk. However, due to the distinct characteristics such as erase-before-write architecture, asymmetric operation speed and unit, disk-based systems and applications may result in severe performance degradation when directly implementing them on NAND flash memory. Especially when a B-tree is implemented on NAND flash memory, intensive overwrite operations may be caused by record inserting, deleting, and reorganizing. These may result in severe performance degradation. Although ${\mu}$-tree has been proposed in order to overcome this problem, it suffers from frequent node split and rapid increment of its height. In this paper, we propose Log-Structured B-Tree(LSB-Tree) where the corresponding log node to a leaf node is allocated for update operation and then the modified data in the log node is stored at only one write operation. LSB-tree reduces additional write operations by deferring the change of parent nodes. Also, it reduces the write operation by switching a log node to a new leaf node when inserting the data sequentially by the key order. Finally, we show that LSB-tree yields a better performance on NAND flash memory by comparing it to ${\mu}$-tree through various experiments.

Generating Rank-Comparison Decision Rules with Variable Number of Genes for Cancer Classification (순위 비교를 기반으로 하는 다양한 유전자 개수로 이루어진 암 분류 결정 규칙의 생성)

  • Yoon, Young-Mi;Bien, Sang-Jay;Park, Sang-Hyun
    • The KIPS Transactions:PartD
    • /
    • v.15D no.6
    • /
    • pp.767-776
    • /
    • 2008
  • Microarray technology is extensively being used in experimental molecular biology field. Microarray experiments generate quantitative expression measurements for thousands of genes simultaneously, which is useful for the phenotype classification of many diseases. One of the two major problems in microarray data classification is that the number of genes exceeds the number of tissue samples. The other problem is that current methods generate classifiers that are accurate but difficult to interpret. Our paper addresses these two problems. We performed a direct integration of individual microarrays with same biological objectives by transforming an expression value into a rank value within a sample and generated rank-comparison decision rules with variable number of genes for cancer classification. Our classifier is an ensemble method which has k top scoring decision rules. Each rule contains a number of genes, a relationship among involved genes, and a class label. Current classifiers which are also ensemble methods consist of k top scoring decision rules. However these classifiers fix the number of genes in each rule as a pair or a triple. In this paper we generalized the number of genes involved in each rule. The number of genes in each rule is in the range of 2 to N respectively. Generalizing the number of genes increases the robustness and the reliability of the classifier for the class prediction of an independent sample. Also our classifier is readily interpretable, accurate with small number of genes, and shed a possibility of the use in a clinical setting.

A Study on Selection Process of Open Source Software (오픈소스 소프트웨어의 선정 절차에 관한 연구)

  • Lee, Young-Min;Rhew, Sung-Yul;Kim, Jong-Bae
    • The KIPS Transactions:PartD
    • /
    • v.15D no.6
    • /
    • pp.793-802
    • /
    • 2008
  • The development methods which utilize OSSs have been tried as new alternative to solve limits of the previous software developments such as the quality of software, and time and cost of developments. Especially, small and medium companies are suffering from difficulty in applying large-scale development methodology whenever they develop softwares. Therefore, in the current situation that demand for small-scale development methodology is increasing, the methods of utilizing OSSs can become an efficient way to save costs and reduce a development period. Accordingly, analysis of open-source is no fixed procedure or method to utilize open-source for software developments in the field. In this thesis, to solve such problems, we propose the procedures and methods for identifying and selecting suitable open-source, and effective methods for improvement and integration through least modification on the basis of synthesis of existing researches and experiences in development projects. For selection of OSSs, we did identify correct requirements for the software to be developed, investigate the open-source just matching with such requirements, draw a candidate index, establish assessment criteria, and the consequently present a method to select OSSs. And, we deduced the validity and improvement of each detailed activity from practical application to an actual project and assessment.

An Adjustable Process of Requirements Prioritizing for Release Plan (릴리즈 플랜의 적응적 요구사항 우선순위 프로세스)

  • Seong, Jae-Seok;Kang, Dong-Su;Song, Chee-Yang;Baik, Doo-Kwon
    • The KIPS Transactions:PartD
    • /
    • v.15D no.6
    • /
    • pp.841-856
    • /
    • 2008
  • The priority of requirement is important because the priority is a critical activity of release plan especially in software development which has an open market customer. Also, it is important for stakeholders to select a method and aspects to prioritize requirements. The selection is based on the organizational experience of a priority process, the process model of the product, goals and a type of products, and dependencies between requirements. But, the current researches considered only static dependency between requirements and did not suggest a systematic priority process. In addition, the current researches only suggest limited aspects to prioritize requirements, such as cost and value. Therefore, this paper proposes an adjustable priority process based on a priority model to select a method and aspects for the suitable priority for product and organization. Especially, this paper enhances the completeness of a release plan by a definition of static and dynamic dependency types between requirements. This paper suggests a priority model, which considers the dependencies between requirement and various viewpoint of software development. Based on the priority model, the paper suggests a systematic priority process to promote reasonable decisions to the priority and release plan of requirement.