• Title/Summary/Keyword: Information processing

Search Result 42,378, Processing Time 0.066 seconds

Design and Implementation of JBI Component Deployment tool Based on the open sources (오픈소스 기반의 JBI 컴포넌트배치도구의 설계 및 구현)

  • Park, Jae-Won;Choi, Jae-Hyun;Lee, Jae-Sung;Lee, Nam-Yong
    • The KIPS Transactions:PartD
    • /
    • v.16D no.5
    • /
    • pp.715-728
    • /
    • 2009
  • The Enterprise Service Bus based on Java Business Integration is an web service standard and one of the methods for implementing distribution channels of Service Oriented Architecture. Consisting of open source group, extensive venders and users, the ESB based JBI has the problems of ineffectiveness as well as advantages of extensibility of service plug-in. That is, in case users need to use Service plug-in, manual connection of packaging process and sequential distribution method is required. This study, therefore, proposes as a way of trouble-shooting the user-oriented component deployment tool which can manage entire process for deploying The ESB middleware platform to Service unit. At the same time, this study elicited the requirements based on issues of JBI-based ESB and has developed the modeling property, packaging, distribution and evaluation thru Schema analysis of JBI-compatible component. Using the deployment tool this study proposed, users will be able to perform and manage the whole deploying process without additional manual work for connecting component. Not only it is expected that interface based on Graphic User Interface provide usability and convenience but they can also minimize the errors rate through component and route validity verification function provided in deployment tool.

Weight Based Technique For Improvement Of New User Recommendation Performance (신규 사용자 추천 성능 향상을 위한 가중치 기반 기법)

  • Cho, Sun-Hoon;Lee, Moo-Hun;Kim, Jeong-Seok;Kim, Bong-Hoi;Choi, Eui-In
    • The KIPS Transactions:PartD
    • /
    • v.16D no.2
    • /
    • pp.273-280
    • /
    • 2009
  • Today, many services and products that used to be only provided on offline have been being provided on the web according to the improvement of computing environment and the activation of web usage. These web-based services and products tend to be provided to customer by customer's preferences. This paradigm that considers customer's opinions and features in selecting is called personalization. The related research field is a recommendation. And this recommendation is performed by recommender system. Generally the recommendation is made from the preferences and tastes of customers. And recommender system provides this recommendation to user. However, the recommendation techniques have a couple of problems; they do not provide suitable recommendation to new users and also are limited to computing space that they generate recommendations which is dependent on ratings of products by users. Those problems has gathered some continuous interest from the recommendation field. In the case of new users, so similar users can't be classified because in the case of new users there is no rating created by new users. The problem of the limitation of the recommendation space is not easy to access because it is related to moneywise that the cost will be increasing rapidly when there is an addition to the dimension of recommendation. Therefore, I propose the solution of the recommendation problem of new user and the usage of item quality as weight to improve the accuracy of recommendation in this paper.

A CNV detection algorithm based on statistical analysis of the aligned reads (정렬된 리드의 통계적 분석을 기반으로 하는 CNV 검색 알고리즘)

  • Hong, Sang-Kyoon;Hong, Dong-Wan;Yoon, Jee-Hee;Kim, Baek-Sop;Park, Sang-Hyun
    • The KIPS Transactions:PartD
    • /
    • v.16D no.5
    • /
    • pp.661-672
    • /
    • 2009
  • Recently it was found that various genetic structural variations such as CNV(copy number variation) exist in the human genome, and these variations are closely related with disease susceptibility, reaction to treatment, and genetic characteristics. In this paper we propose a new CNV detection algorithm using millions of short DNA sequences generated by giga-sequencing technology. Our method maps the DNA sequences onto the reference sequence, and obtains the occurrence frequency of each read in the reference sequence. And then it detects the statistically significant regions which are longer than 1Kbp as the candidate CNV regions by analyzing the distribution of the occurrence frequency. To select a proper read alignment method, several methods are employed in our algorithm, and the performances are compared. To verify the superiority of our approach, we performed extensive experiments. The result of simulation experiments (using a reference sequence, build 35 of NCBI) revealed that our approach successfully finds all the CNV regions that have various shapes and arbitrary length (small, intermediate, or large size).

A Requirement Priority Process of Embedded Systems based on the Dependency and Aspect (의존과 관점 기반 임베디드 시스템의 요구사항 우선순위 프로세스)

  • Hwang, Wi-Yong;Kang, Dong-Su;Song, Chee-Yang;Seong, Jae-Seok;Baik, Doo-Kwon
    • The KIPS Transactions:PartD
    • /
    • v.16D no.5
    • /
    • pp.767-790
    • /
    • 2009
  • Setting up a priorityfor an embedded system is greatly significant because a release plan at the early stage of product developments can properly be established through right decision making procedures based on the priorities. For instance, both dependencies among requirements and the aspects of product developers should be considered into the priorities to improve the embedded system. Especially, trade-offs among the requirements, which are quite different depending on H/W and S/W architecture styles they use, should be acknowledged without exception. However, the selection process on the priority has hitherto been fairly systematic in the existing environment where hardware and software are not being considered at once. Therefore, this paper suggests an dependency and aspect-based model and process for the requirements of the priority. For this, the paper analyzes the trade-offs between the requirements depending on the disparate Architecture styles of H/W and S/W, and it also reflects the viewpoints of the developers. For thelast thing, the model and process suggested will be applied to the case of the development of both cell phones and cameras to gain authenticity and reliability. In conclusion, the danger occurring when the release plan is constructed can be minimized by screening the priorities that optimizes the embedded system more explicitly.

A Performance Improvement of Linux TCP Networking by Data Structure Reuse (자료 구조 재사용을 이용한 리눅스 TCP 네트워킹 성능 개선)

  • Kim, Seokkoo;Chung, Kyusik
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.3 no.8
    • /
    • pp.261-270
    • /
    • 2014
  • As Internet traffic increases recently, much effort has been put on improving the performance of a web server. In addition to hardware side solutions such as replacement by high-end hardware or expansion of the number of servers, there are software side solutions to improve performance. Recent studies on these software side solutions have been actively performed. In this paper, we identify performance degradation problems occurring in a conventional TCP networking reception process and propose a way to solve them. We improve performance by combining three kinds of existing methods for Linux Networking Performance Improvement and two kinds of newly proposed methods in this paper. The three existing methods include 1) an allocation method of a packet flow to a core in a multi-core environment, 2) ITR(Interrupt Throttle Rate) method to control excessive interrupt requests, and 3) sk_buff data structure recycling. The two newly proposed methods are fd data structure recycling and epoll_event data structure recycling. Through experiments in a web server environment, we verify the effect of our two proposed methods and its combination with the three existing methods for performance improvement, respectively. We use three kinds of web servers: a simple web server, Lighttpd generally used in Linux, and Apache. In a simple web server environment, fd data structure recycling and epoll_event data structure recycling bring out performance improvement by about 7 % and 6%, respectively. If they are combined with the three existing methods, performance is improved by up to 40% in total. In a Lighttpd and an Apache web server environment, the combination of five methods brings out performance improvement by up to 36% and 20% in total, respectively.

Development of a Simulation Prediction System Using Statistical Machine Learning Techniques (통계적 기계학습 기술을 이용한 시뮬레이션 결과 예측 시스템 개발)

  • Lee, Ki Yong;Shin, YoonJae;Choe, YeonJeong;Kim, SeonJeong;Suh, Young-Kyoon;Sa, Jeong Hwan;Lee, JongSuk Luth;Cho, Kum Won
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.11
    • /
    • pp.593-606
    • /
    • 2016
  • Computer simulation is widely used in a variety of computational science and engineering fields, including computational fluid dynamics, nano physics, computational chemistry, structural dynamics, and computer-aided optimal design, to simulate the behavior of a system. As the demand for the accuracy and complexity of the simulation grows, however, the cost of executing the simulation is rapidly increasing. It, therefore, is very important to lower the total execution time of the simulation especially when that simulation makes a huge number of repetitions with varying values of input parameters. In this paper we develop a simulation service system that provides the ability to predict the result of the requested simulation without actual execution for that simulation: by recording and then returning previously obtained or predicted results of that simulation. To achieve the goal of avoiding repetitive simulation, the system provides two main functionalities: (1) storing simulation-result records into database and (2) predicting from the database the result of a requested simulation using statistical machine learning techniques. In our experiments we evaluate the prediction performance of the system using real airfoil simulation result data. Our system on average showed a very low error rate at a minimum of 0.9% for a certain output variable. Using the system any user can receive the predicted outcome of her simulation promptly without actually running it, which would otherwise impose a heavy burden on computing and storage resources.

Energy-Efficient Multipath Routing Protocol for Supporting Mobile Events in Wireless Sensor Networks (무선 센서 네트워크에서 이동 이벤트를 지원하기 위한 에너지 효율적인 멀티패스 라우팅 프로토콜)

  • Kim, Hoewon;Lee, Euisin
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.5 no.12
    • /
    • pp.455-462
    • /
    • 2016
  • Wireless sensor networks have been researched to gather data about events on sensor fields from sources at sinks. Multipath routing is one of attractive approaches to reliably send data against the problem of frequent breakages on paths from sources to sinks due to node and link failures. As mobile events such as humans, animals, and vehicles are considered, sources may be continuously generated according to the movement of the mobile event. Thus, mobile events provide new challenging issue in multipath routing. However, the research on multipath routing mainly focus on both efficient multipath construction from sources to static sinks and fast multipath reconstruction against path breakages. Accordingly, the previous multipath routing protocols request each source continuously generated by a mobile event to construct individual multipath from the source to sinks. This induces the increase of multipath construction cost in the previous protocols in proportion to the number of source. Therefore, we propose efficient multipath routing protocol for supporting continuous sources generated by mobile events. In the proposed protocol, new source efficiently reconstructs its multipath by exploiting the existing multipath of previous sources. To do this, the proposed protocol selects one among three reconstruction methods: a local reconstruction, a global partial one, and a global full one. For a selection decision, we provide an analytical energy consumption cost model that calculates the summation of both the multipath reconstruction cost and the data forwarding cost. Simulation results show that the proposed protocol has better performance than the previous protocol to provide multipath routing for mobile events.

A Study to Improve Recovery Ratio of Deleted File Using the Parsing Algorithm of the HFS + Journal File (HFS+ 저널 파일 파싱 알고리즘을 이용한 삭제된 파일 복구 기법 향상 방안)

  • Bang, Seung Gyu;Jeon, Sang Jun;Kim, Do Hyun;Lee, Sang Jin
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.5 no.12
    • /
    • pp.463-470
    • /
    • 2016
  • With the growing demand for MAC-based system, the need for digital forensic techniques of these system has been increasing. In the digital forensic analysis process, sometimes analysts have recovered the deleted files when they prove the allegations if system user try to remove the evidence deliberately. Research and analysis that recover the deleted files from a file system constantly been made and HFS+ that is a file system of MAC-based system also has been researched. Carving techniques primarily has been used to recover the deleted file from HFS+ a file system because metadata of folder or file overwrite metadata of a deleted file when file is deleted from a file system on HFS+ characteristic. But if the file content is saved by separated state in a file system, Carving techniques also can't recover the whole or a part of the deleted file. In this paper we describe technique the deleted file recovery technique using HFS+ file system a journal. This technique that is suggested by existing research and analysis result is the technique that recover the deleted file by metadata that is maintained in a journal on HFS+ file system. but this technique excludes specific files and this problem needs to be reformed. In this paper we suggest algorithm that analysis a journal of HFS+ file system in detail. And we demonstrate that the deleted file cat be recovered from the extracted metadata by this algorithm without the excluded file.

Energy and Delay-Efficient Multipath Routing Protocol for Supporting Mobile Sink in Wireless Sensor Networks (무선 센서 네트워크에서 이동 싱크를 지원하기 위한 다중 경로 라우팅 프로토콜)

  • Lee, Hyun Kyu;Lee, Euisin
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.5 no.12
    • /
    • pp.447-454
    • /
    • 2016
  • The research on multipath routing has been studied to solve the problem of frequent path breakages due to node and link failures and to enhance data delivery reliability in wireless sensor networks. In the multipath routing, mobile sinks such as soldiers in battle fields and rescuers in disaster areas bring about new challenge for handling their mobility. The sink mobility requests new multipath construction from sources to mobile sinks according to their movement path. Since mobile sinks have continuous mobility, the existing multipath can be exploited to efficiently reconstruct to new positions of mobile sinks. However, the previous protocols do not address this issue. Thus, we proposed an efficient multipath reconstruction protocol called LGMR for mobile sinks in wireless sensor networks. The LGMR address three multipath reconstruction methods based on movement types of mobile sinks: a single hop movement-based local multipath reconstruction, a multiple hop movement-based local multipath reconstruction, and a multiple hop movement-based global multipath reconstruction. Simulation results showed that the LGMR has better performance than the previous protocol in terms of energy consumption and data delivery delay.

Direct Pass-Through based GPU Virtualization for Biologic Applications (바이오 응용을 위한 직접 통로 기반의 GPU 가상화)

  • Choi, Dong Hoon;Jo, Heeseung;Lee, Myungho
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.2 no.2
    • /
    • pp.113-118
    • /
    • 2013
  • The current GPU virtualization techniques incur large overheads when executing application programs mainly due to the fine-grain time-sharing scheduling of the GPU among multiple Virtual Machines (VMs). Besides, the current techniques lack of portability, because they include the APIs for the GPU computations in the VM monitor. In this paper, we propose a low overhead and high performance GPU virtualization approach on a heterogeneous HPC system based on the open-source Xen. Our proposed techniques are tailored to the bio applications. In our virtualization framework, we allow a VM to solely occupy a GPU once the VM is assigned a GPU instead of relying on the time-sharing the GPU. This improves the performance of the applications and the utilization of the GPUs. Our techniques also allow a direct pass-through to the GPU by using the IOMMU virtualization features embedded in the hardware for the high portability. Experimental studies using microbiology genome analysis applications show that our proposed techniques based on the direct pass-through significantly reduce the overheads compared with the previous Domain0 based approaches. Furthermore, our approach closely matches the performance for the applications to the bare machine or rather improves the performance.