• Title/Summary/Keyword: Web Server Cluster System

Search Result 39, Processing Time 0.027 seconds

Autonomic Self Healing-Based Load Assessment for Load Division in OKKAM Backbone Cluster

  • Chaudhry, Junaid Ahsenali
    • Journal of Information Processing Systems
    • /
    • v.5 no.2
    • /
    • pp.69-76
    • /
    • 2009
  • Self healing systems are considered as cognation-enabled sub form of fault tolerance system. But our experiments that we report in this paper show that self healing systems can be used for performance optimization, configuration management, access control management and bunch of other functions. The exponential complexity that results from interaction between autonomic systems and users (software and human users) has hindered the deployment and user of intelligent systems for a while now. We show that if that exceptional complexity is converted into self-growing knowledge (policies in our case), can make up for initial development cost of building an intelligent system. In this paper, we report the application of AHSEN (Autonomic Healing-based Self management Engine) to in OKKAM Project infrastructure backbone cluster that mimics the web service based architecture of u-Zone gateway infrastructure. The 'blind' load division on per-request bases is not optimal for distributed and performance hungry infrastructure such as OKKAM. The approach adopted assesses the active threads on the virtual machine and does resource estimates for active processes. The availability of a certain server is represented through worker modules at load server. Our simulation results on the OKKAM infrastructure show that the self healing significantly improves the performance and clearly demarcates the logical ambiguities in contemporary designs of self healing infrastructures proposed for large scale computing infrastructures.

Analysis of Transfer Characteristics and Time-delay of Control System based on Clustering Web Server (클러스터 웹서버 제어시스템의 Time-delay 및 전달 특성 분석)

  • Nahm, Eui-Seok
    • Journal of Digital Convergence
    • /
    • v.12 no.8
    • /
    • pp.219-227
    • /
    • 2014
  • Ethernet, ATM, and CAN are wide-utilized communication protocols for information transfer by internet. Many researches about Network Time-delay have been based on network modeling. But almost of them have not shown an optimal solution in various communication environments. So, asynchronous sample system modeling based internet is needed to be robust in various network environments. Also as closed loop system in internet has a different operational characteristics and noise characteristics comparing with conventional control system, new robust control method is needed in instruments which demand to be safe and precise for internet environments. In order to achieve the safe and precise real-time control in remote plant, this paper is aimed to analysis the transfer characteristics and time-delay of control system in cluster web server.

A Customized Cancer Radiation Treatment Planning Simulation (ccRTPs) System via Web and Network (웹과 네트워크 기술을 이용한 환자 맞춤식 암치료 계획 시뮬레이션 시스템)

  • Khm, O-Yeon
    • Progress in Medical Physics
    • /
    • v.17 no.3
    • /
    • pp.144-152
    • /
    • 2006
  • The telemedicine using independent client-server system via networks can provide high quality normalized services to many hospitals, specifically to local/rural area hospitals. This will eventually lead to a decreased medical cost because the centralized institute can handle big computer hardware systems and complicated software systems efficiently and economically, Customized cancer radiation treatment planning for each patient Is very useful for both a patient and a doctor because it makes possible for the most effective treatment with the least possible dose to patient. Radiation planners know that too small a dose to the tumor can result in recurrence of the cancer, while too large a dose to healthy tissue can cause complications or even death. The best solution is to build an accurate planning simulation system to provide better treatment strategies based on each patient's computerized tomography (CT) image. We are developing a web-based and a network-based customized cancer radiation therapy simulation system consisting of four Important computer codes; a CT managing code for preparing the patients target data from their CT image files, a parallel Monte Carlo high-energy beam code (PMCEPT code) for calculating doses against the target generated from the patient CT image, a parallel linear programming code for optimizing the treatment plan, and scientific data visualization code for efficient pre/post evaluation of the results. The whole softwares will run on a high performance Beowulf PC cluster of about 100-200 CPUs. Efficient management of the hardware and software systems is not an easy task for a hospital. Therefore, we integrated our system into the client-sewer system via network or web and provide high quality normalized services to many hospitals. Seamless communication with doctors is maintained via messenger function of the server-client system.

  • PDF

A Hashing Scheme using Round Robin in a Wireless Internet Proxy Server Cluster System (무선 인터넷 프록시 서버 클러스터 시스템에서 라운드 로빈을 이용한 해싱 기법)

  • Kwak, Huk-Eun;Chung, Kyu-Sik
    • The KIPS Transactions:PartA
    • /
    • v.13A no.7 s.104
    • /
    • pp.615-622
    • /
    • 2006
  • Caching in a Wireless Internet Proxy Server Cluster Environment has an effect that minimizes the time on the request and response of Internet traffic and Web user As a way to increase the hit ratio of cache, we can use a hash function to make the same request URLs to be assigned to the same cache server. The disadvantage of the hashing scheme is that client requests cannot be well-distributed to all cache servers so that the performance of the whole system can depend on only a few busy servers. In this paper, we propose an improved load balancing scheme using hashing and Round Robin scheme that distributes client requests evenly to cache servers. In the existing hashing scheme, if a hashing value for a request URL is calculated, the server number is statically fixed at compile time while in the proposed scheme it is dynamically fixed at run time using round robin method. We implemented the proposed scheme in a Wireless Internet Proxy Server Cluster Environment and performed experiments using 16 PCs. Experimental results show the even distribution of client requests and the 52% to 112% performance improvement compared to the existing hashing method.

CERES: A Log-based, Interactive Web Analytics System for Backbone Networks (CERES: 백본망 로그 기반 대화형 웹 분석 시스템)

  • Suh, Ilhyun;Chung, Yon Dohn
    • KIISE Transactions on Computing Practices
    • /
    • v.21 no.10
    • /
    • pp.651-657
    • /
    • 2015
  • The amount of web traffic has increased as a result of the rapid growth of the use of web-based applications. In order to obtain valuable information from web logs, we need to develop systems that can support interactive, flexible, and efficient ways to analyze and handle large amounts of data. In this paper, we present CERES, a log-based, interactive web analytics system for backbone networks. Since CERES focuses on analyzing web log records generated from backbone networks, it is possible to perform a web analysis from the perspective of a network. CERES is designed for deployment in a server cluster using the Hadoop Distributed File System (HDFS) as the underlying storage. We transform and store web log records from backbone networks into relations and then allow users to use a SQL-like language to analyze web log records in a flexible and interactive manner. In particular, we use the data cube technique to enable the efficient statistical analysis of web log. The system provides users a web-based, multi-modal user interface.

A high speed processing method of web server cluster through round robin load balancing (라운드로빈 부하균형을 통한 웹 서버 클러스터 고속화 처리기법)

  • Sung Kyung;Kim Seok-soo;Park Gil-cheol
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.8 no.7
    • /
    • pp.1524-1531
    • /
    • 2004
  • This study analyzes a load balancing technique using Round Robin Algorithm. The study uses two software packages (Packet Capture and Round Robin Test Package) to check packet volume from Virtual Network structure (data generator, virtual server, Server1,2,3), and finds out traffic distribution toward Server1,2 and 3. The functions of implemented Round Robin Load Balancing Monitoring System include round robin testing, system monitoring, and graphical indication of data transmission and packet volume. As the result of the study shows, Round Robin Algorithm allows servers to ensure definite traffic distribution, unless incoming data loads differ much. Although error levels are high in some cases, they were eventually alleviated by repeated tests for a long period of time.

Distributed Intrusion Detection System for Safe E-Business Model (안전한 E-Business 모델을 위한 분산 침입 탐지 시스템)

  • 이기준;정채영
    • Journal of Internet Computing and Services
    • /
    • v.2 no.4
    • /
    • pp.41-53
    • /
    • 2001
  • Multi-distributed web cluster model built for high availability E-Business model exposes internal system nodes on its structural characteristics and has a potential that normal job performance is impossible due to the intentional prevention and attack by an illegal third party. Therefore, the security system which protects the structured system nodes and can correspond to the outflow of information from illegal users and unfair service requirements effectively is needed. Therefore the suggested distributed invasion detection system is the technology which detects the illegal requirement or resource access of system node distributed on open network through organic control between SC-Agents based on the shared memory of SC-Server. Distributed invasion detection system performs the examination of job requirement packet using Detection Agent primarily for detecting illegal invasion, observes the job process through monitoring agent when job is progressed and then judges the invasion through close cooperative works with other system nodes when there is access or demand of resource not permitted.

  • PDF

Construction of web-based Database for Haliotis SNP (웹기반 전복류 (Haliotis) SNP 데이터베이스 구축)

  • Jeong, Ji-Eun;Lee, Jae-Bong;Kang, Se-Won;Baek, Moon-Ki;Han, Yeon-Soo;Choi, Tae-Jin;Kang, Jung-Ha;Lee, Yong-Seok
    • The Korean Journal of Malacology
    • /
    • v.26 no.2
    • /
    • pp.185-188
    • /
    • 2010
  • The Web-based the genus Haliotis SNP database was constructed on the basis of Intel Server Platform ZSS130 dual Xeon 3.2 GHz cpu and Linux-based (Cent OS) operating system. Haliotis related sequences (2,830 nucleotide sequences, 9,102 EST sequences) were downloaded through NCBI taxonomy browser. In order to eliminate vector sequences, we conducted vector masking step using cross match software with vector sequence database. In addition, poly-A tails were removed using Trimmest software from EMBOSS package. The processed sequences were clustered and assembled by TGICL package (TIGR tools) equipped with CAP3 software. A web-based interface (Haliotis SNP Database, http://www.haliotis.or.kr) was developed to enable optimal use of the clustered assemblies. The Clustering Res. menu shows the contig sequences from the clustering, the alignment results and sequences from each cluster. And also we can compare any sequences with Haliotis related sequences in BLAST menu. The search menu is equipped with its own search engine so that it is possible to search all of the information in the database using the name of a gene, accession number and/or species name. Taken together, the Web-based SNP database for Haliotis will be valuable to develop SNPs of Haliotis in the future.

Recommendation of Best Empirical Route Based on Classification of Large Trajectory Data (대용량 경로데이터 분류에 기반한 경험적 최선 경로 추천)

  • Lee, Kye Hyung;Jo, Yung Hoon;Lee, Tea Ho;Park, Heemin
    • KIISE Transactions on Computing Practices
    • /
    • v.21 no.2
    • /
    • pp.101-108
    • /
    • 2015
  • This paper presents the implementation of a system that recommends empirical best routes based on classification of large trajectory data. As many location-based services are used, we expect the amount of location and trajectory data to become big data. Then, we believe we can extract the best empirical routes from the large trajectory repositories. Large trajectory data is clustered into similar route groups using Hadoop MapReduce framework. Clustered route groups are stored and managed by a DBMS, and thus it supports rapid response to the end-users' request. We aim to find the best routes based on collected real data, not the ideal shortest path on maps. We have implemented 1) an Android application that collects trajectories from users, 2) Apache Hadoop MapReduce program that can cluster large trajectory data, 3) a service application to query start-destination from a web server and to display the recommended routes on mobile phones. We validated our approach using real data we collected for five days and have compared the results with commercial navigation systems. Experimental results show that the empirical best route is better than routes recommended by commercial navigation systems.