• Title/Summary/Keyword: 고객-서버 모델

Search Result 33, Processing Time 0.02 seconds

A Study on Model of Realtime Automation for Website Authoring Tool using Live Site Concept (Live Site 개념을 도입한 웹사이트 저작도구의 실시간 자동화 모델에 관한 연구)

  • Chang, Young-Hyun;Park, Dae-Woo;Lee, Yeo-Won
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2011.06a
    • /
    • pp.175-177
    • /
    • 2011
  • 본 논문의 Live Site 개념을 도입한 웹사이트 저작도구의 실시간 자동화 모델에 관한 연구에서는 MVC(Model-View-Controller) 패턴의 시스템으로 사용자가 요구 사항을 전달하면 jsp에서 jsp 간의 호출을 통해 서버로 변경 사항을 넘기고 그 결과물을 다시 사용자에게 보여주는 형식으로 진행된다. 본 시스템 개발에 사용한 Jquery는 자바스크립트와 HTML 사이의 상호작용을 강조하는 경량화된 web application framework로 일반적 웹 스크립팅에 폭넓게 사용 될 수 있는 추상적 계층을 제공하여 스크립팅에서 필요로 하는 거의 모든 상황에 사용 할 수 있다. 본 논문에서는 추상화된 데이터를 제공하여 일상적인 작업들을 일반화 하고 코드의 크기를 줄이며 극도로 단순하게 개발이 가능한 jquery를 사용하여 거의 모든 브라우저에 호환이 가능한, 사용자 각 개인의 경향에 맞춘 웹사이트 저작 도구를 개발하였다. 본 논문에서는 추상화된 데이터를 제공하므로 일상적인 작업들을 일반화 하고 코드의 크기를 줄이며 극도로 단순하게 개발이 가능한 jquery를 사용하여 거의 모든 브라우저에 호환이 가능한, 사용자 각 개인의 경향에 맞춘 웹 저작 도구를 연구하였다. 전 세계적으로 웹 시장이 대두 되는 이 시점에 본 프로그램은 다양한 웹 제작 공급에 대한 새로운 시장을 형성해 주며, 새로운 콘텐츠 제작 방식의 도입으로 인한 활발한 인터넷 시장이 형성 되리라 기대한다. 현재 일부 생소한 Live Site 개념 즉, '사용자가 직접 보고 느끼며 원하는 대로 만드는 웹' 이란 개념의 가지고 고객 만족 커뮤니티라는 목적에 중점을 둔 본 프로그램 개발은 최근 웹 경향에 따른 이상적인 시스템이라 할 수 있다.

  • PDF

Overlay Multicast Network for IPTV Service using Bandwidth Adaptive Distributed Streaming Scheme (대역폭 적응형 분산 스트리밍 기법을 이용한 IPTV 서비스용 오버레이 멀티캐스트 네트워크)

  • Park, Eun-Yong;Liu, Jing;Han, Sun-Young;Kim, Chin-Chol;Kang, Sang-Ug
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.12
    • /
    • pp.1141-1153
    • /
    • 2010
  • This paper introduces ONLIS(Overlay Multicast Network for Live IPTV Service), a novel overlay multicast network optimized to deliver live broadcast IPTV stream. We analyzed IPTV reference model of ITU-T IPTV standardization group in terms of network and stream delivery from the source networks to the customer networks. Based on the analysis, we divide IPTV reference model into 3 networks; source network, core network and access network, ION(Infrastructure-based Overlay Multicast Network) is employed for the source and core networks and PON(P2P-based Overlay Multicast Network) is applied to the access networks. ION provides an efficient, reliable and stable stream distribution with very negligible delay while PON provides bandwidth efficient and cost effective streaming with a little tolerable delay. The most important challenge in live P2P streaming is to reduce end-to-end delay without sacrificing stream quality. Actually, there is always a trade-off between delay & stream quality in conventional live P2P streaming system. To solve this problem, we propose two approaches. Firstly, we propose DSPT(Distributed Streaming P2P Tree) which takes advantage of combinational overlay multicasting. In DSPT, a peer doesn't fully rely on SP(Supplying Peer) to get the live stream, but it cooperates with its local ANR(Access Network Relay) to reduce delay and improve stream quality. When RP detects bandwidth drop in SP, it immediately switches the connection from SP to ANR and continues to receive stream without any packet loss. DSPT uses distributed P2P streaming technique to let the peer share the stream to the extent of its available bandwidth. This means, if RP can't receive the whole stream from SP due to lack of SP's uploading bandwidth, then it receives only partial stream from SP and the rest from the ANR. The proposed distributed P2P streaming improves P2P networking efficiency.

Design and Implementation of MongoDB-based Unstructured Log Processing System over Cloud Computing Environment (클라우드 환경에서 MongoDB 기반의 비정형 로그 처리 시스템 설계 및 구현)

  • Kim, Myoungjin;Han, Seungho;Cui, Yun;Lee, Hanku
    • Journal of Internet Computing and Services
    • /
    • v.14 no.6
    • /
    • pp.71-84
    • /
    • 2013
  • Log data, which record the multitude of information created when operating computer systems, are utilized in many processes, from carrying out computer system inspection and process optimization to providing customized user optimization. In this paper, we propose a MongoDB-based unstructured log processing system in a cloud environment for processing the massive amount of log data of banks. Most of the log data generated during banking operations come from handling a client's business. Therefore, in order to gather, store, categorize, and analyze the log data generated while processing the client's business, a separate log data processing system needs to be established. However, the realization of flexible storage expansion functions for processing a massive amount of unstructured log data and executing a considerable number of functions to categorize and analyze the stored unstructured log data is difficult in existing computer environments. Thus, in this study, we use cloud computing technology to realize a cloud-based log data processing system for processing unstructured log data that are difficult to process using the existing computing infrastructure's analysis tools and management system. The proposed system uses the IaaS (Infrastructure as a Service) cloud environment to provide a flexible expansion of computing resources and includes the ability to flexibly expand resources such as storage space and memory under conditions such as extended storage or rapid increase in log data. Moreover, to overcome the processing limits of the existing analysis tool when a real-time analysis of the aggregated unstructured log data is required, the proposed system includes a Hadoop-based analysis module for quick and reliable parallel-distributed processing of the massive amount of log data. Furthermore, because the HDFS (Hadoop Distributed File System) stores data by generating copies of the block units of the aggregated log data, the proposed system offers automatic restore functions for the system to continually operate after it recovers from a malfunction. Finally, by establishing a distributed database using the NoSQL-based Mongo DB, the proposed system provides methods of effectively processing unstructured log data. Relational databases such as the MySQL databases have complex schemas that are inappropriate for processing unstructured log data. Further, strict schemas like those of relational databases cannot expand nodes in the case wherein the stored data are distributed to various nodes when the amount of data rapidly increases. NoSQL does not provide the complex computations that relational databases may provide but can easily expand the database through node dispersion when the amount of data increases rapidly; it is a non-relational database with an appropriate structure for processing unstructured data. The data models of the NoSQL are usually classified as Key-Value, column-oriented, and document-oriented types. Of these, the representative document-oriented data model, MongoDB, which has a free schema structure, is used in the proposed system. MongoDB is introduced to the proposed system because it makes it easy to process unstructured log data through a flexible schema structure, facilitates flexible node expansion when the amount of data is rapidly increasing, and provides an Auto-Sharding function that automatically expands storage. The proposed system is composed of a log collector module, a log graph generator module, a MongoDB module, a Hadoop-based analysis module, and a MySQL module. When the log data generated over the entire client business process of each bank are sent to the cloud server, the log collector module collects and classifies data according to the type of log data and distributes it to the MongoDB module and the MySQL module. The log graph generator module generates the results of the log analysis of the MongoDB module, Hadoop-based analysis module, and the MySQL module per analysis time and type of the aggregated log data, and provides them to the user through a web interface. Log data that require a real-time log data analysis are stored in the MySQL module and provided real-time by the log graph generator module. The aggregated log data per unit time are stored in the MongoDB module and plotted in a graph according to the user's various analysis conditions. The aggregated log data in the MongoDB module are parallel-distributed and processed by the Hadoop-based analysis module. A comparative evaluation is carried out against a log data processing system that uses only MySQL for inserting log data and estimating query performance; this evaluation proves the proposed system's superiority. Moreover, an optimal chunk size is confirmed through the log data insert performance evaluation of MongoDB for various chunk sizes.