• Title/Summary/Keyword: TCP 버퍼 튜닝

Search Result 6, Processing Time 0.02 seconds

TCP Buffer Tuning based on MBT for High-Speed Transmissions in Wireless LAN (무선 랜 고속전송을 위한 최대버퍼한계 기반 TCP 버퍼튜닝)

  • Mun, Sung-Gon;Lee, Hong-Seok;Choo, Hyun-Seung;Kong, Won-Young
    • Journal of Internet Computing and Services
    • /
    • v.8 no.1
    • /
    • pp.15-23
    • /
    • 2007
  • Wireless LAN (IEEE 802.11) uses traditional TCP for reliable data transmission, But it brings the unintentional packet loss which is not congestion loss caused by handoff, interference, and fading in wireless LAN. In wireless LAN, TCP experiences performance degradation because it consumes that the cause of packet loss is congestion, and it decrease the sending rate by activating congestion control algorithm. This paper analyzes that correlation of throughput and buffer size for wireless buffer tuning. We find MBT (Maximum Buffer Threshold) which does not increase the throughput through the analysis, For calculation of MBT, we experiment the throughput by using high volume music data which is creased by real-time performance of piano. The experiment results is shown that buffer tuing based on MBT shows 20.3%, 21.4%, and 45.4% throughput improvement under 5ms RTT, 10ms RTT, and 20ms RTT, respectively, comparing with the throughput of operation system default buffer size, In addition, we describe that The setting of TCP buffer size by exceeding MBT does not have an effect on the performance of TCP.

  • PDF

An Adaptive Buffer Tuning Mechanism for striped transport layer connection on multi-homed mobile host (멀티홈 모바일 호스트상에서 스트라이핑 전송계층 연결을 위한 적응형 버퍼튜닝기법)

  • Khan, Faraz-Idris;Huh, Eui-Nam
    • Journal of Internet Computing and Services
    • /
    • v.10 no.4
    • /
    • pp.199-211
    • /
    • 2009
  • Recent advancements in wireless networks have enabled support for mobile applications to transfer data over heterogeneous wireless paths in parallel using data striping technique [2]. Traditionally, high performance data transfer requires tuning of multiple TCP sockets, at sender's end, based on bandwidth delay product (BDP). Moreover, traditional techniques like Automatic TCP Buffer Tuning (ATBT), which balance memory and fulfill network demand, is designed for wired infrastructure assuming single flow on a single socket. Hence, in this paper we propose a buffer tuning technique at senders end designed to ensure high performance data transfer by striping data at transport layer across heterogeneous wireless paths. Our mechanism has the capability to become a resource management system for transport layer connections running on multi-homed mobile host supporting features for wireless link i.e. mobility, bandwidth fluctuations, link level losses. We show that our proposed mechanism performs better than ATBT, in efficiently utilizing memory and achieving aggregate throughput.

  • PDF

QoS Implementation by using PID Control of TCP Buffer in Data Grid (데이타 그리드상에서 TCP 버퍼의 PID 제어를 통한 QoS 구현)

  • Lee Hong-Seok;Huh Eui-Nam;Choo Hyun-Seung;Cho Kum-Won
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.33 no.7
    • /
    • pp.363-369
    • /
    • 2006
  • It is aware that Bandwidth management with dramatically increasing traffic on account of frequent and large file transmission in a data grid environment is one of essential needs. A this paper we propose new method which guarantees QoS (Quality of Service) by being in control of resources in TCP layer based on existing studies that manage bandwidth over TCP buffer tuning. General QoS solutions manage network resources subsequent to observing them in IP or link layer, but the scheme in the paper is able to control network resources in TCP layer that is network upper layer. Consequently, bandwidth allocation to each user can be efficiently controlled depending on an authority each user is given so that users could be use different bandwidth. It is expected that a new paradigm is supposed in network resource management and the method of levies for users' bandwidth uses.

Design of High-Performance Lambda Network Based on DRS Model (DRS 모델에 기반한 고성능 람다 네트워크의 설계)

  • Noh, Min-Ki;Ahn, Sung-Jin
    • The Journal of Korean Association of Computer Education
    • /
    • v.12 no.2
    • /
    • pp.77-86
    • /
    • 2009
  • Large-scale applications, that needs large-capacity R&D resources and realtime data transmission, have demanded more stable and high-performance network environment than current Internet environments. Recently, global R&D networks have focuses on utilizing Lambda networking technologies and resource reservation systems to be satisfied with various applications' requirements. In this paper, we modify the existing DRS (Dynamic Right-Sizing) model to reflect various advantages in terms of the stability and high-capacity of Lambda network. In addition, we suggest the design methodology of high-performance Lambda network, which can integrate NRPS (Network Resource Provisioning System) into our modified DRS model.

  • PDF

A Maximum Mechanism of Data Transfer Rate using Parallel Transmission Technology on High Performance Network (고성능 네트워크에서 병렬 전송 기술을 이용한 전송률 극대화 메커니즘)

  • Kim, Young-Shin;Huh, Eui-Nam
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.34 no.9
    • /
    • pp.425-434
    • /
    • 2007
  • Even though Internet backbone speeds have increased in the last few years due to projects like Internet 2 and NGI, many high performance distributed applications are able to achieve only a small fraction of the available bandwidth. The cause of such problem is due to a character of TCP/IP. The primary goal of this protocol is reliable data transmission. Therefore high speed data transmission didn't be considered when TCP/IP is designed. Hence several researchers have been studied in order to solve the problem of TCP/IP. One of these research results, parallel transfer technique, solves this problem to use parallel TCP connections on application level. Additionally, this technique is compatibility. Recently, these researchers have been studied a mechanism to decide the number of parallel TCP connections. However, some researchers reported the number of parallel TCP connection base on only empirical results. Although hardware performance of host affects transmission rate, the hardware performance didn't be considered in their works. Hence, we collect all data related to transmission rate, such as hardware state information (cpu utilization, interrupt, context switch). Then, we analyzed collected data. And, we suggest a new mechanism determining number of parallel TCP connections for maximization of performance based on our analysis.

Modeling and Performance Evaluation of the Web server supporting Persistent Connection (Persistent Connection을 지원하는 웹서버 모델링 및 성능분석)

  • Min, Byeong-Seok;Nam, Ui-Seok;Lee, Sang-Mun;Sim, Yeong-Seok;Kim, Hak-Bae
    • The KIPS Transactions:PartC
    • /
    • v.9C no.4
    • /
    • pp.605-614
    • /
    • 2002
  • Amount of the web traffic web server handles are explosively increasing, which requires that the performance of the web server should be improved for the various web services. Although the analysis for the HTTP traffic with the proper tuning for the web server is essential, the research relevant to the subject are insignificant. In particular, although most of applications are implemented over HTTP 1.1 protocol, the researches mostly deal with the performance evaluation of the HTTP 1.0 protocol. Consequently, the modeling approach and the performance evaluation over HTTP 1.1 protocol have not been well formed. Therefore, basing on the HTTP 1.1 protocol supporting persistent connection, we present an analytical end-to-end tandem queueing model for web server to consider the specific hardware configuration inside web server beginning at accepting the user request until completing the service. we compare various performances between HTTP 1.0 and HTTP 1.1 under the overloading condition, and then analyze the characteristics of the HTTP traffic that include file size requested to web server, the OFF time between file transfers, the frequency of requests, and the temporal locality of requests. Presented model is verified through the comparing the server throughput according to varying requests rate with the real web server. Thereafter, we analyze the performance evaluation of the web server, according to the interrelation between TCP Listen queue size, the number of HTTP threads and the size of the network buffers.