• Title/Summary/Keyword: file caching

Search Result 52, Processing Time 0.027 seconds

Partition and Caching Mechanism for GML Visualization on Mobile Device (모바일 디바이스에서 GML 가시화를 위한 분할 및 캐싱 기법)

  • Song, Eun-Ha;Park, Yong-Jin;Han, Won-Hee;Jeong, Young-Sik
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.7
    • /
    • pp.1025-1034
    • /
    • 2008
  • In this paper, we developed GridGML for efficiently supplying a GML and visualizing the map with partitioning map and caching method to a mobile device. In order to overcome the weighting of a file, which is the biggest weakness of a GML, GridGML extracts only the most necessary parts for the visualization of the map among GML attributes, and makes the file light as a class instance by applying an offset value. GridGML manages a partition based on the visualization area of a mobile device to visualize the map to a mobile device in real time, and transmits the partition area by serializing it for the benefit of transmission. Also, the received partition area is compounded in a mobile device and is visualized by being partitioned again as four visible areas based on the display of a mobile device. Then, the area is managed by applying a caching algorithm in consideration of repetitiveness for a received map for the efficient operation of resources. Also, in order to prevent the delay in transmission time as regards the instance density area of the map, an adaptive map partition mechanism is proposed for maintaining the transmission time uniformly.

  • PDF

Performance Enhancement Algorithm for Remote file service in VDFS (VDFS에서 원격 화일 서비스의 성능 향상 알고리즘)

  • Yun, Dong-Sik;Lee, Byeong-Gwan
    • The Transactions of the Korea Information Processing Society
    • /
    • v.2 no.2
    • /
    • pp.156-167
    • /
    • 1995
  • VDFS(V-kernel Distributed File System) constructed in this paper proposes an asymmetric Distributed File System so that various problems of existing symmetric structures like NFS & RFS may be solved in small-sized system. Particulary, in this paper, server workload can be reduced by using a caching mechanism and remote access transparency can be improved by using a sharing file duplicately.

  • PDF

Prefetching Policy based on File Acess Pattern and Cache Area (파일 접근 패턴과 캐쉬 영역을 고려한 선반입 기법)

  • Lim, Jae-Deok;Hwang-Bo, Jun-Hyeong;Koh, Kwang-Sik;Seo, Dae-Hwa
    • The KIPS Transactions:PartA
    • /
    • v.8A no.4
    • /
    • pp.447-454
    • /
    • 2001
  • Various caching and prefetching algorithms have been investigated to identify and effective method for improving the performance of I/O devices. A prefetching algorithm decreases the processing time of a system by reducing the number of disk accesses when an I/O is needed. This paper proposes an AMBA prefetching method that is an extended version of the OBA prefetching method. The AMBA prefetching method will prefetching blocks continuously as long as disk bandwidth is enough. In this method, though there were excessive data request rate, we would expect efficient prefetching. And in the AMBA prefetching method, to prevent the cache pollution, it limits the number of data blocks to be prefetched within the cache area. It can be implemented in a user-level File System based on a Linux Operating System. In particular, the proposed prefetching policy improves the system performance by about 30∼40% for large files that are accessed sequentially.

  • PDF

A Scalable Cache Group Configuration Policy using Role-Partitioned Cache (캐쉬의 역할 구분을 이용한 확장성이 있는 캐쉬 그룹 구성 정책)

  • 현진일;민준식
    • The Journal of the Korea Contents Association
    • /
    • v.3 no.3
    • /
    • pp.63-73
    • /
    • 2003
  • Today, in exponential growth of internet, the importance of file caching which could reduce the sun load, the volume of network traffic, and the latency of response has emerged. Actually, in one network, the traffic has reduced by using the cache and this means that file caching can improve the internet environment by cost a fraction of link upgrades. In this paper, we address a dynamic cache group configuration policy, to solve the scalable problem. The simulation result shows that the cache group using our proposal policy reduces the latency of response time and it means that out cache group configuration is more scalable than the static cache configuration.

  • PDF

File-System-Level SSD Caching for Improving Application Launch Time (응용프로그램의 기동시간 단축을 위한 파일 시스템 수준의 SSD 캐싱 기법)

  • Han, Changhee;Ryu, Junhee;Lee, Dongeun;Kang, Kyungtae;Shin, Heonshik
    • Journal of KIISE
    • /
    • v.42 no.6
    • /
    • pp.691-698
    • /
    • 2015
  • Application launch time is an important performance metric to user experience in desktop and laptop environment, which mostly depends on the performance of secondary storage. Application launch times can be reduced by utilizing solid-state drive (SSD) instead of hard disk drive (HDD). However, considering a cost-performance trade-off, utilizing SSDs as caches for slow HDDs is a practicable alternative in reducing the application launch times. We propose a new SSD caching scheme which migrates data blocks from HDDs to SSDs. Our scheme operates entirely in the file system level and does not require an extra layer for mapping SSD-cached data that is essential in most other schemes. In particular, our scheme does not incur mapping overheads that cause significant burdens on the main memory, CPU, and SSD space for mapping table. Experimental results conducted with 8 popular applications demonstrate our scheme yields 56% of performance gain in application launch, when data blocks along with metadata are migrated.

IT-based Technology An Efficient Global Buffer Management ,algorithm for SAN Environments (SAN 환경을 위한 효율적인 전역버퍼 관리 알고리즘)

  • 이석재;박새미;송석일;유재수;이장선
    • The Journal of the Korea Contents Association
    • /
    • v.4 no.3
    • /
    • pp.71-80
    • /
    • 2004
  • In distributed file-systems, cooperative caching algorithm which owns the data cached at each node jointly is used to reduce an expense of disk access. Cooperative caching algorithm is the method that increases a cache hit-ratio and decrease a disk access as it holds the cache information of distributed systems in common and makes cache larger virtually. Recently, several cooperative caching algorithms decrease the message costs by using approximate information of the cache and increase the cache hit-ratio by using local and global cache fields dynamically. And they have an advantage that increases the whole field hit-ratio by sending a replaced buffer to the idle node on buffers replacement in order to maintain the replaced cache in the cache field. However the wrong approximate information deteriorates the performance, the consistency maintenance goes to great expense to exchange messages and the cost that manages Age-information of each node to choose the idle node increases. In this thesis, we propose a cooperative cache algorithm that maintains correct cache information, minimizes the maintenance cost for consistency and the management cost for buffer Age-information. Also, we show the superiority of our algorithm through the performance evaluation.

  • PDF

Fips : Dynamic File Prefetching Scheme based on File Access Patterns (Fips : 파일 접근 유형을 고려한 동적 파일 선반입 기법)

  • Lee, Yoon-Young;Kim, Chei-Yol;Seo, Dae-Wha
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.29 no.7
    • /
    • pp.384-393
    • /
    • 2002
  • A Parallel file system is normally used to support excessive file requests from parallel applications in a cluster system, whereas prefetching is useful for improving the file system performance. This paper proposes a new prefetching method, Fips(dynamic File Prefetching Scheme based on file access patterms), that is particularly suitable for parallel scientific applications and multimedia web services in a parallel file system. The proposed prefetching method introduces a dynamic prefetching scheme to predict data blocks precisely in run-time although the file access patterns are irregular. In addition, it includes an algorithm to determine whether and when the prefetching is performed using the current available I/O bandwidth. Experimental results confirmed that the use of the proposed prefetching policy in a parallel file system produced a higher file system performance.

Regular File Access of Embedded System Using Flash Memory as a Storage (플래시 메모리를 저장매체로 사용하는 임베디드 시스템에서의 정규파일 접근)

  • 이은주;박현주
    • Journal of Information Technology Applications and Management
    • /
    • v.11 no.1
    • /
    • pp.189-200
    • /
    • 2004
  • Recently Flash Memory which is small and low-powered is widely used as a storage of embedded system, because an embedded system requests portability and a fast response. To resolve a difference of access time between a storage and RAM, Linux is using disk caching which copies a part of file on disk into RAM. It is not also an exception on embedded system. A READ access-time of flash memory is similar to RAMs. So, when a process on an embedded system reads data, it is similar to the time to access cached data in RAM and to access directly data on a flash memory. On the embedded system using limited memory, using a disk cache is that wastes much time and memory spaces to manage it and can not reflects the characteristic of a flash memory. This paper proposes the regular file access of limited using a page cache in the file system based on a flash memory and reflects the characteristic of a flash memory. The proposed algorithm minimizes power consumption because access numbers of the RAM are reduced and doesn't waste a memory space because it accesses directly to a flash memory Therefore, the performance improvement of the system applying the proposed algorithm is expected.

  • PDF

A Study on Improvement of Buffer Cache Performance for File I/O in Deep Learning (딥러닝의 파일 입출력을 위한 버퍼캐시 성능 개선 연구)

  • Jeongha Lee;Hyokyung Bahn
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.24 no.2
    • /
    • pp.93-98
    • /
    • 2024
  • With the rapid advance in AI (artificial intelligence) and high-performance computing technologies, deep learning is being used in various fields. Deep learning proceeds training by randomly reading a large amount of data and repeats this process. A large number of files are randomly repeatedly referenced during deep learning, which shows different access characteristics from traditional workloads with temporal locality. In order to cope with the difficulty in caching caused by deep learning, we propose a new sampling method that aims at reducing the randomness of dataset reading and adaptively operating on existing buffer cache algorithms. We show that the proposed policy reduces the miss rate of the buffer cache by 16% on average and up to 33% compared to the existing method, and improves the execution time by up to 24%.

Design of Network Protocol based on P2P Collaboration and User's Content Using Information (P2P 협업 및 사용자 콘텐츠 이용 정보 기반의 네트워크 프로토콜 설계)

  • Nahm, Eui-Seok
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.66 no.3
    • /
    • pp.575-580
    • /
    • 2017
  • In these days, the big-size and high resolution multimedia file is widely used through networks. To transfer and service effectively, the internet network technology is necessary to substitute broadcasting. Normally Content Delivery Network(CDN) is widely used in conventional internet for multimedia services. But it has a small bandwidth to service. So to solve this problems, many researchers have suggest the protocol for download, content distribution/saving, server synchronization, caching, pushing rate, and streaming etc. But all of these has some defects like low resolution, packets loss and delay, real application implementations etc. So, this paper suggests a new method of network protocol based on P2P collaboration and user's content using information. And it evaluated the performance of suggested method. As the results, it showed the effectiveness of 4 performances indices : download speed, decreasing rate of connected user in same time, adaptive hit ratio, traffic decreasing rate.