• Title/Summary/Keyword: 선인출

Search Result 49, Processing Time 0.027 seconds

A Study on the Prefetch Method for Multimedia Presentations to play on the Low Bandwidth Mobile Devices (낮은 대역폭을 갖는 이동 단말기에서 멀티미디어 프리젠테이션을 재생하기 위한 프리패치 방법)

  • Hong Maria;Yang Hyuck;Lim Young-Haw
    • Journal of Internet Computing and Services
    • /
    • v.4 no.4
    • /
    • pp.65-74
    • /
    • 2003
  • Recently, It is required that multimedia data display on not only PC but also the Mobile device such as cellar phone, PDA. Mobile divieces has not sufficient of storage space or main memory and has slow network speed. So we can't process the multimedia data with the existing way on PC. This paper presents various methods for MultimediaPresentation which are able to display on the mobile device with the low-bandwidth and small storage. There Is an analysis that Is playable on the mobile device using presentation area devision method and replayable analysis method. If the display of multimedia presentation is impossible from this analysis, to apply prefetch techniques a using EPOB apoint and as User QoS of multimedia data through prefetch techniques have keep, do so that playable. Then, it is proposed to playable smooth multimedia presentation as that each of the presentations through prefetch Scheduling techniques point and quantity of data which are prefetch.

  • PDF

An Adaptive Prefetching Technique for Software Distributed Shared Memory Systems (소프트웨어 분산공유메모리시스템을 위한 적응적 선인출 기법)

  • Lee, Sang-Kwon;Yun, Hee-Chul;Lee, Joon-Won;Maeng, Seung-Ryoul
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.28 no.9
    • /
    • pp.461-468
    • /
    • 2001
  • Though shared virtual memory (SVM) system promise low cost solutions for high performance computing they suffer from long memory latencies. These latencies are usually caused by repetitive invalidations on shared data. Since shared data are accessed through synchronization and the patterns by which threads synchronizes are repetitive, a prefetching scheme bases on such repetitiveness would reduce memory latencies. Based on this observation, we propose a prefetching technique which predicts future access behavior by analyzing access history per synchronization variable. Our technique was evaluated on an 8-node SVM system using the SPLASH-2 benchmark. The results show the our technique could achieve 34%~45% reduction in memory access latencies.

  • PDF

Buffer Prefatching for The Streaming Multimedia Service in Mobile Devices (모바일 이동 단말기에서 스트리밍 미디어서비스를 위한 버퍼 선인출)

  • Kim, Jongwoo;Kim, Daeho;Jeon, Soobin;Chae, Myunghoon;Jung, Inbum
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2009.04a
    • /
    • pp.1247-1249
    • /
    • 2009
  • 최근 첨단 IT 분야 기반 기술들의 상용화가 시작되고 있다. 이들 기술들을 바탕으로 이동 단말에서 다양한 서비스를 즐길 수 있는 시대로 진입하고 있다. 이동 중인 사용자는 핸드오프가 빈번히 발생되는 환경에서 스트리밍 서비스를 받게 된다. 이러한 환경에서 스트리밍 미디어의 QoS를 만족시키기 위해서는 이동성이 고려된 미디어 데이터선인출이 사용되어야 한다. 본 논문에서는 미디어 서버에서 사용자 단말로 스트리밍 미디어가 서비스 될 때 빈번한 핸드오프에 의한 기지국(Base Station : BS) 스트리밍 서버의 자원낭비를 연구한다. 논문에서는 BS의 버퍼용량을 속도와 이동거리에 따라 변화하는 량을 조사하여 선인출 용량 변화가 버퍼 사용률에 영향을 미침을 지적한다.

Application-Oriented Context Pre-fetch Method for Enhancing Inference Performance in Ontology-based Context Management (온톨로지 기반의 상황정보관리에서 추론 성능 향상을 위한 어플리케이션 지향적 상황정보 선인출 기법)

  • Lee Jae-Ho;Park In-Suk;Lee Dong-Man;Hyun Soon-Joo
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.12 no.4
    • /
    • pp.254-263
    • /
    • 2006
  • Ontology-based context models are widely used in ubiquitous computing environment because they have advantages in the acquisition of conceptual context through inferencing, context sharing, and context reusing. Among the benefits, inferencing enables context-aware applications to use conceptual contexts which cannot be acquired by sensors. However, inferencing causes processing delay and thus becomes the major obstacle to the implementation of context-aware applications. The delay becomes longer as the amount of contexts increases. In this paper, we propose a context pre-fetching method to reduce the size of contexts to be processed in a working memory in attempt to speed up inferencing. For this, we extend the query-tree method to identify contexts relevant to the queries of a context-aware application. Maintaining the pre-fetched contexts optimal in a working memory, the processing delay of inference reduces without the loss of the benefits of ontology-based context model. We apply the proposed scheme to our ubiquitous computing middleware, Active Surroundings, and demonstrate the performance enhancement by experiments.

Research on Web Cache Infection Methods and Countermeasures (웹 캐시 감염 방법 및 대응책 연구)

  • Hong, Sunghyuck;Han, Kun-Hee
    • Journal of Convergence for Information Technology
    • /
    • v.9 no.2
    • /
    • pp.17-22
    • /
    • 2019
  • Cache is a technique that improves the client's response time, thereby reducing the bandwidth and showing an effective side. However, there are vulnerabilities in the cache technique as well as in some techniques. Web caching is convenient, but it can be exploited by hacking and cause problems. Web cache problems are mainly caused by cache misses and excessive cache line fetch. If the cache miss is high and excessive, the cache will become a vulnerability, causing errors such as transforming the secure data and causing problems for both the client and the system of the user. If the user is aware of the cache infection and the countermeasure against the error, the user will no longer feel the cache error or the problem of the infection occurrence. Therefore, this study proposed countermeasures against four kinds of cache infections and errors, and suggested countermeasures against web cache infections.

A Level One Cache Organization for Chip-Size Limited Single Processor (칩의 크기가 제한된 단일칩 프로세서를 위한 레벨 1 캐시구조)

  • Ju YoungKwan;Kim Sukil
    • The KIPS Transactions:PartA
    • /
    • v.12A no.2 s.92
    • /
    • pp.127-136
    • /
    • 2005
  • This paper measured a proper ratio of the size of demand fetch cache $L_1$ to that of prefetch cache $L_P$ by imulation when the size of $L_1$ and $L_P$ are constant which organize space-limited level 1 cache of a single microprocessor chip. The analysis of our experiment showed that in the condition of the sum of the size of $L_1$ and $L_P$ are 16 KB, the level 1 cache organization by constituting $L_P$ with 4 KB and employing OBL and FIFO as a prefetch technique and a cache replacement policy respectively resulted in the best performance. Also, this analysis showed that in the condition of the sum of the size of $L_1$ and $L_P$ are over 32 KB, employing dynamic filtering as prefetch technique of $L_P$ are more advantageous and splitting level 1 cache by constituting $L_1$ with 28 KB and $L_P$ with 4 KB in the case of 32 KB of space are available, by constituting $L_1$ with 48 KB and $L_P$ with 16 KB in the case of 64 KB elicited the best performance.

An Efficient H.264/AVC Decoding Technique Using Prefetching Mechanism (선인출 메커니즘을 이용한 효율적인 H.264/AVC 복호화 기법)

  • Ji, Shin-Haeng;Park, Jung-Wook;Kim, Shin-Dug
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2005.11a
    • /
    • pp.946-948
    • /
    • 2005
  • H.264/AVC는 SoC/IEC MPEG와 ITU-T Video Coding Experts Group에서 함께 발표한 비디오 코딩을 위한 가장 최근의 표준이다. 기존의 표준들 보다 적은 비트로 높은 압축률과 좋은 화질을 제공하고 있다. 그러나 1/4 화소 움직임 예측과 보상의 지원과 7가지의 가변블록에 대한 움직임예측과 블록모드별 RD(Rate-Distortion)를 수행하고 CAVLC등 H.264/AVC 표준에서 채택한 여러 가지 비디오 압축방식으로 인해 그 복잡도가 훨씬 증가하였다. 이 논문에서 H.264/AVC의 복호화기에서 복잡도의 약 $40\%$ 이상을 차지하는 움직임보상 모듈을 효율적으로 수행하고 최적화하기 위한 방법을 제안한다. 예측된 모션벡터에 따라 창조하는 프레임에서 매크로블록을 만들어 내는 움직임 보상 과정을 수행하는 데 있어서 접근 지연시간이 큰 외부 메모리 창조를 선인출 메커니즘을 이용하여 미리 예측하여 수행함으로써 전체 수행시간을 줄이는 기법을 적용하였다. 이를 통하여 가변길이 복호화 모듈과 움직임 보상모듈을 수정하여 반복적으로 읽고, 쓰기를 수행해야 하는 횟수를 줄였다. 실험 결과 본 논문에서 제안하는 방법을 이용하여 복호화 과정을 수행했을 때 PSNR(Peak Signal to Noise Ratio)의 손실은 전혀 없으면서 복호화기의 전체 실행시간을 약 $5\%$ 향상시키고, 핵심 모듈인 움직임 보상과정에서 약 $20\%$ 정도 실행시간을 향상시키는 등 높은 성능 향상을 보였다.

  • PDF

Design and Implementation of Prefetching Mechanism for Software Streaming Systems in Mobile Environment (모바일 환경 기반의 소프트웨어 스트리밍 시스템을 위한 선인출 기법의 설계 및 구현)

  • Lee Dae-Woo;Park Seon-Yeong;Kim Jin-Soo;Maeng Seung-Ryoul
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2006.05a
    • /
    • pp.1329-1332
    • /
    • 2006
  • 온디맨드 소프트웨어 스트리밍(On-Demand Software Streaming)이란 서버가 제공하는 소프트웨어를 클라이언트에 설치하지 않고 실행하는 기술로, 서버로부터 필요한 부분만 스트리밍으로 전송 받아 실행하는 기술을 말한다. 이 기술을 이용하면 소프트웨어 제공자는 소프트웨어 관리를 용이하게 할 수 있고, 소프트웨어 사용자는 적은 저장 공간으로 많은 소프트웨어를 사용할 수 있다는 이점을 얻게 된다. 하지만 모바일 환경에서 이를 이용하는 경우에는, 느린 무선 네트워크를 통해 소프트웨어 이미지를 전송해야 하기 때문에 소프트웨어 실행 속도가 매우 느리다는 문제가 생긴다. 이를 해결하기 위해서 본 논문에서는 온디맨드 소프트웨어 스트리밍을 사용하는 시스템의 성능 향상을 위해 효율적인 선인출 기법을 설계하고 실제로 구현하였다. 실험 결과, 애플리케이션이 데이터를 읽을 때 걸리는 시간이 무선랜 환경에서는 평균 50%, CDMA 환경에서는 평균 20% 정도 감소했으며, 특히 네트워크 지연시간이 증가할수록 더 많이 감소하였다.

  • PDF

Analysis of Web Server Referencing Characteristics and performance Improvement of Web Server (웹 서버의 참조 특성 분석과 성능 개선)

  • Ahn, Hyo-Beom;Cho, Kyung-San
    • The KIPS Transactions:PartA
    • /
    • v.8A no.3
    • /
    • pp.201-208
    • /
    • 2001
  • Explosive growth of the Web and the non-uniform characteristics of client requests result in the performance degradation of Web servers, and server cache has been recognized as the solution. We analyzed Web server accessing characteristics-repetition, size, and locality of access. Based on the result, we analyzed the cache removal policies and proposed a prefetch strategy to improve the hit ratio of server caches. In addition, through the trace-driven simulation based on the traces from real Web sites, we showed the performance improvement by our proposal.

  • PDF

The Reducting Technique of compulsory Misses for S/W managed TLB (S/W관리 TLB의 초기접근실패 감소 기법)

  • Park, Jang-Suk
    • The Transactions of the Korea Information Processing Society
    • /
    • v.5 no.3
    • /
    • pp.620-632
    • /
    • 1998
  • This paper introduces a new teehniquc for reducing the compulsory misses of software-managed TLBs by prefetching necessary TLB entries before being used. This technique is not inherently limited to specific applications. The key of this scheme is to perform the prefetch operations to update the TLB entries before first accesses so that TLB misses can be avoided. For the identifications of the prefctch pages, the new classification is introduced, which is based 0n the view of an object code execution. Then, the algorithms and the implementation technique arc described. Using a quantitative analysis, the proposed scheme is evaluated to prove that it is a useful technique for the perronnall~~ ~nhan~"ment of the S/W managed TLBs. in addition, it is discussed that reducing the miss rate by the prefeteh scheme reduces the total miss penalty and bus traffics in S/W-managed TLBs.

  • PDF