Nothing Special   »   [go: up one dir, main page]

Skip to main content

Improve Prefetch Performance by Splitting the Cache Replacement Queue

  • Conference paper
Advanced Infocomm Technology (ICAIT 2012)

Part of the book series: Lecture Notes in Computer Science ((LNCCN,volume 7593))

Included in the following conference series:

Abstract

The performance of a prefetch cache is dependent on both the prefetch technique and the cache replacement policy. Both these algorithms execute independently of each other, but they share a data structure - the cache replacement queue. This paper shows that even with a simple prefetch technique, there is an increase in hit rate when the LRU replacement queue is split into two equal sized queues. A more significant performance improvement is possible with a sophisticated prefetch technique and by splitting the queue unequally.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. Bhatia, S., Varki, E., Merchant, A.: Sequential prefetch cache sizing for maximal hit rate. In: 18th Annual IEEE/ACM International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems, pp. 89–98 (2010)

    Google Scholar 

  2. Butt, A.R., Gniady, C., Hu, Y.C.: The performance impact of kernel prefetching on buffer cache replacement algorithms. IEEE Transactions on Computers 56(7), 889–908 (2007)

    Article  MathSciNet  Google Scholar 

  3. Gill, B.S., Modha, D.S.: SARC: Sequential prefetching in adaptive replacement cache. In: Proc. of USENIX 2005 Annual Technical Conference, pp. 293–308 (2005)

    Google Scholar 

  4. Jiang, S.: Dulo: An effective buffer cache management scheme to exploit both temporal and spatial localities. In: USENIX Conference on File and Storage Technologies (FAST) (2005)

    Google Scholar 

  5. Li, M., Varki, E., Bhatia, S., Merchant, A.: TaP: Table-based prefetching for storage caches. In: 6th USENIX Conference on File and Storage Technologies (FAST 2008), pp. 81–97 (2008)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Varki, E., Hubbe, A., Merchant, A. (2013). Improve Prefetch Performance by Splitting the Cache Replacement Queue. In: Guyot, V. (eds) Advanced Infocomm Technology. ICAIT 2012. Lecture Notes in Computer Science, vol 7593. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-38227-7_13

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-38227-7_13

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-38226-0

  • Online ISBN: 978-3-642-38227-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics