Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/1242531.1242566acmconferencesArticle/Chapter ViewAbstractPublication PagescfConference Proceedingsconference-collections
Article

Memory-miser: a performance-constrained runtime system for power-scalable clusters

Published:07 May 2007Publication History

ABSTRACT

Main memory in clusters may dominate total system power. The resulting energy consumption increases system operating cost and the heat produced reduces reliability. Emergent memory technology will provide servers with the ability to dynamically turn-on (online) and turn-off (offline) memory devices at runtime. This technology, coupled with slack in memory demand, offers the potential for significant energy savings in clusters of servers. Enabling power-aware memory and conserving energy in clusters are non-trivial. First, power-aware memory techniques must be scalable to thousands of devices. Second, techniques must not negatively impact the performance of parallel scientific applications. Third, techniques must be transparent to the user to be practical. We propose a Memory Management Infra-Structure for Energy Reduction (Memory MISER). Memory MISER is transparent, performance-neutral, and scalable. It consists of a prototype Linux kernel that manages memory at device granularity and a userspace daemon that monitors memory demand systemically to control devices and implement energy- and performance-constrained policies. Experiments on an 8-node cluster show our control daemon reduces memory energy up to 56.8% with <1% performance degradation for several classes of parallel scientific codes. Our daemon uses a PID controller to conservatively offline memory and aggressively online memory at runtime. For multi-user workloads where memory demand often spikes dramatically, Memory MISER can save up to 67.94% of memory energy with <1% performance degradation. Current IBM eServer systems support up to 2 terabytes of SDRAM per node and 16 processors. For a server-based cluster with 8 90-watt processors and 32 GB of SDRAM per processor, Memory MISER can save about 30% total system energy for multi-user parallel workloads.

References

  1. N. Adiga, G. Almasi, et al., "An Overview of the BlueGene/L Supercomputer," Proceedings of IEEE/ACM SC 2002, Baltimore, MD, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. K. J. Astrom and B. Wittenmark, Adaptive Control: Adison-Wesley, 1995. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. L. Benini and G. De Micheli, "System-level power optimization: techniques and tools," ACM TODAES, vol. 5, pp. 115--192, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. K. W. Cameron, X. Feng, and R. Ge, "Performance-and Energy-Conscious Distributed DVS Scheduling for Scientific Applications on Power-aware Clusters," Proceedings of 17th High Performance Computing, Networking and Storage Conference (SC 2005), Seattle, WA, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. K. W. Cameron, R. Ge, and X. Feng, "High-performance, power-aware, distributed computing for scientific applications," IEEE Computer, vol. 38, pp. 40--47, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. G. Chen, K. Malkowski, and P. Raghavan, "Reducing Power with Performance Constraints for Parallel Sparse Applications," Proceedings of 1st Workshop on High-performance, power-aware computing, Denver, CO, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. V. Delaluz, M. Kandemir, N. Vijaykrishnan, A. Sivasubramaniam, and M. J. Irwin, "Hardware and Software Techniques for Controlling DRAM Power Modes," IEEE Transactions On Computers, vol. 50, pp. 1154--1173, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. W. Feng, M. Warren, and E. Weigle, "The Bladed Beowulf: A Cost-Effective Alternative to Traditional Beowulfs," Proceedings of IEEE International Conference on Cluster Computing (CLUSTER'02), Chicago, Illinois, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. X. Feng, R. Ge, and K. W. Cameron, "Power and Energy Profiling of Scientific Applications on Distributed Systems," Proceedings of 19th International Parallel and Distributed Processing Symposium (IPDPS 05), Denver, CO, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. V. Freeh and D. K. Lowenthal, "Using Multiple Energy Gears in MPI Programs on a Power-Scalable Cluster," Proceedings of 10th AMC Symposium on Principles and Practice of Parallel Programming (PPOPP), Chicago, IL, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. J. Haas and P. Vogt, "Fully-buffered DIMM Technology Moves Enterprise Platforms to the Next Level," in Technology@Intel Magazine, vol. 3, 2005.Google ScholarGoogle Scholar
  12. C.-H. Hsu and U. Kremer, "Compiler-directed dynamic voltage scaling for memory-bound applications," Department of Computer Science Rutgers University, Piscataway August 2002.Google ScholarGoogle Scholar
  13. H. Huang, C. Lefurgy, T. Keller, and K. G. Shin, "Memory Traffic Reshaping for Energy-Efficient Memory," Proceedings of International Symposium on Low Power Electronics and Design (ISLPED), San Diego, CA, 2005.Google ScholarGoogle Scholar
  14. A. R. Lebeck, X. Fan, H. Zeng, and C. Ellis, "Power Aware Page Allocation," Proceedings of ASPLOS--IX, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. C. Lefurgy, K. Rajamani, et al., "Energy Management for Commercial Servers," IEEE Computer, vol. 36, pp. 39--48, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. P. Mochel, "The sysfs Filesystem," Proceedings of Annual Linux Symposium, Ottawa, Canada, 2005.Google ScholarGoogle Scholar
  17. Rambus, "Rambus RDRAM," 1999.Google ScholarGoogle Scholar
  18. M. E. Tolentino, "Flexible Operating System Structure for Dynamic Memory Management," M.S. Thesis in Computer Science: University of Washington, 2004, pp. 88.Google ScholarGoogle Scholar
  19. M. E. Tolentino, J. Turner, and K. W. Cameron, "An Implementation of Page Allocation Shaping for Energy Efficiency," Proceedings of 3rd Workshop on High-Performance, Power-Aware Computing, Long Beach, CA, 2007.Google ScholarGoogle Scholar

Index Terms

  1. Memory-miser: a performance-constrained runtime system for power-scalable clusters

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      CF '07: Proceedings of the 4th international conference on Computing frontiers
      May 2007
      300 pages
      ISBN:9781595936837
      DOI:10.1145/1242531

      Copyright © 2007 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 7 May 2007

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • Article

      Acceptance Rates

      Overall Acceptance Rate240of680submissions,35%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader