Nothing Special   »   [go: up one dir, main page]

skip to main content
10.5555/2872599.2872611acmconferencesArticle/Chapter ViewAbstractPublication Pageshp3cConference Proceedingsconference-collections
research-article

Throughput studies on an InfiniBand interconnect via all-to-all communications

Published: 12 April 2015 Publication History

Abstract

Distributed-memory clusters are the most important type of parallel computer today, and they dominate the TOP500 list. The InfiniBand interconnect is the most popular network for distributed-memory compute clusters. Contention of communications across a switched network that connects multiple compute nodes in a distributed-memory cluster may seriously degrade performance of parallel code. This contention is maximized when communicating large blocks of data among all parallel processes simultaneously. This communication pattern arises in many important algorithms such as parallel sorting. The cluster tara in the UMBC High Performance Computing Facility (HPCF) with a quad-data rate InfiniBand interconnect provides an opportunity to test if the capacity of a switched network can become a limiting factor in algorithmic performance. We find that we can design a test case of a problem involving increasing usage of memory that does not scale any more on the InifiniBand interconnect, thus becoming a limiting factor for parallel scalability. However, for the case of stable memory usage of the problem, the InifiniBand communications get faster and will not inhibit parallel scalability. The tests in this paper are designed to involve only basic MPI commands for wide reproducibility, and the paper provides the detailed motivation of the design of the memory usage needed for the tests.

References

[1]
Balaji, P., Buntinas, D., Goodell, D., Gropp, W., Kumar, S., Lusk, E., Thakur, R., and Träff, J. L. MPI on a million processors. In Recent Advances in Parallel Virtual Machine and Message Passing Interface. Springer, 2009, 20--30.
[2]
Balaji, P., Narravula, S., Vaidyanathan, K., Krishnamoorthy, S., Wu, J., and Panda, D. K. Sockets direct protocol over InfiniBand in clusters: Is it beneficial? In Performance Analysis of Systems and Software, 2004 IEEE International Symposium on-ISPASS, IEEE (2004), 28--35.
[3]
Hoefler, T., Lumsdaine, A., and Rehm, W. Implementation and performance analysis of non-blocking collective operations for MPI. In Supercomputing, 2007. SC'07. Proceedings of the 2007 ACM/IEEE Conference on, IEEE (2007), 1--10.
[4]
Kandalla, K., Mancini, E. P., Sur, S., and Panda, D. K. Designing power-aware collective communication algorithms for InfiniBand clusters. In Parallel Processing (ICPP), 2010 39th International Conference on, IEEE (2010), 218--227.
[5]
Kandalla, K., Subramoni, H., Tomko, K., Pekurovsky, D., and Panda, D. A novel functional partitioning approach to design high-performance MPI-3 non-blocking Alltoallv collective on multi-core systems. In 42nd International Conference on Parallel Processing (ICPP) 2013, IEEE (2013), 611--620.
[6]
Kandalla, K., Subramoni, H., Tomko, K., Pekurovsky, D., Sur, S., and Panda, D. K. High-performance and scalable non-blocking all-to-all with collective offload on InfiniBand clusters: a study with parallel 3D FFT. Computer Science-Research and Development 26, 3--4 (2011), 237--246.
[7]
Mistry, N., Ramsey, J., Wiley, B., Yanchuck, J., Huang, X., Gobbert, M. K., Mineo, C., and Mountain, D. Contention of communications in switched networks with applications to parallel sorting. Tech. Rep. HPCF-2013-13, UMBC High Performance Computing Facility, University of Maryland, Baltimore County, 2013.
[8]
Pacheco, P. S. Parallel Programming with MPI. Morgan Kaufmann, 1997.
[9]
Raim, A. M., and Gobbert, M. K. Parallel performance studies for an elliptic test problem on the cluster tara. Tech. Rep. HPCF-2010-2, UMBC High Performance Computing Facility, University of Maryland, Baltimore County, 2010.
[10]
Reuther, A., Funk, A., Kepner, J., McCabe, A., Arcand, W., Currie, T., Hubbell, M., and Michaleas, P. Benchmarking the MIT LL HPCMP DHPI system. In DoD High Performance Computing Modernization Program Users Group Conference, 2007, IEEE (2007), 310--316.
[11]
Thibeault, C. M., Minkovich, K., O'Brien, M. J., Harris Jr, F. C., and Srinivasa, N. Efficiently passing messages in distributed spiking neural network simulation. Frontiers in Computational Neuroscience 7 (2013).
[12]
White, C. M. Data Communications and Computer Networks: A Business User's Approach. Course Technology, 2013.
[13]
Wu, J., Wyckoff, P., and Panda, D. PVFS over InfiniBand: Design and performance evaluation. In Parallel Processing, 2003. Proceedings. 2003 International Conference on, IEEE (2003), 125--132.

Index Terms

  1. Throughput studies on an InfiniBand interconnect via all-to-all communications

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    HPC '15: Proceedings of the Symposium on High Performance Computing
    April 2015
    253 pages
    ISBN:9781510801011

    Sponsors

    Publisher

    Society for Computer Simulation International

    San Diego, CA, United States

    Publication History

    Published: 12 April 2015

    Check for updates

    Author Tags

    1. InfiniBand interconnect
    2. MPI
    3. all-to-all communications
    4. network contention
    5. scalability studies

    Qualifiers

    • Research-article

    Conference

    SpringSim '15
    Sponsor:
    SpringSim '15: 2015 Spring Simulation Multiconference
    April 12 - 15, 2015
    Virginia, Alexandria

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 40
      Total Downloads
    • Downloads (Last 12 months)2
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 14 Nov 2024

    Other Metrics

    Citations

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media