Abstract
Different programming paradigms utilize a variety of collective communication operations, often with different semantics. We present the component collective messaging interface (CCMI), that can support asynchronous non-blocking collectives and is extensible to different programming paradigms and architectures. CCMI is designed with components written in the C++ programming language, allowing it to have reuse and extendability. Collective algorithms are embodied in topological schedules and executors that execute them. Portability across architectures is enabled by the multisend data movement component. CCMI includes a programming language adaptor used to implement different APIs with different semantics for different paradigms. We study the effectiveness of CCMI on Blue Gene/P and evaluate its performance for the barrier, broadcast, and allreduce collective operations. We also present the performance of the barrier collective on the Abe Infiniband cluster.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Forum, M.P.I.: MPI-2: Extensions to the message-passing interface (1997), http://www.mpi-forum.org/docs/mpi-20-html/mpi2-report.html
Hoefler, T., Lumsdaine, A., Rehm, W.: Implementation and Performance Analysis of Non-Blocking Collective Operations for MPI. In: Proceedings of SC 2007. IEEE Computer Society/ACM (2007)
Nishtala, R., Almasi, G., Cascaval, C.: Performance without Pain = Productivity. In: Proceedings of PPOPP 2008 (2008)
Kale, L.V., Krishnan, S.: Charm++: Parallel Programming with Message-Driven Objects. In: Wilson, G.V., Lu, P. (eds.) Parallel Programming using C++, pp. 175–213. MIT Press, Cambridge (1996)
Department of Computer Science,University of Illinois at Urbana-Champaign, Urbana, IL, The CHARM (5.9) Programming language manual (2007)
IBM Blue Gene Team, Overview of the Blue Gene/P project, IBM J. Res. Dev., vol. 52 (January 2008), http://www.research.ibm.com/journal/rd/521/team.html
NCSA Abe Cluster, http://www.ncsa.uiuc.edu/UserInfo/Resources/Hardware/Intel64Cluster
Gara, A., Blumrich, M.A., Chen, D., Chiu, G.L.-T., Coteus, P., Giampapa, M.E., Haring, R.A., Heidelberger, P., Hoenicke, D., Kopcsay, G.V., Liebsch, T.A., Ohmacht, M., Steinmacher-Burow, B.D., Takken, T., Vranas, P.: Overview of the Blue Gene/L System Architecture. IBM Journal of Research and Development 49(2/3), 195–212 (2005)
Petrini, F., chun Feng, W., Hoisie, A., Coll, S., Frachtenberg, E.: The quadrics network: high-performance clustering technology. IEEE Micro. 22(1), 46–57 (2002)
Hensgen, D., Finkel, R., Manber, U.: Two algorithms for barrier synchrofization. International Journal of Parallel Programming 17(1) (1988)
Thakur, R., Rabenseifner, R., Gropp, W.: Optimization of Collective Communication Operations in MPICH. International Journal of High Performance Computing Applications 19, 49–66 (2005)
Almasi, G., et al.: Design and implementation of message-passing services for the Blue Gene/L supercomputer. IBM J. Res. Dev. 49, 393–406 (2005)
Almasi, G., Archer, C., Erway, C., Heidelberger, P., Martorell, X., Moreira, J., Steinmacher-Burow, B., Zheng, Y.: Optimization of MPI Collective Communication on BlueGene/L Systems. In: Proceedings of the 19th annual international conference on Supercomputing ICS 2005, pp. 253–262. ACM Press, New York (2005)
Kumar, S., Dozsa, G., Almasi, G., Chen, D., Heidelberger, P., Giampapa, M.E., Blocksome, M., Faraj, A., Parker, J., Ratterman, J., Smith, B., Archer, C.: The Deep Computing Messaging Framework: Generalized Scalable Message passing on the BlueGene/P Supercomputer. In: International Conference on Supercomputing ICS 2008 (to appear, 2008)
DCMF (2008), http://dcmf.anl-external.org/wiki .
Huang, W., Santhanaraman, G., Jin, H.-W., Gao, Q., Panda, D.K.x.D.K.: Design of High Performance MVAPICH2: MPI2 over InfiniBand. In: CCGRID 2006: Proceedings of the Sixth IEEE International Symposium on Cluster Computing and the Grid, pp. 43–48. IEEE Computer Society, Washington (2006)
Almasi, G., Dozsa, G., Erway, C.C., Steinmacher-Burow, B.D.: Efficient Implementation of Allreduce on BlueGene/L Collective Network. In: Proceedings of Euro PVM/MPI, pp. 57–66 (2005)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2008 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Kumar, S. et al. (2008). Architecture of the Component Collective Messaging Interface. In: Lastovetsky, A., Kechadi, T., Dongarra, J. (eds) Recent Advances in Parallel Virtual Machine and Message Passing Interface. EuroPVM/MPI 2008. Lecture Notes in Computer Science, vol 5205. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-87475-1_10
Download citation
DOI: https://doi.org/10.1007/978-3-540-87475-1_10
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-87474-4
Online ISBN: 978-3-540-87475-1
eBook Packages: Computer ScienceComputer Science (R0)