ARMCI: a portable remote memory copy library for distributed array libraries and compiler run-time systems The purpose of the Aggregate Remote Memory Copy (ARMCI) library is to provide a general-purpose, efficient, and widely portable remote memory access (RMA) operations (one-sided communication) optimized for contiguous and noncontiguous (strided, scatter/gather, I/O vector) data transfers. In addition, ARMCI includes a set of atomic and mutual exclusion operations. The development ARMCI is driven by the need to support the global-addres space communication model in context of distributed regular or irregular distributed data structures, communication libraries, and compilers. ARMCI is a standalone system that could be used to support user-level libraries and applications that use MPI or PVM. Development of ARMCI has been supported by the DoE2000 Global Arrays project, EMSL, and recently by the Center for Programming Models for Scalable Parallel Computing.

References in zbMATH (referenced in 14 articles )

Showing results 1 to 14 of 14.
Sorted by year (citations)

  1. Hori, Atsushi; Lee, Jinpil; Sato, Mitsuhisa: Audit: a new synchronization API for the GET/PUT protocol (2012) ioport
  2. Lu, Qingda; Gao, Xiaoyang; Krishnamoorthy, Sriram; Baumgartner, Gerald; Ramanujam, J.; Sadayappan, P.: Empirical performance model-driven data layout optimization and library call selection for tensor contraction expressions (2012) ioport
  3. Yu, Weikuan; Que, Xinyu; Tipparaju, Vinod; Vetter, Jeffrey S.: Hicoo: hierarchical cooperation for scalable communication in global address space programming models on Cray XT systems (2012) ioport
  4. Sottile, Matthew J; Mattson, Timothy G; Rasmussen, Craig E.: Introduction to concurrency in programming languages. (2010)
  5. Bader, Michael; Hanigk, Sebastian; Huckle, Thomas: Parallelisation of block-recursive matrix multiplication in prefix computations (2008)
  6. He, Jian; Verstak, Alex; Watson, Layne T.; Sosonkina, Masha: Design and implementation of a massively parallel version of DIRECT (2008)
  7. Shet, Aniruddha G.; Sadayappan, P.; Bernholdt, David E.; Nieplocha, Jarek; Tipparaju, Vinod: A framework for characterizing overlap of communication and computation in parallel applications (2008) ioport
  8. Coarfa, Cristian; Dotsenko, Yuri; Mellor-Crummey, John: Experiences with sweep3D implementations in co-array fortran (2006) ioport
  9. Coarfa, Cristian; Dotsenko, Yuri; Eckhardt, Jason; Mellor-Crummey, John: Co-array Fortran performance and potential: An NPB experimental study (2004)
  10. Rasmussen, Craig E.; Sottile, Matthew J.; Nieplocha, Jarek; Numrich, Robert W.; Jones, Eric: Co-array Python: A parallel extension to the Python language (2004)
  11. Alexeev, Yuri; Kendall, Ricky A.; Gordon, Mark S.: The distributed data SCF (2002)
  12. Barker, Kevin; Chrisochoides, Nikos; Dobbelaere, Jeffrey; Nave, Démian; Pingali, Keshav: Date movement and control substrate for parallel adaptive applications (2002)
  13. Bernholdt, D. E.: Scalability of correlated electronic structure calculations on parallel computers: A case study of the RI-MP2 method (2000)
  14. Nobes, R. H.; Rendell, A. P.; Nieplocha, J.: Computational chemistry on Fujitsu vector--parallel processors: Hardware and programming environment (2000)

Further publications can be found at: