Nicholas Bacon, Patrick G. Bridges, Scott Levy, Kurt B. Ferreira, and Amanda Bienz. 2023. Evaluating the Viability of LogGP for Modeling MPI Performance with Non-contiguous Datatypes on Modern Architectures. In Proceedings of the 30th European MPI Users’ Group Meeting.
Amanda Bienz, Derek Schafer, Anthony Skjellum. 2023. MPI Advance: Open-Source Message Passing Optimizations. EuroMPI’23: 30th European MPI Users' Group Meeting. https://eurompi23.github.io/assets/papers/EuroMPI23_paper_33.pdf
Amanda Bienz. 2023. Benchmarking and Optimizing Data Movement on Emerging Heterogeneous Architectures. 2023 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPWS), pages 611-614. https://ieeexplore.ieee.org/document/10196592
Pouya Haghi, William Krska, Cheng Tan, Tong Geng, Po Hao Chen, Connor Greenwood, Anqi Guo, Thomas Hines, Chunshu Wu, and Ang Li. 2023. FLASH: FPGA-Accelerated Smart Switches with GCN Case Study. Proceedings of the 37th International Conference on Supercomputing, p. 450-462.
Pouya Haghi, Ryan Marshall, Po Hao Chen, Anthony Skjellum, and Martin Herbordt. 2023. A Survey of Potential MPI Complex Collectives: Large-Scale Mining and Analysis of HPC Applications. arXiv e-Print https://arxiv.org/abs/2305.19946
Shelby Lockhart, Amanda Bienz, William Gropp, and Luke Olson. 2023. Performance Analysis and Optimal Node-aware Communication for Enlarged Conjugate Gradient Methods. ACM Trans. Parallel Comput. 10, 1, Article 2 (March 2023), 25 pages. https://doi.org/10.1145/3580003
Shelby Lockhart, Amanda Bienz, William D. Gropp, and Luke N. Olson. 2023. Characterizing the performance of node-aware strategies for irregular point-to-point communication on heterogeneous architectures, Parallel Computing, Volume 116, 103021. https://doi.org/10.1016/j.parco.2023.103021.
William Pepper Marts, Matthew G. F. Dosanjh, Whit Schonbein, Scott Levy, and Patrick G. Bridges. 2023. Measuring Thread Timing to Assess the Feasibility of Early-bird Message Delivery. In Proceedings of the 52nd International Conference on Parallel Processing Workshops (ICPP Workshops '23). Association for Computing Machinery, New York, NY, USA, 119–126. https://doi.org/10.1145/3605731.3605884
W. Pepper Marts, Andrew Worley, Prema Soundarajan, Derek Schafer, Matthew G. F. Dosanjh, Ryan E. Grant, Purushotham V. Bangalore, Anthony Skjellum, and Sheikh K. Ghafoor. 2023. Design of a portable implementation of partitioned point-to-point communication primitives. Concurrency and Computation: Practice and Experience, 35(20):e7655. https://doi.org/10.1002/cpe.7655.
Evelyn Namugwanya, Amanda Bienz, Derek Schafer, and Anthony Skjellum. 2023. Collective-Optimized FFTs. arXiv e-Print https://arxiv.org/pdf/2306.16589.pdf
Grace Nansamba, Amani Altarawneh, and Anthony Skjellum. 2023 A Fault-Model-Relevant Classification of Consensus Mechanisms for MPI and HPC. International Journal of Parallel Programming, V51, #2-3, p. 128-149.
David Walker and Anthony Skjellum. 2023. The Impact of Space-Filling Curves on Data Movement in Parallel Systems. arXiv e-Print https://arxiv.org/pdf/2307.07828.pdf
David Walker. 2023. Algorithms for 3D Hilbert Encoding and Decoding. arXiv e-Print. https://arxiv.org/pdf/2308.05673.pdf
Gerald Collom, Rui Peng Li, and Amanda Bienz. 2023. Optimizing Irregular Communication with Neighborhood Collectives and Locality-Aware Parallelism. Accepted to ExaMPI 2023 Workshop at SC23. https://arxiv.org/abs/2306.01876
Evan D. Suggs, Stephen D. Olivier, Jan Ciesko, and Anthony Skjellum. 2023. View-aware Message Passing Through the Integration of Kokkos and ExaMPI. EuroMPI 2023, Bristol, UK.
Amanda Bienz. What Parallel Performance Really Looks Like. 2023. LANL HPC Summer School, Los Alamos, New Mexico. (slides)
Amanda Bienz. ASHES workshop. 2023. International Parallel & Distributed Processing Symposium, St. Petersburg, Florida. (slides)
Gerald Collom. Optimizing Hypre Communication with Node Aware Parallelism. 2023. SIAM CSE23, Amsterdam, Netherlands. (slides)
William Pepper Marts. Partitioned Communication and Message Aggregation (Poster). 2023. EuroMPI’23: 30th European MPI Users' Group Meeting. Bristol, United Kingdom.
Derek Schafer. Leveraging Modern MPI+GPU Communication Strategies. 2023. SIAM CSE23 Amsterdam, Netherlands.
Anthony Skjellum. 2023. MPI Advance: Open-Source Message Passing Optimizations. EuroMPI 2023, Bristol, UK.
Evan Suggs. 2023. View-aware Message Passing Through the Integration of Kokkos and ExaMPI. EuroMPI 2023, Bristol, UK.
“Beatnik: A prototype High Performance Parallel Interface Benchmark, Software Release 1.0.” Patrick G. Bridges, Jason Stewart, Jacob McCullough, Jered Dominguez-Trujillo, and Thomas Hines. 2023. https://github.com/CUP-ECS/beatnik.
“MPIPCL: MPI Partitioned Communication Library, Software Release 1.0.” Derek Schafer, Andrew Worely, Puri Bangalore, Sheikh Ghafoor, and Tony Skjellum. 2022. https://github.com/mpi-advance/MPIPCL/releases/tag/v1.0.0
“MPIPCL: MPI Partitioned Communication Library, Software Release 1.1.1.” Derek Schafer, Puri Bangalore, and Tony Skjellum. 2023. https://github.com/mpi-advance/MPIPCL/releases/tag/v1.1.1
Bienz, S. Gautam and A. Kharel. 2022. A Locality-Aware Bruck Allgather. Proceedings of 2022 EuroMPI conference, Chattanooga, Tennessee.
Po Hao Chen, Pouya Haghi, Jae Yoon Chung, Tong Geng, Richard West, Anthony Skjellum, and Martin C. Herbordt. 2022. The Viability of Using Online Prediction to Perform Extra Work while Executing BSP Applications. 2022 IEEE High Performance Extreme Computing Conference (HPEC), p. 1-7.
P. Haghi, Guo, A, Xiong, Q, et al. 2022. Reconfigurable switches for high performance and flexible MPI collectives. Concurrency Computat Pract Exper., 34(6):e6769. 10.1002/cpe.6769
B. E. Romero, S. Poroseva, P. Vorobieff, and J. Reisner. 2022. Three-Dimensional Simulations of a Shock-Gas Column Interaction. AIAA SCITECH 2022 Forum, p. 1072. https://doi.org/10.2514/6.2022-1072
B. E. Romero. 2022. FIESTA and Shock-Driven Flows. PhD diss. University of New Mexico.
Bienz, L. N. Olson, W. D. Gropp and S. Lockhart. 2021. Modeling Data Movement Performance on Heterogeneous Architectures. 2021 IEEE High Performance Extreme Computing Conference (HPEC), pp. 1-7. https://doi.org/10.1109/HPEC49654.2021.9622742
M. G.F. Dosanjh, A. Worley, D. Schafer, P. Soundararajan, S. Ghafoor, A. Skjellum, P. V. Bangalore, R. E. Grant. 2021. Implementation and evaluation of MPI 4.0 partitioned communication libraries, Parallel Computing, Volume 108. https://doi.org/10.1016/j.parco.2021.102827
S. Ghosh, et al. 2021. Towards Modern C++ Language Support for MPI. In 2021 Workshop on Exascale MPI (ExaMPI), St. Louis, MO, USA, pp. 27-35. https://doi.org/10.1109/ExaMPI54564.2021.00009
D. Holmes, et al. 2021. Partitioned Collective Communication. In 2021 Workshop on Exascale MPI (ExaMPI), St. Louis, MO, USA, 2021 pp. 9-17. https://doi.org/10.1109/ExaMPI54564.2021.00007
W. P. Marts, M. G. F. Dosanjh, S. Levy, W. Schonbein, R. E. Grant and P. G. Bridges. 2021. MiniMod: A Modular Miniapplication Benchmarking Framework for HPC. 2021 IEEE International Conference on Cluster Computing (CLUSTER), pp. 12-22, https://doi.org/10.1109/Cluster48925.2021.00028
B. E. Romero, S. Poroseva, P. Vorobieff, and J. Reisner. 2021. Comparison of 2D and 3D Simulations of a Shock Accelerated Inclined Gas Column. APS Division of Fluid Dynamics Meeting Abstracts, pp. P10-012.
D. Schafer, T. Hines, E. D. Suggs, M. Rüfenacht and A. Skjellum. 2021. Overlapping Communication and Computation with ExaMPI's Strong Progress and Modern C++ Design. 2021 Workshop on Exascale MPI (ExaMPI), pp. 18-26, https://doi.org/10.1109/ExaMPI54564.2021.00008