This document summarizes basic communication operations for parallel computing including:
- One-to-all broadcast and all-to-one reduction which involve sending a message from one processor to all others or combining messages from all processors to one.
- All-to-all broadcast and reduction where all processors simultaneously broadcast or reduce messages.
- Collective operations like all-reduce and prefix-sum which combine messages from all processors using associative operators.
- Examples of implementing these operations on different network topologies like rings, meshes and hypercubes are presented along with analyzing their communication costs. The document provides an overview of fundamental communication patterns in parallel computing.