This document discusses Riak Pipe, which allows distributing work across Riak clusters. Riak Pipe uses a dataflow-based approach where work is split into phases or "fittings" connected by queues. Work items flow from fitting to fitting, with each fitting potentially running on different nodes. This allows scaling out processing of large datasets. Common uses of Riak Pipe include Riak's MapReduce functionality and calculating tenant usage metrics across clusters. The document covers concepts like vnodes, workers, and failure handling, as well as existing fittings and tools for monitoring and debugging distributed pipelines.