aboutsummaryrefslogtreecommitdiff
path: root/content/pipelines.article
diff options
context:
space:
mode:
Diffstat (limited to 'content/pipelines.article')
-rw-r--r--content/pipelines.article306
1 files changed, 306 insertions, 0 deletions
diff --git a/content/pipelines.article b/content/pipelines.article
new file mode 100644
index 0000000..ab5d1ce
--- /dev/null
+++ b/content/pipelines.article
@@ -0,0 +1,306 @@
+Go Concurrency Patterns: Pipelines and cancellation
+13 Mar 2014
+Tags: concurrency, pipelines, cancellation
+
+Sameer Ajmani
+
+* Introduction
+
+Go's concurrency primitives make it easy to construct streaming data pipelines
+that make efficient use of I/O and multiple CPUs. This article presents
+examples such pipelines, highlights subtleties that arise when operations fail,
+and introduces techniques for dealing with failures cleanly.
+
+* What is a pipeline?
+
+There's no formal definition of a pipeline in Go; it's just one of many kinds of
+concurrent programs. Informally, a pipeline is a series of _stages_ connected
+by channels, where each stage is a group of goroutines running the same
+function. In each stage, the goroutines
+
+- receive values from _upstream_ via _inbound_ channels
+- perform some function on that data, usually producing new values
+- send values _downstream_ via _outbound_ channels
+
+Each stage has any number of inbound and outbound channels, except the
+first and last stages, which have only outbound or inbound channels,
+respectively. The first stage is sometimes called the _source_ or
+_producer_; the last stage, the _sink_ or _consumer_.
+
+We'll begin with a simple example pipeline to explain the ideas and techniques.
+Later, we'll present a more realistic example.
+
+* Squaring numbers
+
+Consider a pipeline with three stages.
+
+The first stage, `gen`, is a function that converts a list of integers to a
+channel that emits the integers in the list. The `gen` function starts a
+goroutine that sends the integers on the channel and closes the channel when all
+the values have been sent:
+
+.code pipelines/square.go /func gen/,/^}/
+
+The second stage, `sq`, receives integers from a channel and returns a
+channel that emits the square of each recevied integer. After the
+inbound channel is closed and this stage has sent all the values
+downstream, it closes the outbound channel:
+
+.code pipelines/square.go /func sq/,/^}/
+
+The `main` function sets up the pipeline and runs the final stage: it receives
+values from the second stage and prints each one, until the channel is closed:
+
+.code pipelines/square.go /func main/,/^}/
+
+Since `sq` has the same type for its inbound and outbound channels, we
+can compose it any number of times. We can also rewrite `main` as a
+range loop, like the other stages:
+
+.code pipelines/square2.go /func main/,/^}/
+
+* Fan-out, fan-in
+
+Multiple functions can read from the same channel until that channel is closed;
+this is called _fan-out_. This provides a way to distribute work amongst a group
+of workers to parallelize CPU use and I/O.
+
+A function can read from multiple inputs and proceed until all are closed by
+multiplexing the input channels onto a single channel that's closed when all the
+inputs are closed. This is called _fan-in_.
+
+We can change our pipeline to run two instances of `sq`, each reading from the
+same input channel. We introduce a new function, _merge_, to fan in the
+results:
+
+.code pipelines/sqfan.go /func main/,/^}/
+
+The `merge` function converts a list of channels to a single channel by starting
+a goroutine for each inbound channel that copies the values to the sole outbound
+channel. Once all the `output` goroutines have been started, `merge` starts one
+more goroutine to close the outbound channel after all sends on that channel are
+done.
+
+Sends on a closed channel panic, so it's important to ensure all sends
+are done before calling close. The
+[[http://golang.org/pkg/sync/#WaitGroup][`sync.WaitGroup`]] type
+provides a simple way to arrange this synchronization:
+
+.code pipelines/sqfan.go /func merge/,/^}/
+
+* Stopping short
+
+There is a pattern to our pipeline functions:
+
+- stages close their outbound channels when all the send operations are done.
+- stages keep receiving values from inbound channels until those channels are closed.
+
+This pattern allows each receving stage to be written as a `range` loop and
+ensures that all goroutines exit once all values have been successfully sent
+downstream.
+
+But in real pipelines, stages don't always receive all the inbound
+values. Sometimes this is by design: the receiver may only need a
+subset of values to make progress. More often, a stage exits early
+because an inbound value represents an error in an earlier stage. In
+either case the receiver should not have to wait for the remaining
+values to arrive, and we want earlier stages to stop producing values
+that later stages don't need.
+
+In our example pipeline, if a stage fails to consume all the inbound values, the
+goroutines attempting to send those values will block indefinitely:
+
+.code pipelines/sqleak.go /first value/,/^}/
+
+This is a resource leak: goroutines consume memory and runtime resources, and
+heap references in goroutine stacks keep data from being garbage collected.
+Goroutines are not garbage collected; they must exit on their own.
+
+We need to arrange for the upstream stages of our pipeline to exit even when the
+downstream stages fail to receive all the inbound values. One way to do this is
+to change the outbound channels to have a buffer. A buffer can hold a fixed
+number of values; send operations complete immediately if there's room in the
+buffer:
+
+ c := make(chan int, 2) // buffer size 2
+ c <- 1 // succeeds immediately
+ c <- 2 // succeeds immediately
+ c <- 3 // blocks until another goroutine does <-c and receives 1
+
+When the number of values to be sent is known at channel creation time, a buffer
+can simplify the code. For example, we can rewrite `gen` to copy the list of
+integers into a buffered channel and avoid creating a new goroutine:
+
+.code pipelines/sqbuffer.go /func gen/,/^}/
+
+Returning to the blocked goroutines in our pipeline, we might consider adding a
+buffer to the outbound channel returned by `merge`:
+
+.code pipelines/sqbuffer.go /func merge/,/unchanged/
+
+While this fixes the blocked goroutine in this program, this is bad code. The
+choice of buffer size of 1 here depends on knowing the number of values `merge`
+will receive and the number of values downstream stages will consume. This is
+fragile: if we pass an additional value to `gen`, or if the downstream stage
+reads any fewer values, we will again have blocked goroutines.
+
+Instead, we need to provide a way for downstream stages to indicate to the
+senders that they will stop accepting input.
+
+* Explicit cancellation
+
+When `main` decides to exit without receiving all the values from
+`out`, it must tell the goroutines in the upstream stages to abandon
+the values they're trying it send. It does so by sending values on a
+channel called `done`. It sends two values since there are
+potentially two blocked senders:
+
+.code pipelines/sqdone1.go /func main/,/^}/
+
+The sending goroutines replace their send operation with a `select`
+statement that proceeds either when the send on `out` happens or when
+they receive a value from `done`. The value type of `done` is the
+empty struct because the value doesn't matter: it is the receive event
+that indicates the send on `out` should be abandoned. The `output`
+goroutines continue looping on their inbound channel, `c`, so the
+upstream stages are not blocked:
+
+.code pipelines/sqdone1.go /func merge/,/unchanged/
+
+But this approach has a problem: the downstream receiver needs to know the
+number of potentially blocked upstream senders. Keeping track of these counts
+is tedious and error-prone.
+
+We need a way to tell an unknown and unbounded number of goroutines to
+stop sending their values downstream. In Go, we can do this by
+closing a channel, because
+[[http://golang.org/ref/spec#Receive_operator][a receive operation on a closed channel can always proceed immediately, yielding the element type's zero value.]]
+
+This means that `main` can unblock all the senders simply by closing
+the `done` channel. This close is effectively a broadcast signal to
+the senders. We extend each of our pipeline functions to accept
+`done` as a parameter and arrange for the close to happen via a
+`defer` statement, so that all return paths from `main` will signal
+the pipeline stages to exit.
+
+.code pipelines/sqdone3.go /func main/,/^}/
+
+Each of our pipeline stages is now free to return early: `sq` can return from
+the middle of its loop since we know that if `done` is closed here, it is also
+closed for the upstream `gen` stage. `sq` ensures its `out` channel is closed
+on all return paths via a `defer` statement.
+
+.code pipelines/sqdone3.go /func sq/,/^}/
+
+Here are the guidlines for pipeline construction:
+
+- stages close their outbound channels when all the send operations are done.
+- stages keep receiving values from inbound channels until those channels are closed or the senders are unblocked.
+
+Pipelines unblock senders either by ensuring there's enough buffer for all the
+values that are sent or by explicitly signalling senders when the receiver may
+abandon the channel.
+
+* Digesting a tree
+
+Let's consider a more realistic pipeline.
+
+MD5 is a message-digest algorithm that's useful as a file checksum. The command
+line utility `md5sum` prints digest values for a list of files.
+
+ % md5sum *.go
+ d47c2bbc28298ca9befdfbc5d3aa4e65 bounded.go
+ ee869afd31f83cbb2d10ee81b2b831dc parallel.go
+ b88175e65fdcbc01ac08aaf1fd9b5e96 serial.go
+
+Our example program is like `md5sum` but instead takes a single directory as an
+argument and prints the digest values for each regular file under that
+directory, sorted by path name.
+
+ % go run serial.go .
+ d47c2bbc28298ca9befdfbc5d3aa4e65 bounded.go
+ ee869afd31f83cbb2d10ee81b2b831dc parallel.go
+ b88175e65fdcbc01ac08aaf1fd9b5e96 serial.go
+
+The main function of our program invokes a helper function `MD5All`, which
+returns a map from path name to digest value, then sorts and prints the results:
+
+.code pipelines/serial.go /func main/,/^}/
+
+The `MD5All` function is the focus of our discussion. In
+[[pipelines/serial.go][serial.go]], the implementation uses no concurrency and
+simply reads and sums each file as it walks the tree.
+
+.code pipelines/serial.go /MD5All/,/^}/
+
+* Parallel digestion
+
+In [[pipelines/parallel.go][parallel.go]], we split `MD5All` into a two-stage
+pipeline. The first stage, `sumFiles`, walks the tree, digests each file in
+a new goroutine, and sends the results on a channel with value type `result`:
+
+.code pipelines/parallel.go /type result/,/}/ HLresult
+
+`sumFiles` returns two channels: one for the `results` and another for the error
+returned by `filepath.Walk`. The walk function starts a new goroutine to
+process each regular file, then checks `done`. If `done` is closed, the walk
+stops immediately:
+
+.code pipelines/parallel.go /func sumFiles/,/^}/
+
+`MD5All` receives the digest values from `c`. `MD5All` returns early on error,
+closing `done` via a `defer`:
+
+.code pipelines/parallel.go /func MD5All/,/^}/ HLdone
+
+* Bounded parallelism
+
+The `MD5All` implementation in [[pipelines/parallel.go][parallel.go]]
+starts a new goroutine for each file. In a directory with many large
+files, this may allocate more memory than is available on the machine.
+
+We can limit these allocations by bounding the number of files read in
+parallel. In [[pipelines/bounded.go][bounded.go]], we do this by
+creating a fixed number of goroutines for reading files. Our pipeline
+now has three stages: walk the tree, read and digest the files, and
+collect the digests.
+
+The first stage, `walkFiles`, emits the paths of regular files in the tree:
+
+.code pipelines/bounded.go /func walkFiles/,/^}/
+
+The middle stage starts a fixed number of `digester` goroutines that receive
+file names from `paths` and send `results` on channel `c`:
+
+.code pipelines/bounded.go /func digester/,/^}/ HLpaths
+
+Unlike our previous examples, `digester` does not close its output channel, as
+multiple goroutines are sending on a shared channel. Instead, code in `MD5All`
+arranges for the channel to be closed when all the `digesters` are done:
+
+.code pipelines/bounded.go /fixed number/,/End of pipeline/ HLc
+
+We could instead have each digester create and return its own output
+channel, but then we would need additional goroutines to fan-in the
+results.
+
+The final stage receives all the `results` from `c` then checks the
+error from `errc`. This check cannot happen any earlier, since before
+this point, `walkFiles` may block sending values downstream:
+
+.code pipelines/bounded.go /m := make/,/^}/ HLerrc
+
+* Conclusion
+
+This article has presented techniques for constructing streaming data pipelines
+in Go. Dealing with failures in such pipelines is tricky, since each stage in
+the pipeline may block attempting to send values downstream, and the downstream
+stages may no longer care about the incoming data. We showed how closing a
+channel can broadcast a "done" signal to all the goroutines started by a
+pipeline and defined guidelines for constructing pipelines correctly.
+
+Further reading:
+
+- [[http://talks.golang.org/2012/concurrency.slide#1][Go Concurrency Patterns]] ([[https://www.youtube.com/watch?v=f6kdp27TYZs][video]]) presents the basics of Go's concurrency primitives and several ways to apply them.
+- [[http://blog.golang.org/advanced-go-concurrency-patterns][Advanced Go Concurrency Patterns]] ([[http://www.youtube.com/watch?v=QDDwwePbDtw][video]]) covers more complex uses of Go's primitives, especially `select`.
+- Douglas McIlroy's paper [[http://swtch.com/~rsc/thread/squint.pdf][Squinting at Power Series]] shows how Go-like concurrency provides elegant support for complex calculations.