diff options
Diffstat (limited to 'content/pipelines.article')
-rw-r--r-- | content/pipelines.article | 306 |
1 files changed, 306 insertions, 0 deletions
diff --git a/content/pipelines.article b/content/pipelines.article new file mode 100644 index 0000000..ab5d1ce --- /dev/null +++ b/content/pipelines.article @@ -0,0 +1,306 @@ +Go Concurrency Patterns: Pipelines and cancellation +13 Mar 2014 +Tags: concurrency, pipelines, cancellation + +Sameer Ajmani + +* Introduction + +Go's concurrency primitives make it easy to construct streaming data pipelines +that make efficient use of I/O and multiple CPUs. This article presents +examples such pipelines, highlights subtleties that arise when operations fail, +and introduces techniques for dealing with failures cleanly. + +* What is a pipeline? + +There's no formal definition of a pipeline in Go; it's just one of many kinds of +concurrent programs. Informally, a pipeline is a series of _stages_ connected +by channels, where each stage is a group of goroutines running the same +function. In each stage, the goroutines + +- receive values from _upstream_ via _inbound_ channels +- perform some function on that data, usually producing new values +- send values _downstream_ via _outbound_ channels + +Each stage has any number of inbound and outbound channels, except the +first and last stages, which have only outbound or inbound channels, +respectively. The first stage is sometimes called the _source_ or +_producer_; the last stage, the _sink_ or _consumer_. + +We'll begin with a simple example pipeline to explain the ideas and techniques. +Later, we'll present a more realistic example. + +* Squaring numbers + +Consider a pipeline with three stages. + +The first stage, `gen`, is a function that converts a list of integers to a +channel that emits the integers in the list. The `gen` function starts a +goroutine that sends the integers on the channel and closes the channel when all +the values have been sent: + +.code pipelines/square.go /func gen/,/^}/ + +The second stage, `sq`, receives integers from a channel and returns a +channel that emits the square of each recevied integer. After the +inbound channel is closed and this stage has sent all the values +downstream, it closes the outbound channel: + +.code pipelines/square.go /func sq/,/^}/ + +The `main` function sets up the pipeline and runs the final stage: it receives +values from the second stage and prints each one, until the channel is closed: + +.code pipelines/square.go /func main/,/^}/ + +Since `sq` has the same type for its inbound and outbound channels, we +can compose it any number of times. We can also rewrite `main` as a +range loop, like the other stages: + +.code pipelines/square2.go /func main/,/^}/ + +* Fan-out, fan-in + +Multiple functions can read from the same channel until that channel is closed; +this is called _fan-out_. This provides a way to distribute work amongst a group +of workers to parallelize CPU use and I/O. + +A function can read from multiple inputs and proceed until all are closed by +multiplexing the input channels onto a single channel that's closed when all the +inputs are closed. This is called _fan-in_. + +We can change our pipeline to run two instances of `sq`, each reading from the +same input channel. We introduce a new function, _merge_, to fan in the +results: + +.code pipelines/sqfan.go /func main/,/^}/ + +The `merge` function converts a list of channels to a single channel by starting +a goroutine for each inbound channel that copies the values to the sole outbound +channel. Once all the `output` goroutines have been started, `merge` starts one +more goroutine to close the outbound channel after all sends on that channel are +done. + +Sends on a closed channel panic, so it's important to ensure all sends +are done before calling close. The +[[http://golang.org/pkg/sync/#WaitGroup][`sync.WaitGroup`]] type +provides a simple way to arrange this synchronization: + +.code pipelines/sqfan.go /func merge/,/^}/ + +* Stopping short + +There is a pattern to our pipeline functions: + +- stages close their outbound channels when all the send operations are done. +- stages keep receiving values from inbound channels until those channels are closed. + +This pattern allows each receving stage to be written as a `range` loop and +ensures that all goroutines exit once all values have been successfully sent +downstream. + +But in real pipelines, stages don't always receive all the inbound +values. Sometimes this is by design: the receiver may only need a +subset of values to make progress. More often, a stage exits early +because an inbound value represents an error in an earlier stage. In +either case the receiver should not have to wait for the remaining +values to arrive, and we want earlier stages to stop producing values +that later stages don't need. + +In our example pipeline, if a stage fails to consume all the inbound values, the +goroutines attempting to send those values will block indefinitely: + +.code pipelines/sqleak.go /first value/,/^}/ + +This is a resource leak: goroutines consume memory and runtime resources, and +heap references in goroutine stacks keep data from being garbage collected. +Goroutines are not garbage collected; they must exit on their own. + +We need to arrange for the upstream stages of our pipeline to exit even when the +downstream stages fail to receive all the inbound values. One way to do this is +to change the outbound channels to have a buffer. A buffer can hold a fixed +number of values; send operations complete immediately if there's room in the +buffer: + + c := make(chan int, 2) // buffer size 2 + c <- 1 // succeeds immediately + c <- 2 // succeeds immediately + c <- 3 // blocks until another goroutine does <-c and receives 1 + +When the number of values to be sent is known at channel creation time, a buffer +can simplify the code. For example, we can rewrite `gen` to copy the list of +integers into a buffered channel and avoid creating a new goroutine: + +.code pipelines/sqbuffer.go /func gen/,/^}/ + +Returning to the blocked goroutines in our pipeline, we might consider adding a +buffer to the outbound channel returned by `merge`: + +.code pipelines/sqbuffer.go /func merge/,/unchanged/ + +While this fixes the blocked goroutine in this program, this is bad code. The +choice of buffer size of 1 here depends on knowing the number of values `merge` +will receive and the number of values downstream stages will consume. This is +fragile: if we pass an additional value to `gen`, or if the downstream stage +reads any fewer values, we will again have blocked goroutines. + +Instead, we need to provide a way for downstream stages to indicate to the +senders that they will stop accepting input. + +* Explicit cancellation + +When `main` decides to exit without receiving all the values from +`out`, it must tell the goroutines in the upstream stages to abandon +the values they're trying it send. It does so by sending values on a +channel called `done`. It sends two values since there are +potentially two blocked senders: + +.code pipelines/sqdone1.go /func main/,/^}/ + +The sending goroutines replace their send operation with a `select` +statement that proceeds either when the send on `out` happens or when +they receive a value from `done`. The value type of `done` is the +empty struct because the value doesn't matter: it is the receive event +that indicates the send on `out` should be abandoned. The `output` +goroutines continue looping on their inbound channel, `c`, so the +upstream stages are not blocked: + +.code pipelines/sqdone1.go /func merge/,/unchanged/ + +But this approach has a problem: the downstream receiver needs to know the +number of potentially blocked upstream senders. Keeping track of these counts +is tedious and error-prone. + +We need a way to tell an unknown and unbounded number of goroutines to +stop sending their values downstream. In Go, we can do this by +closing a channel, because +[[http://golang.org/ref/spec#Receive_operator][a receive operation on a closed channel can always proceed immediately, yielding the element type's zero value.]] + +This means that `main` can unblock all the senders simply by closing +the `done` channel. This close is effectively a broadcast signal to +the senders. We extend each of our pipeline functions to accept +`done` as a parameter and arrange for the close to happen via a +`defer` statement, so that all return paths from `main` will signal +the pipeline stages to exit. + +.code pipelines/sqdone3.go /func main/,/^}/ + +Each of our pipeline stages is now free to return early: `sq` can return from +the middle of its loop since we know that if `done` is closed here, it is also +closed for the upstream `gen` stage. `sq` ensures its `out` channel is closed +on all return paths via a `defer` statement. + +.code pipelines/sqdone3.go /func sq/,/^}/ + +Here are the guidlines for pipeline construction: + +- stages close their outbound channels when all the send operations are done. +- stages keep receiving values from inbound channels until those channels are closed or the senders are unblocked. + +Pipelines unblock senders either by ensuring there's enough buffer for all the +values that are sent or by explicitly signalling senders when the receiver may +abandon the channel. + +* Digesting a tree + +Let's consider a more realistic pipeline. + +MD5 is a message-digest algorithm that's useful as a file checksum. The command +line utility `md5sum` prints digest values for a list of files. + + % md5sum *.go + d47c2bbc28298ca9befdfbc5d3aa4e65 bounded.go + ee869afd31f83cbb2d10ee81b2b831dc parallel.go + b88175e65fdcbc01ac08aaf1fd9b5e96 serial.go + +Our example program is like `md5sum` but instead takes a single directory as an +argument and prints the digest values for each regular file under that +directory, sorted by path name. + + % go run serial.go . + d47c2bbc28298ca9befdfbc5d3aa4e65 bounded.go + ee869afd31f83cbb2d10ee81b2b831dc parallel.go + b88175e65fdcbc01ac08aaf1fd9b5e96 serial.go + +The main function of our program invokes a helper function `MD5All`, which +returns a map from path name to digest value, then sorts and prints the results: + +.code pipelines/serial.go /func main/,/^}/ + +The `MD5All` function is the focus of our discussion. In +[[pipelines/serial.go][serial.go]], the implementation uses no concurrency and +simply reads and sums each file as it walks the tree. + +.code pipelines/serial.go /MD5All/,/^}/ + +* Parallel digestion + +In [[pipelines/parallel.go][parallel.go]], we split `MD5All` into a two-stage +pipeline. The first stage, `sumFiles`, walks the tree, digests each file in +a new goroutine, and sends the results on a channel with value type `result`: + +.code pipelines/parallel.go /type result/,/}/ HLresult + +`sumFiles` returns two channels: one for the `results` and another for the error +returned by `filepath.Walk`. The walk function starts a new goroutine to +process each regular file, then checks `done`. If `done` is closed, the walk +stops immediately: + +.code pipelines/parallel.go /func sumFiles/,/^}/ + +`MD5All` receives the digest values from `c`. `MD5All` returns early on error, +closing `done` via a `defer`: + +.code pipelines/parallel.go /func MD5All/,/^}/ HLdone + +* Bounded parallelism + +The `MD5All` implementation in [[pipelines/parallel.go][parallel.go]] +starts a new goroutine for each file. In a directory with many large +files, this may allocate more memory than is available on the machine. + +We can limit these allocations by bounding the number of files read in +parallel. In [[pipelines/bounded.go][bounded.go]], we do this by +creating a fixed number of goroutines for reading files. Our pipeline +now has three stages: walk the tree, read and digest the files, and +collect the digests. + +The first stage, `walkFiles`, emits the paths of regular files in the tree: + +.code pipelines/bounded.go /func walkFiles/,/^}/ + +The middle stage starts a fixed number of `digester` goroutines that receive +file names from `paths` and send `results` on channel `c`: + +.code pipelines/bounded.go /func digester/,/^}/ HLpaths + +Unlike our previous examples, `digester` does not close its output channel, as +multiple goroutines are sending on a shared channel. Instead, code in `MD5All` +arranges for the channel to be closed when all the `digesters` are done: + +.code pipelines/bounded.go /fixed number/,/End of pipeline/ HLc + +We could instead have each digester create and return its own output +channel, but then we would need additional goroutines to fan-in the +results. + +The final stage receives all the `results` from `c` then checks the +error from `errc`. This check cannot happen any earlier, since before +this point, `walkFiles` may block sending values downstream: + +.code pipelines/bounded.go /m := make/,/^}/ HLerrc + +* Conclusion + +This article has presented techniques for constructing streaming data pipelines +in Go. Dealing with failures in such pipelines is tricky, since each stage in +the pipeline may block attempting to send values downstream, and the downstream +stages may no longer care about the incoming data. We showed how closing a +channel can broadcast a "done" signal to all the goroutines started by a +pipeline and defined guidelines for constructing pipelines correctly. + +Further reading: + +- [[http://talks.golang.org/2012/concurrency.slide#1][Go Concurrency Patterns]] ([[https://www.youtube.com/watch?v=f6kdp27TYZs][video]]) presents the basics of Go's concurrency primitives and several ways to apply them. +- [[http://blog.golang.org/advanced-go-concurrency-patterns][Advanced Go Concurrency Patterns]] ([[http://www.youtube.com/watch?v=QDDwwePbDtw][video]]) covers more complex uses of Go's primitives, especially `select`. +- Douglas McIlroy's paper [[http://swtch.com/~rsc/thread/squint.pdf][Squinting at Power Series]] shows how Go-like concurrency provides elegant support for complex calculations. |