So far in this book, you've seen functions and objects that process their inputs immediately using a single thread of execution where the code runs to completion and produces useful results or state changes. In this chapter, you turn your attention to concurrent, parallel, asynchronous, and reactive programs. These represent substantially different approaches to programming from those you've seen so far. Some of the reasons for turning to these techniques are as follows:
To achieve better responsiveness in a graphical user interface (GUI)
To report progress results during a long-running computation and to support cancellation of these computations
To achieve greater throughput in a reactive application or service
To achieve faster processing rates on a multiprocessor machine or cluster
To take advantage of the I/O parallelism available in modern disk drives or network connections
To sustain processing while network and disk I/O operations are in process
This chapter covers some of the techniques that can help achieve these outcomes:
Using .NET threads and the BackgroundWorker
class for background computations
Using events and messages to report results back to a GUI
Using F# asynchronous workflows and the .NET thread pool to handle network requests and other asynchronous I/O operations
Using F# pattern-matching to process message queues
Using low-level .NET shared-memory primitives to implement new concurrency techniques and control access to mutable data structures
Chapter 11 looked at the most common type of reactive program: GUI programs that respond to events raised on the GUI thread. The inner loop of such an application (contained in the Windows Forms library) spends most of its time blocked and waiting for the underlying operating system to notify it of a relevant event, such as a click from the user or a timer event from the operating system. This notification is received as an event in a message queue. Many GUI programs have only a single thread of execution, so all computation happens on the GUI thread. This can lead to problems such as nonresponsive user interfaces. This is one of many reasons it's important to master some of the techniques of concurrent and asynchronous programming.
Let's begin by looking more closely at some terminology:
Processes are, in the context of this chapter, standard operating system (OS) processes. Each instance of the .NET Common Language Runtime (CLR) runs in its own process, and multiple instances of the .NET CLR are often running on the same machine.
Threads are, in the context of this chapter, standard .NET threads. On most implementations of .NET, these correspond to operating system threads. Each .NET process has many threads running within the one process.
Concurrent programs are ones with multiple threads of execution, each typically executing different code, or at different execution points within the same code. Simultaneous execution may be simulated by scheduling and descheduling the threads, which is done by the OS. For example, most OS services and GUI applications are concurrent.
Parallel programs are one or more processes or threads executing simultaneously. For example, many modern microprocessors have two or more physical CPUs capable of executing processes and threads in parallel. Parallel programs can also be data parallel. For example, a massively parallel device such as a graphics processor unit (GPU) can process arrays and images in parallel. Parallel programs can also be a cluster of computers on a network, communicating via message passing. Historically, some parallel scientific programs have even used e-mail for communication!
Asynchronous programs perform requests that don't complete immediately but are fulfilled at a later time and where the program issuing the request has to do meaningful work in the meantime. For example, most network I/O is inherently asynchronous. A web crawler is also a highly asynchronous program, managing hundreds or thousands of simultaneous network requests.
Reactive programs are ones whose normal mode of operation is to be in a state of waiting for some kind of input, such as waiting for user input or for input from a message queue via a network socket. For example, GUI applications and web servers are reactive programs.
Parallel, asynchronous, concurrent, and reactive programs bring many interesting challenges. For example, these programs are nearly always nondeterministic. This makes debugging more challenging because it's difficult to step through a program; even pausing a running program with outstanding asynchronous requests may cause timeouts. Most dramatically, incorrect concurrent programs may deadlock, which means all threads are waiting on results from some other thread and no thread can make progress. Programs may also livelock, where processing is occurring and messages are being sent between threads but no useful work is being performed.
One of the easiest ways to get going with concurrency and parallelism is to use the System.ComponentModel.BackgroundWorker
class of the .NET Framework. A BackgroundWorker
class runs on its own dedicated operating system thread. These objects can be used in many situations but are especially useful for coarse-grained concurrency and parallelism such as checking the spelling of a document in the background. This section shows some simple uses of BackgroundWorker
and how to build similar objects that use BackgroundWorker
internally.
Listing 13-1 shows a simple use of BackgroundWorker
that computes the Fibonacci numbers on the worker thread.
Example 13.1. A Simple BackgroundWorker
open System.ComponentModel open System.Windows.Forms let worker = new BackgroundWorker() let numIterations = 1000 worker.DoWork.Add(fun args -> let rec computeFibonacci resPrevPrev resPrev i =// Compute the next result
let res = resPrevPrev + resPrev// At the end of the computation write the result into mutable state
if i = numIterations then args.Result <- box res else// Compute the next result
computeFibonacci resPrev res (i+1) computeFibonacci 1 1 2) worker.RunWorkerCompleted.Add(fun args -> MessageBox.Show(sprintf "Result = %A" args.Result) |> ignore)// Execute the worker
worker.RunWorkerAsync()
Table 13-1 shows the primary members of a BackgroundWorker
object. The execution sequence of the code in Listing 13-1 is as follows:
The main application thread creates and configures a BackgroundWorker
object.
After configuration is complete, the main application thread calls the RunWorkerAsync
method on the BackgroundWorker
object. This causes the DoWork
event to be raised on the worker thread.
The DoWork
event handler is executed in the worker thread and computes the one-thousandth Fibonacci number. At the end of the computation, the result is written into args.Result
, a mutable storage location in the event arguments for the DoWork
event. The DoWork
event handler then completes.
At some point after the DoWork
event handler completes, the RunWorkerCompleted
event is automatically raised on the main application thread. This displays a message box with the result of the computation, retrieved from the args
field of the event arguments.
Table 13.1. Primary Members of the BackgroundWorker
Class
Member and Type | Description |
---|---|
| Starts the process on a separate thread asynchronously. Called from the main thread. |
| Sets the |
| Set to |
| Set to |
| Set to |
| Indicates the progress of the operation. Used by the worker thread. |
| Fires in response to a call to |
| Fires when the background operation is canceled, when the operation is completed, or when an exception is thrown. Invoked on the main thread. |
| Fires whenever the |
Objects such as a BackgroundWorker
are two-faced: they have some methods and events that are for use from the main thread and some that are for use on the worker thread. This is common in concurrent programming. In particular, be careful to understand which thread an event is raised on. For BackgroundWorker
, the RunWorkerAsync
and CancelAsync
methods are for use from the GUI thread, and the ProgressChanged
and RunWorkerCompleted
events are raised on the GUI thread. The DoWork
event is raised on the worker thread, and the ReportProgress
method and the CancellationPending
property are for use from the worker thread when handling this event.
The members in Table 13-1 show two additional facets of BackgroundWorker
objects: they can optionally support protocols for cancellation and reporting progress. To report progress percentages, a worker must call the ReportProgress
method, which raises the ProgressChanged
event in the GUI thread. For cancellation, a worker computation need only check the CancellationPending
property at regular intervals, exiting the computation as a result.
Capturing common control patterns such as cancellation and progress reporting is an absolutely essential part of mastering concurrent programming. However, one of the problems with .NET classes such as BackgroundWorker
is that they're often more imperative than you may want, and they force other common patterns to be captured by using mutable data structures shared between threads. This leads to the more difficult topic of shared-memory concurrency, discussed later in the chapter. Furthermore, the way BackgroundWorker
handles cancellation means that you must insert flag-checks and early-exit paths in the executing background process. Finally, BackgroundWorker
isn't useful for background threads that perform asynchronous operations, because the background thread exits too early, before the callbacks for the asynchronous operations have executed.
For this reason, it can often be useful to build abstractions that are similar to BackgroundWorker
but that capture richer or different control patterns, preferably in a way that doesn't rely on the use of mutable state and that interferes less with the structure of the overall computation. Much of the rest of this chapter looks at various techniques to build these control structures.
You start with a case study in which you build a type IterativeBackgroundWorker
that represents a variation on the BackgroundWorker
design pattern. Listing 13-2 shows the code.
Example 13.2. A Variation on the BackgroundWorker
Design Pattern for Iterative Computations
open System.ComponentModel open System.Windows.Forms/// An IterativeBackgroundWorker follows the BackgroundWorker design pattern
/// but instead of running an arbitrary computation it iterates a function
/// a fixed number of times and reports intermediate and final results.
/// The worker is paramaterized by its internal state type.
///
/// Percentage progress is based on the iteration number. Cancellation checks
/// are made at each iteration. Implemented via an internal BackgroundWorker.
type IterativeBackgroundWorker<'T>(oneStep:('T -> 'T), initialState:'T, numIterations:int) = let worker = new BackgroundWorker(WorkerReportsProgress=true, WorkerSupportsCancellation=true)// Create the events that we will later trigger
let completed = new Event<_>() let error = new Event<_>() let cancelled = new Event<_>() let progress = new Event<_>() do worker.DoWork.Add(fun args ->
// This recursive function represents the computation loop.
// It runs at "maximum speed", i.e. is an active rather than
// a reactive process, and can only be controlled by a
// cancellation signal.
let rec iterate state i =// At the end of the computation terminate the recursive loop
if worker.CancellationPending then args.Cancel <- true elif i < numIterations then// Compute the next result
let state' = oneStep state// Report the percentage computation and the internal state
let percent = int ((float (i+1)/float numIterations) * 100.0) do worker.ReportProgress(percent, box state);// Compute the next result
iterate state' (i+1) else args.Result <- box state iterate initialState 0) do worker.RunWorkerCompleted.Add(fun args -> if args.Cancelled then cancelled.Trigger() elif args.Error <> null then error.Trigger args.Error else completed.Trigger (args.Result :?> 'T)) do worker.ProgressChanged.Add(fun args -> progress.Trigger (args.ProgressPercentage,(args.UserState :?> 'T))) member x.WorkerCompleted = completed.Publish member x.WorkerCancelled = cancelled.Publish member x.WorkerError = error.Publish member x.ProgressChanged = progress.Publish// Delegate the remaining members to the underlying worker
member x.RunWorkerAsync() = worker.RunWorkerAsync() member x.CancelAsync() = worker.CancelAsync()
The types inferred for the code in Listing 13-2 are as follows:
type IterativeBackgroundWorker<'State> =
new : ('State -> 'State) * 'State * int -> IterativeBackgroundWorker<'State>
member RunWorkerAsync : unit -> unit
member CancelAsync : unit -> unit
member ProgressChanged : IEvent<int * 'State>
member WorkerCompleted : IEvent<'State>
member WorkerCancelled : IEvent<unit>
member WorkerError : IEvent<exn>
Let's look at this signature first, because it represents the design of the type. The worker constructor is given a function of type 'State -> 'State
to compute successive iterations of the computation, plus an initial state and the number of iterations to compute. For example, you can compute the Fibonacci numbers using the following iteration function:
let fibOneStep (fibPrevPrev:bigint,fibPrev) = (fibPrev, fibPrevPrev+fibPrev);;
The type of this function is as follows:
val fibOneStep : bigint * bigint -> bigint * bigint
The RunWorkerAsync
and CancelAsync
members follow the BackgroundWorker
design pattern, as do the events, except that you expand the RunWorkerCompleted
event into three events to correspond to the three termination conditions and modify the ProgressChanged
to include the state. You can instantiate the type as follows:
> let worker = new IterativeBackgroundWorker<_>( fibOneStep,(1I,1I),100);;val worker : IterativeBackgroundWorker<bigint * bigint>
> worker.WorkerCompleted.Add(fun result -> MessageBox.Show(sprintf "Result = %A" result) |> ignore);;val it : unit = ()
> worker.ProgressChanged.Add(fun (percentage, state) -> printfn "%d%% complete, state = %A" percentage state);;val it : unit = ()
> worker.RunWorkerAsync();;1% complete, state = (1I, 1I)
2% complete, state = (1I, 2I)
3% complete, state = (2I, 3I)
4% complete, state = (3I, 5I)
...
98% complete, state = (135301852344706746049I, 218922995834555169026I)
99% complete, state = (218922995834555169026I, 354224848179261915075I)
100% complete, state = (354224848179261915075I, 573147844013817084101I)
val it : unit = ()
One difference here is that cancellation and percentage progress reporting are handled automatically based on the iterations of the computation. This is assuming each iteration takes roughly the same amount of time. Other variations on the BackgroundWorker
design pattern are possible. For example, reporting percentage completion of fixed tasks such as installation is often performed by timing sample executions of the tasks and adjusting the percentage reports appropriately.
You implement IterativeBackgroundWorker
via delegation rather than inheritance. This is because its external members are different from those of BackgroundWorker
. The .NET documentation recommends that you use implementation inheritance for this, but we disagree. Implementation inheritance can only add complexity to the signature of an abstraction and never makes things simpler, whereas an IterativeBackgroundWorker
is in many ways simpler than using a BackgroundWorker
, despite that it uses an instance of the latter internally. Powerful, compositional, simple abstractions are the primary building blocks of functional programming.
Often, you need to raise additional events from objects that follow the BackgroundWorker
design pattern. For example, let's say you want to augment IterativeBackgroundWorker
to raise an event when the worker starts its work and for this event to pass the exact time that the worker thread started as an event argument. Listing 13-3 shows the extra code you need to add to the IterativeBackgroundWorker
type to make this happen. You use this extra code in the next section.
Example 13.3. Code to Raise GUI-Thread Events from an IterativeBackgroundWorker
Object
open System open System.Threading// Pseudo-code for adding event-raising to this object
type IterativeBackgroundWorker<'T>(...) = let worker = ...// The constructor captures the synchronization context. This allows us to post
// messages back to the GUI thread where the BackgroundWorker was created.
let syncContext = SynchronizationContext.Current do if syncContext = null then failwith "no synchronization context found" let started = new Event<_>()// Raise the event when the worker starts. This is done by posting a message
// to the captured synchronization context.
do worker.DoWork.Add(fun args -> syncContext.Post(SendOrPostCallback(fun _ -> started.Trigger(DateTime.Now)), state=null) .../// The Started event gets raised when the worker starts. It is
/// raised on the GUI thread (i.e. in the synchronization context of
/// the thread where the worker object was created).
// It has type IEvent<DateTime>
member x.Started = started.Publish
The simple way to raise additional events is often wrong. For example, it's tempting to create an event, arrange for it to be triggered, and publish it, as you would do for a GUI control. However, if you do that, you end up triggering the event on the background worker thread, and its event handlers run on that thread. This is dangerous, because most GUI objects (and many other objects) can be accessed only from the thread they were created on; this is a restriction enforced by most GUI systems.
One of the nice features of the BackgroundWorker
class is that it automatically arranges to raise the RunWorkerCompleted
and ProgressChanged
events on the GUI thread. Listing 13-3 shows how to achieve this. Technically speaking, the extended IterativeBackgroundWorker
object captures the synchronization context of the thread where it was created and posts an operation back to that synchronization context. A synchronization context is an object that lets you post operations back to another thread. For threads such as a GUI thread, this means that posting an operation posts a message through the GUI event loop.
To round off this section on the BackgroundWorker
design pattern, Listing 13-4 shows the full code required to build a small application with a background worker task that supports cancellation and reports progress.
Example 13.4. Connecting an IterativeBackgroundWorker
to a GUI
open System.Drawing
open System.Windows.Forms
let form = new Form(Visible=true,TopMost=true)
let panel = new FlowLayoutPanel(Visible=true,
Height = 20,
Dock=DockStyle.Bottom,
BorderStyle=BorderStyle.FixedSingle)
let progress = new ProgressBar(Visible=false,
Anchor=(AnchorStyles.Bottom ||| AnchorStyles.Top),
Value=0)
let text = new Label(Text="Paused",
Anchor=AnchorStyles.Left,
Height=20,
TextAlign= ContentAlignment.MiddleLeft)
panel.Controls.Add(progress)
panel.Controls.Add(text)
form.Controls.Add(panel)
let fibOneStep (fibPrevPrev:bigint,fibPrev) = (fibPrev, fibPrevPrev+fibPrev)
// Run the iterative algorithm 500 times before reporting intermediate results
let rec repeatMultipleTimes n f s =
if n <= 0 then s else repeatMultipleTimes (n-1) f (f s)// Burn some additional cycles to make sure it runs slowly enough
let rec burnSomeCycles n f s = if n <= 0 then f s else ignore (f s); burnSomeCycles (n-1) f s let step = (repeatMultipleTimes 500 (burnSomeCycles 1000 fibOneStep))// Create the iterative worker.
let worker = new IterativeBackgroundWorker<_>(step,(1I,1I),80) worker.ProgressChanged.Add(fun (progressPercentage,state)-> progress.Value <- progressPercentage) worker.WorkerCompleted.Add(fun (_,result) -> progress.Visible <- false; text.Text <- "Paused"; MessageBox.Show(sprintf "Result = %A" result) |> ignore) worker.WorkerCancelled.Add(fun () -> progress.Visible <- false; text.Text <- "Paused"; MessageBox.Show(sprintf "Cancelled OK!") |> ignore) worker.WorkerError.Add(fun exn -> text.Text <- "Paused"; MessageBox.Show(sprintf "Error: %A" exn) |> ignore) form.Menu <- new MainMenu() let workerMenu = form.Menu.MenuItems.Add("&Worker") workerMenu.MenuItems.Add(new MenuItem("Run",onClick=(fun _ args -> text.Text <- "Running"; progress.Visible <- true; worker.RunWorkerAsync()))) workerMenu.MenuItems.Add(new MenuItem("Cancel",onClick=(fun _ args -> text.Text <- "Cancelling"; worker.CancelAsync()))) form.Closed.Add(fun _ -> worker.CancelAsync())
When you run the code in F# Interactive, a window appears, as shown in Figure 13-1.
The two background worker samples shown so far run at full throttle. In other words, the computations run on the background threads as active loops, and their reactive behavior is limited to flags that check for cancellation. In reality, background threads often have to do different kinds of work, either by responding to completed asynchronous I/O requests, by processing messages, by sleeping, or by waiting to acquire shared resources. Fortunately, F# comes with a powerful set of techniques for structuring asynchronous programs in a natural way. These are called asynchronous workflows. The next three sections cover how to use asynchronous workflows to structure asynchronous and message-processing tasks in ways that preserve the essential logical structure of your code.
One of the most intuitive asynchronous tasks is fetching a web page; we all use web browsers that can fetch multiple pages simultaneously. The samples in Chapter 2 showed how to fetch pages synchronously. This is useful for many purposes, but browsers and high-performance web crawlers have tens or thousands of connections in flight at once.
The type Microsoft.FSharp.Control.Async<'T>
lies at the heart of F# asynchronous workflows. A value of type Async<'T>
represents a program fragment that will generate a value of type 'T
at some point in the future. Listing 13-5 shows how to use asynchronous workflows to fetch several web pages simultaneously. (This example uses a method AsyncReadToEnd
defined in the open source library called the F# Power Pack. If you like, you can replace this with a call to the synchronous method ReadToEnd
, defined in the standard .NET libraries. Doing so may increase the thread count used by executing this code.)
Example 13.5. Fetching Three Web Pages Simultaneously
#r "FSharp.PowerPack.dll" // contains the definition for AsyncReadToEnd
open System.Net
open System.IO
let museums = ["MOMA", "http://moma.org/";
"British Museum", "http://www.thebritishmuseum.ac.uk/";
"Prado", "http://museoprado.mcu.es"]
let fetchAsync(nm,url:string) =
async { printfn "Creating request for %s..." nm
let req = WebRequest.Create(url)
let! resp = req.AsyncGetResponse()
printfn "Getting response stream for %s..." nm
let stream = resp.GetResponseStream()
printfn "Reading response for %s..." nm
let reader = new StreamReader(stream)
let! html = reader.AsyncReadToEnd()
printfn "Read %d characters for %s..." html.Length nm }
for nm,url in museums do
Async.Start (fetchAsync(nm,url))
The types of these functions and values are as follows:
val museums : (string * string) list
val fetchAsync : string * string -> Async<unit>
When run on one of our machines via F# Interactive, the output of the code from Listing 13-5 is as follows:
Creating request for MOMA...
Creating request for British Museum...
Creating request for Prado...
Getting response for MOMA...
Reading response for MOMA...
Getting response for Prado...
Reading response for Prado...
Read 188 characters for Prado...
Read 41635 characters for MOMA...
Getting response for British Museum...
Reading response for British Museum...
Read 24341 characters for British Museum...
The heart of the code in Listing 13-5 is the construct introduced by async { ... }
. This is an application of the workflow syntax introduced in Chapter 9.
Let's take a closer look at Listing 13-5. The key operations are the two let!
operations within the workflow expression:
async { ... let! resp = req.AsyncGetResponse() ... let! html = reader.AsyncReadToEnd() ... }
Within asynchronous workflow expressions, the language construct let! var = expr in body
means "perform the asynchronous operation expr
and bind the result to var
when the operation completes. Then, continue by executing the rest of the computation body
."
With this in mind, you can now see what fetchAsync
does:
It synchronously requests a web page.
It asynchronously awaits a response to the request.
It gets the response Stream
and StreamReader
synchronously after the asynchronous request completes.
It reads to the end of the stream asynchronously.
After the read completes, it prints the total number of characters read synchronously.
Finally, you use the method Async.Start
to initiate the execution of a number of asynchronous computations. This works by queuing the computations in the .NET thread pool. The following section explains the .NET thread pool in more detail.
Asynchronous computations are different from normal, synchronous computations: an asynchronous computation tends to hop between different underlying .NET threads. To see this, let's augment the asynchronous computation with diagnostics that show the ID of the underlying .NET thread at each point of active execution. You can do this by replacing uses of printfn
in the function fetchAsync
with uses of the following function:
let tprintfn fmt = printf "[.NET Thread %d]" System.Threading.Thread.CurrentThread.ManagedThreadId; printfn fmt
After doing this, the output changes to the following:
[.NET Thread 12]Creating request for MOMA...
[.NET Thread 13]Creating request for British Museum...
[.NET Thread 12]Creating request for Prado...
[.NET Thread 8]Getting response for MOMA...
[.NET Thread 8]Reading response for MOMA...
[.NET Thread 9]Getting response for Prado...
[.NET Thread 9]Reading response for Prado...
[.NET Thread 9]Read 188 characters for Prado...
[.NET Thread 8]Read 41635 characters for MOMA...
[.NET Thread 8]Getting response for British Museum...
[.NET Thread 8]Reading response for British Museum...
[.NET Thread 8]Read 24341 characters for British Museum...
Note how each individual Async
program hops between threads; the MOMA request started on .NET thread 12 and finished life on .NET thread 8. Each asynchronous computation in Listing 13-5 executes in the following way:
It starts life as a work item in the .NET thread pool. (The .NET thread pool is explained in the "What Is the .NET Thread Pool?" sidebar.) These are processed by a number of .NET threads.
When the asynchronous computations reach the AsyncGetResponse
and AsyncReadToEnd
calls, the requests are made, and the continuations are registered as I/O completion actions in the .NET thread pool. No thread is used while the request is in progress.
When the requests complete, they trigger a callback in the .NET thread pool. These may be serviced by different threads than those that originated the calls.
Async<'T
> values are essentially a way of writing continuation-passing or callback programs explicitly. Continuations themselves were described in Chapter 8 along with techniques to pass them explicitly. Async<'T>
computations call a success continuation when the asynchronous computation completes and anexception continuation if it fails. They provide a form of managed asynchronous computation, where managed means that several aspects of asynchronous programming are handled automatically:
Exception propagation is added for free: If an exception is raised during an asynchronous step, then the exception terminates the entire asynchronous computation and cleans up any resources declared using use
, and the exception value is then handed to a continuation. Exceptions may also be caught and managed within the asynchronous workflow by using try/with/finally
.
Cancellation checking is added for free: The execution of an Async<'T>
workflow automatically checks a cancellation flag at each asynchronous operation. Cancellation can be controlled through the use of cancellation tokens.
Resource lifetime management is fairly simple: You can protect resources across parts of an asynchronous computation by using use
inside the workflow syntax.
If you put aside the question of cancellation, values of type Async<'T>
are effectively identical to the following type:
type Async<'T> = Async of ('T -> unit) * (exn -> unit) -> unit
Here, the functions are the success continuation and exception continuations, respectively. Each value of type Async<'T>
should eventually call one of these two continuations. The async
object is of type AsyncBuilder
and supports the following methods, among others:
type AsyncBuilder with
member Return : 'T -> Async<'T>
member Delay : (unit -> Async<'T>) -> Async<'T>
member Using: 'T * ('T -> Async<'U>) -> Async<'U> when 'T :> System.IDisposable
member Bind: Async<'T> * ('T -> Async<'U>) -> Async<'U>
The full definition of Async<'T>
values and the implementations of these methods for the async
object are given in the F# library source code. As you saw in Chapter 9, builder objects such as async
containing methods like those shown previously mean that you can use the syntax async { ... }
as a way of building Async<'T>
values.
Table 13-2 shows the common constructs used in asynchronous workflow expressions. For example, the following asynchronous workflow
async { let req = WebRequest.Create("http://moma.org/") let! resp = req.AsyncGetResponse() let stream = resp.GetResponseStream() let reader = new StreamReader(stream) let! html = reader.AsyncReadToEnd() html }
is shorthand for the following code:
async.Delay(fun () -> let req = WebRequest.Create("http://moma.org/") async.Bind(req.AsyncGetResponse(), (fun resp -> let stream = resp.GetResponseStream() let reader = new StreamReader(stream) async.Bind(reader.AsyncReadToEnd(), (fun html -> async.Return html)))
As you saw in Chapter 9, the key to understanding the F# workflow syntax is always to understand the meaning of let!
. In the case of async
workflows, let!
executes one asynchronous computation and schedules the next computation for execution after the first asynchronous computation completes. This is syntactic sugar for the Bind
operation on the async
object.
Table 13.2. Common Constructs Used in async { ... }
Workflow Expressions
Construct | Description |
---|---|
| Executes the asynchronous computation |
| Executes an expression synchronously, and binds its result to |
| Equivalent to |
| Equivalent to |
| Evaluates the expression, and returns its value as the result of the containing asynchronous workflow. Equivalent to |
| Executes the expression as an asynchronous computation, and returns its result as the overall result of the containing asynchronous workflow. Equivalent to |
| Executes the expression immediately, and binds its result immediately. Calls the |
This section shows a slightly longer example of asynchronous I/O processing. The running sample is an application that must read a large number of image files and perform some processing on them. This kind of application may be compute bound (if the processing takes a long time and the file system is fast) or I/O bound (if the processing is quick and the file system is slow). Using asynchronous techniques tends to give good overall performance gains when an application is I/O bound and can also give performance improvements for compute-bound applications if asynchronous operations are executed in parallel on multicore machines.
Listing 13-6 shows a synchronous implementation of the image-transformation program.
Example 13.6. A Synchronous Image Processor
open System.IO let numImages = 200 let size = 512 let numPixels = size * size let makeImageFiles () = printfn "making %d %dx%d images... " numImages size size
let pixels = Array.init numPixels (fun i -> byte i)
for i = 1 to numImages do
System.IO.File.WriteAllBytes(sprintf "Image%d.tmp" i, pixels)
printfn "done."
let processImageRepeats = 20
let transformImage (pixels, imageNum) =
printfn "transformImage %d" imageNum;
// Perform a CPU-intensive operation on the image.
for i in 1 .. processImageRepeats do
pixels |> Array.map (fun b -> b + 1uy) |> ignore
pixels |> Array.map (fun b -> b + 1uy)
let processImageSync i =
use inStream = File.OpenRead(sprintf "Image%d.tmp" i)
let pixels = Array.zeroCreate numPixels
let nPixels = inStream.Read(pixels,0,numPixels);
let pixels' = transformImage(pixels,i)
use outStream = File.OpenWrite(sprintf "Image%d.done" i)
outStream.Write(pixels',0,numPixels)
let processImagesSync () =
printfn "processImagesSync...";
for i in 1 .. numImages do
processImageSync(i)
You assume the image files are already created using the following code:
> System.Environment.CurrentDirectory <- __SOURCE_DIRECTORY__;;
val it : unit = ()
> makeImageFiles();;
val it : unit = ()
You leave the transformation on the image largely unspecified, such as the function transformImage
. By changing the value of processImageRepeats
, you can adjust the computation from compute bound to I/O bound.
The problem with this implementation is that each image is read and processed sequentially, when in practice multiple images can be read and transformed simultaneously, giving much greater throughput. Listing 13-7 shows the implementation of the image processor using an asynchronous workflow.
Example 13.7. The Asynchronous Image Processor
let processImageAsync i = async { use inStream = File.OpenRead(sprintf "Image%d.tmp" i) let! pixels = inStream.AsyncRead(numPixels) let pixels' = transformImage(pixels,i) use outStream = File.OpenWrite(sprintf "Image%d.done" i) do! outStream.AsyncWrite(pixels') }
let processImagesAsync() = printfn "processImagesAsync..."; let tasks = [ for i in 1 .. numImages -> processImageAsync(i) ] Async.RunSynchronously (Async.Parallel tasks) |> ignore printfn "processImagesAsync finished!";
On one of our machines, the asynchronous version of the code ran up to three times as fast as the synchronous version (in total elapsed time), when processImageRepeats
is 20 and numImages
is 200. A factor of 2 was achieved consistently for any number of processImageRepeats
because this machine had two CPUs.
Let's take a closer look at this code. The call Async.Run (Async.Parallel ...)
executes a set of asynchronous operations in the thread pool, collects their results (or their exceptions), and returns the overall array of results to the original code. The core asynchronous workflow is introduced by the async { ... }
construct. Let's look at the inner workflow line by line:
async { use inStream = File.OpenRead(sprintf "Image%d.tmp" i) ... }
This line opens the input stream synchronously using File.OpenRead
. Although this is a synchronous operation, the use of use
indicates that the lifetime of the stream is managed over the remainder of the workflow. The stream is closed when the variable is no longer in scope: that is, at the end of the workflow, even if asynchronous activations occur in between. If any step in the workflow raises an uncaught exception, then the stream is also closed while handling the exception.
The next line reads the input stream asynchronously using inStream.AsyncRead
:
async { use inStream = File.OpenRead(sprintf "Image%d.tmp" i) let! pixels = inStream.AsyncRead(numPixels) ... }
Stream.AsyncRead
is an extension method added to the .NET System.IO.Stream
class defined in the F# library and generates a value of type Async<byte[]>
. The use of let!
executes this operation asynchronously and registers a callback. When the callback is invoked, the value pixels
is bound to the result of the operation, and the remainder of the asynchronous workflow is executed. The next line transforms the image synchronously using transformImage
:
async { use inStream = File.OpenRead(sprintf "Image%d.tmp" i) let! pixels = inStream.AsyncRead(numPixels) let pixels' = transformImage(pixels,i) ... }
Like the first line, the next line opens the output stream. Using use
guarantees that the stream is closed by the end of the workflow regardless of whether exceptions are thrown in the remainder of the workflow:
async { use inStream = File.OpenRead(sprintf "Image%d.tmp" i) let! pixels = inStream.AsyncRead(numPixels) let pixels' = transformImage(pixels,i) use outStream = File.OpenWrite(sprintf "Image%d.done" i) ... }
The final line of the workflow performs an asynchronous write of the image. Once again, AsyncWrite
is an extension method added to the .NET System.IO.Stream
class defined in the F# library:
async { use inStream = File.OpenRead(sprintf "Image%d.tmp" i) let! pixels = inStream.AsyncRead(numPixels) let pixels' = transformImage(pixels,i) use outStream = File.OpenWrite(sprintf "Image%d.done" i) do! outStream.AsyncWrite(pixels') }
If you now return to the first part of the function, you can see that the overall operation of the function is to create numImages
individual asynchronous operations, using a sequence expression that generates a list:
let tasks = [ for i in 1 .. numImages -> processImageAsync(i) ]
You can compose these tasks in parallel using Async.Parallel
and then run the resulting process using Async.Run
. This waits for the overall operation to complete and returns the result:
Async.Run (Async.Parallel tasks)
Table 13-3 shows some of the primitives and combinators commonly used to build asynchronous workflows. Take the time to compare Listings 13-7 and 13-6. Notice the following:
The overall structure and flow of the core of Listing 13-7 is similar to Listing 13-6: that is, the synchronous algorithm, even though it includes steps executed asynchronously.
The performance characteristics of Listing 13-7 are the same as those of Listing 13-6. Any overhead involved in executing the asynchronous workflow is easily dominated by the overall cost of I/O and image processing. It's also much easier to experiment with modifications such as making the write operation synchronous.
Table 13.3. Some Common Primitives Used to Build Async<'T>
Values
Member/Type | Description |
---|---|
| Builds a single primitive asynchronous step of an asynchronous computation. The function that implements the step is passed continuations to call after the step is complete or if the step fails. |
| Builds a single asynchronous computation that runs the given asynchronous computations in parallel and waits for results from all to be returned. Each may either terminate with a value or return an exception. If any raise an exception, then the others are cancelled, and the overall asynchronous computation also raises the same exception. |
Values of type Async<'T>
are usually run using the functions listed in Table 13-4. You can build Async<'T>
values using functions and members in the F# libraries.
Table 13.4. Common Methods in the Async
Type Used to Run Async<'T>
Values
Member/Type | Description |
---|---|
| Runs an operation in the thread pool and waits for its result. |
| Queues the asynchronous computation as an operation in the thread pool. |
| Queues the asynchronous computation, initially as a work item in the thread pool, but inherits the cancellation handle from the current asynchronous computation. |
Asynchronous programming is becoming more widespread because of the use of multicore machines and networks in applications, and many .NET APIs now come with both synchronous and asynchronous versions of their functionality. For example, all web service APIs generated by .NET tools have asynchronous versions of their requests. A quick scan of the .NET API documentation on the Microsoft website reveals the asynchronous operations listed in Table 13-5. These all have equivalent Async<'T>
operations defined in the F# libraries as extensions to the corresponding .NET types.
Table 13.5. Some Asynchronous Operations in the .NET Libraries and Corresponding Expected F# Naming Scheme
.NET Asynchronous Operation | F# Naming Scheme | Description |
---|---|---|
|
| Reads a stream of bytes asynchronously. See also |
|
| Writes a stream of bytes asynchronously. See also |
|
| Accepts an incoming network socket requestasyn chronouly. |
|
| Receives data on a network socket asynchronously. |
| Sends data on a network socket asynchronously. | |
|
| Makes an asynchronous web request. See also |
|
| Executes an |
|
| Executes a read of XML asynchronously. |
|
| Executes a nonreading |
Sometimes, you may need to write a few primitives to map .NET asynchronous operations into the F# asynchronous framework. You see some examples later in this section and in Chapter 14.
Async.Parallel
can appear magical. Computation tasks are created, executed, and resynchronized almost without effort. Listing 13-9 shows that a basic implementation of this operator is simple and again helps you see how Async<'T>
values work under the hood.
Example 13.9. A Basic Implementation of a Fork-Join Parallel Operator
let forkJoinParallel(taskSeq) = Async.FromContinuations (fun (cont,econt,ccont) -> let tasks = Seq.toArray taskSeq let count = ref tasks.Length let results = Array.zeroCreate tasks.Length tasks |> Array.iteri (fun i p -> Async.Start (async { let! res = p results.[i] <- res; let n = System.Threading.Interlocked.Decrement(count) if n=0 then cont results })))
This basic implementation first converts the input task sequence to an array and then creates mutable state count
and results
to record the progress of the parallel computations. It then iterates through the tasks and queues each for execution in the .NET thread pool. Upon completion, each writes its result and decrements the counter using an atomic Interlocked.Decrement
operator, discussed further in the section "Using Shared-Memory Concurrency" at the end of this chapter. The last process to finish calls the continuation with the collected results.
In practice, Async.Parallel
is implemented more efficiently and takes into account exceptions and cancellation; again, see the F# library code for full details.
One of the great advantages of F# async programming is that it can be used for both CPU and I/O parallel programming tasks. For example, you can use it for many CPU parallelism tasks that don't perform any I/O but rather carry out straight CPU-bound computations.
For optimized, partitioned CPU parallelism, this is often done by using Async.Parallel
with a number of tasks that exactly matches the number of physical processors on a machine. For example, the following code shows parallel initialization of an array where each cell is filled by running the input function. The implementation of this function makes careful use of shared memory primitives (a topic discussed later in this book) and is highly efficient:
open System.Threading open System// Initialize an array by a parallel init using all available processors
// Note, this primitive doesn't support cancellation.
let parallelArrayInit n f = let currentLine = ref −1 let res = Array.zeroCreate n let rec loop () = let y = Interlocked.Increment(¤tLine.contents) if y < n then res.[y] <- f y; loop()// Start just the right number of tasks, one for each physical CPU
Async.Parallel [ for i in 1 .. Environment.ProcessorCount -> async { do loop()} ] |> Async.Ignore |> Async.RunSynchronously res > let rec fib x = if x < 2 then 1 else fib (x - 1) + fib (x - 2) > parallelArrayInit 25 (fun x -> fib x);;val it : int [] =
[|1; 1; 2; 3; 5; 8; 13; 21; 34; 55; 89; 144; 233; 377; 610; 987; 1597; 2584;
4181; 6765; 10946; 17711; 28657; 46368; 75025; 121393; 196418; 317811;
514229; 832040|]
Two recurring topics in asynchronous programming are exceptions and cancellation. Let's first explore some of the behavior of asynchronous programs with regard to exceptions:
> let failingTask = async { do failwith "fail" };;val failingTask: Async<unit>
> Async.RunSynchronously failingTask;;Microsoft.FSharp.Core.FailureException: fail
stopped due to error
> let failingTasks = [ async { do failwith "fail A" }; async { do failwith "fail B" }; ];;
val failingTasks: Async<unit>
> Async.RunSynchronously (Async.Parallel failingTasks);;Microsoft.FSharp.Core.FailureException: fail A
stopped due to error
> Async.RunSynchronously (Async.Parallel failingTasks);;Microsoft.FSharp.Core.FailureException: fail B
stopped due to error
From these examples, you can see the following:
Tasks fail only when they're actually executed. The construction of a task using the async { ... }
syntax never fails.
Tasks run using Async.RunSynchronously
report any failure back to the controlling thread as an exception.
It's nondeterministic which task will fail first.
Tasks composed using Async.Parallel
report the first failure from among the collected set of tasks. An attempt is made to cancel other tasks by setting the cancellation flag for the group of tasks, and any further failures are ignored.
You can wrap a task using the Async.Catch
combinator. This has the following type:
static member Catch : Async<'T> -> Async<Choice<'T,exn>>
For example:
> Async.RunSynchronously (Async.Catch failingTask);; val it : Choice<unit,exn> = Choice2_2 (FailureException ())
You can also handle errors by using try
/finally
in an async { ... }
workflow.
A distinction is often made between shared-memory concurrency and message passing concurrency. The former is often more efficient on local machines and is covered in the section "Using Shared-Memory Concurrency" later in this chapter. The latter scales to systems where there is no shared memory—for example, distributed systems—and can also be used to avoid performance problems associated with shared memory. Asynchronous message passing and processing is a common foundation for concurrent programming, and this section looks at some simple examples of message-passing programs.
In a sense, you've already seen a good deal of message passing in this chapter. For example:
In the BackgroundWorker
design pattern, the CancelAsync
method is a simple kind of message.
Whenever you raise events on a GUI thread from a background thread, you are, under the hood, posting a message to the GUI's event queue. On Windows, this event queue is managed by the OS, and the processing of the events on the GUI thread is called the Windows event loop.
This section covers a simple kind of message processing called mailbox processing that's popular in languages such as Erlang. A mailbox is a message queue that you can scan for a message particularly relevant to the message-processing agent you're defining. Listing 13-10 shows a concurrent agent that implements a simple counter by processing a mailbox as messages arrive. The type MailboxProcessor
is defined in the F# library.
Example 13.10. Implementing a Counter Using a MailboxProcessor
let counter = new MailboxProcessor<_>(fun inbox -> let rec loop n = async { printfn "n = %d, waiting..." n let! msg = inbox.Receive() return! loop (n+msg) } loop 0)
The type of counter is MailboxProcessor<int>
, where the type argument indicates that this object expects to be sent messages of type int
:
val counter : MailboxProcessor<int>
The "Message Processing and State Machines" sidebar describes the general pattern of Listing 13-10 and the other MailboxProcessor
examples in this chapter, all of which can be thought of as state machines. With this in mind, let's take a closer look at Listing 13-10. First, let's use counter
on some simple inputs:
> counter.Start();;n = 0, waiting...
> counter.Post(1);;n = 1, waiting...
> counter.Post(2);;n = 3, waiting...
> counter.Post(1);; n = 4, waiting...
Looking at Listing 13-10, note that calling the Start
method causes the processing agent to enter loop
with n
= 0
. The agent then performs an asynchronous Receive
request on the inbox
for the MailboxProcessor
; that is, the agent waits asynchronously until a message has been received. When the message msg
is received, the program calls loop (n+msg)
. As additional messages are received, the internal counter (actually an argument) is incremented further.
You post messages to the agent using counter.Post
. The type of inbox.Receive
is as follows:
member Receive: unit -> Async<'Message>
Using an asynchronous receive ensures that no real threads are blocked for the duration of the wait. This means the previous techniques scale to many thousands of concurrent agents.
Often, it's wise to hide the internals of an asynchronous computation behind an object, because the use of message passing can be seen as an implementation detail. Listing 13-10 doesn't show you how to retrieve information from the counter, except by printing it to the standard output. Furthermore, it doesn't show how to ask the processing agent to exit. Listing 13-11 shows how to implement an object wrapping an agent that supports Increment
, Stop
, and Fetch
messages.
Example 13.11. Hiding a Mailbox and Supporting a Fetch Method
/// The internal type of messages for the agent
type internal msg = Increment of int | Fetch of AsyncReplyChannel<int> | Stop type CountingAgent() = let counter = MailboxProcessor.Start(fun inbox ->// The states of the message-processing state machine...
let rec loop n = async { let! msg = inbox.Receive() match msg with | Increment m ->// increment and continue...
return! loop(n+m) | Stop ->// exit
return () | Fetch replyChannel ->// post response to reply channel and continue
do replyChannel.Reply n return! loop n }// The initial state of the message-processing state machine...
loop(0)) member a.Increment(n) = counter.Post(Increment n) member a.Stop() = counter.Post Stop member a.Fetch() = counter.PostAndReply(fun replyChannel -> Fetch replyChannel)
The inferred public types indicate how the presence of a concurrent agent is successfully hidden by the use of an object:
type CountingAgent =
new : unit -> CountingAgent
member Fetch : unit -> int
member Increment : n:int -> unit
member Stop : unit -> unit
Here, you can see an instance of this object in action:
> let counter = new CountingAgent();;val counter : CountingAgent
> counter.Increment(1);;val it : unit = ()
> counter.Fetch();;val it : int = 1
> counter.Increment(2);;val it : unit = ()
> counter.Fetch();;
val it : int = 3
> counter.Stop();;
val it : unit = ()
Listing 13-11 shows several important aspects of message passing and processing using the mailbox-processing model:
Internal message protocols are often represented using discriminated unions. Here the type msg
has cases Increment
, Fetch
, and Stop
, corresponding to the three methods accepted by the object that wraps the overall agent implementation.
Pattern matching over discriminated unions gives a succinct way to process messages. A common pattern is a call to inbox.Receive()
or inbox.TryReceive()
followed by a match on the message contents.
The PostAndReply
on the MailboxProcessor
type gives a way to post a message and wait for a reply. A temporary reply channel is created and should form part of the message. A reply channel is an object of type Microsoft.FSharp.Control.AsyncReplyChannel<'reply>
, which in turn supports a Post
method. The MailboxProcessor
can use this to post a reply to the waiting caller. In Listing 13-11, the channel is sent to the underlying message-processing agent counter
as part of the Fetch
message.
Table 13-6 summarizes the most important members available on the MailboxProcessor
type.
Table 13.6. Some Members of the MailboxProcessor<'Message>
Type
Member/Type | Description |
---|---|
| Posts a message to a mailbox queue. |
| Returns the next message in the mailbox queue. If no messages are present, performs an asynchronous wait until the message arrives. If a timeout occurs, then raises a |
| Scans the mailbox for a message where the function returns a |
| Like |
| Like |
It's common for a message-processing agent to end up in a state where it's not interested in all messages that may appear in a mailbox but only a subset of them. For example, you may be awaiting a reply from another agent and aren't interested in serving new requests. In this case, it's essential that you use MailboxProcessor.Scan
rather than MailboxProcessor.Receive
. Table 13-6 shows the signatures of both of these. The former lets you choose between available messages by processing them in order, whereas the latter forces you to process every message. Listing 13-12 shows an example of using MailboxProcessor.Scan
.
Example 13.12. Scanning a Mailbox for Relevant Messages
type Message = | Message1 | Message2 of int | Message3 of string let agent = MailboxProcessor.Start(fun inbox -> let rec loop() = inbox.Scan(function | Message1 -> Some (async { do printfn "message 1!" return! loop() }) | Message2 n -> Some (async { do printfn "message 2!" return! loop() }) | Message3 _ -> None) loop())
You can now post these agent messages, including messages of the ignored kind Message3
:
> agent.Post(Message1) ;;message 1!
val it : unit = ()
> agent.Post(Message2(100));;message 2!
val it : unit = ()
> agent.Post(Message3("abc"));;val it : unit = ()
> agent.Post(Message2(100));;message 2!
val it : unit = ()
> agent.CurrentQueueLength;; val it : int = 1
When you send Message3
to the message processor, the message is ignored. However, the last line shows that the unprocessed Message3
is still in the message queue, which you examine using the backdoor property UnsafeMessageQueueContents
.
At the start of this chapter, we mentioned that the rise of the Web and other forms of networks is a major reason for the increasing importance of concurrent and asynchronous programming. Listing 13-13 shows an implementation of a web crawler using asynchronous programming and mailbox-processing techniques.
Example 13.13. A Scalable, Controlled Asynchronous Web Crawler
open System.Collections.Generic open System.Net open System.IO open System.Threading open System.Text.RegularExpressions let limit = 50 let linkPat = "href=s*"[^"h]*(http://[^&"]*)"" let getLinks (txt:string) = [ for m in Regex.Matches(txt,linkPat) -> m.Groups.Item(1).Value ]// A type that helps limit the number of active web requests
type RequestGate(n:int) = let semaphore = new Semaphore(initialCount=n,maximumCount=n) member x.AsyncAcquire(?timeout) = async { let! ok = Async.AwaitWaitHandle(semaphore, ?millisecondsTimeout=timeout) if ok then return { new System.IDisposable with member x.Dispose() = semaphore.Release() |> ignore } else return! failwith "couldn't acquire a semaphore" }// Gate the number of active web requests
let webRequestGate = RequestGate(5)// Fetch the URL, and post the results to the urlCollector.
let collectLinks (url:string) = async {// An Async web request with a global gate
let! html = async {// Acquire an entry in the webRequestGate. Release
// it when 'holder' goes out of scope
use! holder = webRequestGate.AsyncAcquire()
let req = WebRequest.Create(url,Timeout=5)// Wait for the WebResponse
use! response = req.AsyncGetResponse()
// Get the response stream
use reader = new StreamReader(response.GetResponseStream())// Read the response stream (note: a synchronous read)
return reader.ReadToEnd() }// Compute the links, synchronously
let links = getLinks html// Report, synchronously
do printfn "finished reading %s, got %d links" url (List.length links)// We're done
return links }/// 'urlCollector' is a single agent that receives URLs as messages. It creates new
/// asynchronous tasks that post messages back to this object.
let urlCollector = MailboxProcessor.Start(fun self ->// This is the main state of the urlCollector
let rec waitForUrl (visited : Set<string>) = async {// Check the limit
if visited.Count < limit then// Wait for a URL...
let! url = self.Receive() if not (visited.Contains(url)) then// Start off a new task for the new url. Each collects
// links and posts them back to the urlCollector.
do! Async.StartChild (async { let! links = collectLinks url for link in links do self.Post link }) |> Async.Ignore// Recurse into the waiting state
return! waitForUrl(visited.Add(url)) }// This is the initial state.
waitForUrl(Set.empty))
You can initiate a web crawl from a particular URL as follows:
> urlCollector <-- "http://news.google.com";;finished reading http://news.google.com, got 191 links
finished reading http://news.google.com/?output=rss, got 0 links
finished reading http://www.ktvu.com/politics/13732578/detail.html, got 14 links
finished reading http://www.washingtonpost.com/wp-dyn/content/art..., got 218 links
finished reading http://www.newsobserver.com/politics/story/646..., got 56 links
finished reading http://www.foxnews.com/story/0,2933,290307,0...l, got 22 links
...
The key techniques shown in Listing 13-13 are as follows:
The type RequestGate
encapsulates the logic needed to ensure that you place a global limit on the number of active web requests occurring at any point in time. This is instantiated to the particular instance webRequestGate
with limit 5. This uses a System.Threading.Semaphore
object to coordinate access to this shared resource. Semaphores are discussed in more detail in the section "Using Shared-Memory Concurrency."
The RequestGate
type ensures that web requests sitting in the request queue don't block threads but rather wait asynchronously as callback items in the thread pool until a slot in the webRequestGate
becomes available.
The collectLinks
function is a regular asynchronous computation. It first enters the RequestGate
(that is, acquires one of the available entries in the Semaphore
). After a response has been received, it reads off the HTML from the resulting reader, scrapes the HTML for links using regular expressions, and returns the generated set of links.
The urlCollector
is the only message-processing program. It's written using a MailboxProcessor
. In its main state, it waits for a fresh URL and spawns a new asynchronous computation to call collectLinks
once one is received. For each collected link, a new message is sent back to the urlCollector
's mailbox. Finally, you recurse to the waiting state, having added the fresh URL to the overall set of URLs you've traversed so far.
The operator <--
is used as shorthand for posting a message to an agent. This is a recommended abbreviation in F# asynchronous programming.
The AsyncAcquire
method of the RequestGate
type uses a design pattern called a holder. The object returned by this method is an IDisposable
object that represents the acquisition of a resource. This holder object is bound using use
, and this ensures that the resource is released when the computation completes or when the computation ends with an exception.
Listing 13-13 shows that it's relatively easy to create sophisticated, scalable asynchronous programs using a mix of message passing and asynchronous I/O techniques. Modern web crawlers have thousands of outstanding open connections, indicating the importance of using asynchronous techniques in modern scalable web-based programming.
The final topics covered in this chapter are the various primitive mechanisms used for threads, shared-memory concurrency, and signaling. In many ways, these are the assembly language of concurrency.
This chapter has concentrated mostly on techniques that work well with immutable data structures. That isn't to say you should always use immutable data structures. It is, for example, perfectly valid to use mutable data structures as long as they're accessed from only one particular thread. Furthermore, private mutable data structures can often be safely passed through an asynchronous workflow, because at each point the mutable data structure is accessed by only one thread, even if different parts of the asynchronous workflow are executed by different threads. This doesn't apply to workflows that use operators such as Async.Parallel
and Async.StartChild
that start additional threads of computation.
This means we've largely avoided covering shared-memory primitives so far, because F# provides powerful declarative constructs such as asynchronous workflows and message passing that often subsume the need to resort to shared-memory concurrency. However, a working knowledge of thread primitives and shared-memory concurrency is still very useful, especially if you want to implement your own basic constructs or highly efficient concurrent algorithms on shared-memory hardware.
This chapter has avoided showing how to work with threads directly, instead relying on abstractions such as BackgroundWorker
and the .NET thread pool. If you want to create threads directly, here is a short sample:
open System.Threading let t = new Thread(ThreadStart(fun _ -> printfn "Thread %d: Hello" Thread.CurrentThread.ManagedThreadId)); t.Start(); printfn "Thread %d: Waiting!" Thread.CurrentThread.ManagedThreadId t.Join(); printfn "Done!"
When run, this gives the following:
val t : Thread
Thread 1: Waiting!
Thread 10: Hello
Done!
Always avoid using Thread.Suspend
, Thread. Resume
, and Thread.Abort
. These are guaranteed ways to put obscure concurrency bugs in your program. The MSDN website has a good description of why Thread.Abort
may not even succeed. One of the only compelling uses for Thread.Abort
is to implement Ctrl+C in an interactive development environment for a general-purpose language such as F# Interactive.
Many multithreaded applications use mutable data structures shared between multiple threads. Without synchronization, these data structures will almost certainly become corrupt: threads may read data that has been only partially updated (because not all mutations are atomic), or two threads may write to the same data simultaneously (a race condition). Mutable data structures are usually protected by locks, although lock-free mutable data structures are also possible.
Shared-memory concurrency is a difficult and complicated topic, and a considerable amount of good material on .NET shared-memory concurrency is available on the Web. All this material applies to F# when you're programming with mutable data structures such as reference cells, arrays, and hash tables and the data structures can be accessed from multiple threads simultaneously. F# mutable data structures map to .NET memory in fairly predictable ways; for example, mutable references become mutable fields in a .NET class, and mutable fields of word size can be assigned atomically.
On modern microprocessors, multiple threads can see views of memory that aren't consistent; that is, not all writes are propagated to all threads immediately. The guarantees given are called a memory model and are usually expressed in terms of the ordering dependencies between instructions that read/write memory locations. This is, of course, deeply troubling, because you have to think about a huge number of possible reorderings of your code, and it's one of the main reasons why shared mutable data structures are difficult to work with. You can find further details on the .NET memory model at www.expert-fsharp.net/topics/MemoryModel
.
Locks are the simplest way to enforce mutual exclusion between two threads attempting to read or write the same mutable memory location. Listing 13-14 shows an example of code with a race condition.
Example 13.14. Shared-Memory Code with a Race Condition
type MutablePair<'T,'U>(x:'T,y:'U) =
let mutable currentX = x
let mutable currentY = y
member p.Value = (currentX,currentY)
member p.Update(x,y) =
// Race condition: This pair of updates is not atomic
currentX <- x;
currentY <- y
let p = new MutablePair<_,_>(1,2)
do Async.Start (async { do (while true do p.Update(10,10)) })
do Async.Start (async { do (while true do p.Update(20,20)) })
Here is the definition of the F# lock
function:
open System.Threading let lock (lockobj : obj) f = Monitor.Enter lockobj; try f() finally Monitor.Exit lockobj
The pair of mutations in the Update
method isn't atomic; that is, one thread may have written to currentX
, another then writes to both currentX
and currentY
, and the final thread then writes to currentY
, leaving the pair holding the value (10,20)
or (20,10)
. Mutable data structures are inherently prone to this kind of problem if shared between multiple threads. Luckily, F# code tends to have fewer mutations than imperative languages, because functions normally take immutable values and return a calculated value. However, when you do use mutable data structures, they shouldn't be shared between threads, or you should design them carefully and document their properties with respect to multithreaded access.
Here is one way to use the F# lock
function to ensure that updates to the data structure are atomic. Locks are also required on uses of the property p.Value
:
do Async.Start (async { do (while true do lock p (fun () -> p.Update(10,10))) }) do Async.Start (async { do (while true do lock p (fun () -> p.Update(20,20))) })
If you use locks inside data structures, then do so only in a simple way that uses them to enforce the concurrency properties you've documented. Don't lock just for the sake of it, and don't hold locks longer than necessary. In particular, beware of making indirect calls to externally supplied function values, interfaces, or abstract members while a lock is held. The code providing the implementation may not be expecting to be called when a lock is held and may attempt to acquire further locks in an inconsistent fashion.
It's common for mutable data structures to be read more than they're written. Indeed, mutation is often used only to initialize a mutable data structure. In this case, you can use a .NET ReaderWriterLock
to protect access to a resource. For example, consider the following two functions:
open System.Threading let readLock (rwlock : ReaderWriterLock) f = rwlock.AcquireReaderLock(Timeout.Infinite) try f() finally rwlock.ReleaseReaderLock() let writeLock (rwlock : ReaderWriterLock) f = rwlock.AcquireWriterLock(Timeout.Infinite) try f(); Thread.MemoryBarrier() finally rwlock.ReleaseWriterLock()
Listing 13-15 shows how to use these functions to protect the MutablePair
class.
Example 13.15. Shared-Memory Code with a Race Condition
type MutablePair<'T,'U>(x:'T,y:'U) = let mutable currentX = x let mutable currentY = y let rwlock = new ReaderWriterLock() member p.Value = readLock rwlock (fun () -> (currentX,currentY)) member p.Update(x,y) = writeLock rwlock (fun () -> currentX <- x; currentY <- y)
Table 13-7 shows some of the other shared-memory concurrency primitives available in the .NET Framework.
Table 13.7. .NET Shared-Memory Concurrency Primitives
Type | Description |
---|---|
| A synchronization object for signaling the control of threads. |
| A two-state (on/off) |
| A two-state (on/off) |
| A lock-like object that can be shared between operating system processes. |
| Used to limit the number of threads simultaneously accessing a resource. However, use a mutex or lock if at most one thread can access a resource at a time. |
| Atomic operations on memory locations. Especially useful for atomic operations on F# reference cells. |
This chapter covered concurrent, reactive, and asynchronous programming, which is a set of topics of growing importance in modern programming because of the widespread adoption of multicore microprocessors, network-aware applications, and asynchronous I/O channels. It discussed in depth background processing and a powerful F# construct called asynchronous workflows. Finally, the chapter covered applications of asynchronous workflows to message-processing agents and web crawling, and examined covered some of the shared-memory primitives for concurrent programming on the .NET platform. The next chapter looks at web programming, from serving web pages to delivering applications via web browsers.