Difference between revisions of "Typeclassopedia"
(adding diagram) 
Geheimdienst (talk  contribs) (Copied over the full text from the PDF version. Still raw. Now let's do some proofreading) 

Line 1:  Line 1:  
−  The Typeclassopedia, by Brent Yorgey, was an article in [[the Monad.Reader]] issue 13 explaining several of the standard Haskell typeclasses, including Monoid, Applicative, Alternative, Arrow, and Traversable. 

+  :''By Brent Yorgey, byorgey@cis.upenn.edu'' 

−  The Typeclassopedia itself can be found here: [[Media:Typeclassopedia.pdf]]. 

+  :''As published 12 March 2009, [http://www.haskell.org/wikiupload/8/85/TMRIssue13.pdf issue 13] of [http://themonadreader.wordpress.com/ the Monad.Reader]'' 

−  A summary of the entire article is at available at: http://typeclassopedia.citizen428.net/ 

+  :''Alternate formats: [[Media:Typeclassopedia.pdfPDF]] / [http://code.haskell.org/~byorgey/TMR/Issue13/typeclassopedia.lhs tex source] / [http://code.haskell.org/~byorgey/TMR/Issue13/typeclassopedia.bib bibliography]'' 

−  The entire issue of [[the Monad.Reader]] can be found here: http://www.haskell.org/wikiupload/8/85/TMRIssue13.pdf 

+  ''The standard Haskell libraries feature a number of type classes with algebraic or categorytheoretic underpinnings. Becoming a fluent Haskell hacker requires intimate familiarity with them all, yet acquiring this familiarity often involves combing through a mountain of tutorials, blog posts, mailing list archives, and IRC logs.'' 

+  ''The goal of this article is to serve as a starting point for the student of Haskell wishing to gain a firm grasp of its standard type classes. The essentials of each type class are introduced, with examples, commentary, and extensive references for further reading.'' 

+  
+  =Introduction= 

+  
+  Have you ever had any of the following thoughts? 

+  * What the heck is a monoid, and how is it different from a mon<u>a</u>d? 

+  
+  * I finally figured out how to use [[Parsec]] with donotation, and someone told me I should use something called <code>Applicative</code> instead. Um, what? 

+  
+  * Someone in the [[IRC#haskell]] IRC channel used <code>(***)</code>, and when I asked lambdabot to tell me its type, it printed out scary gobbledygook that didn't even fit on one line! Then someone used <code>fmap fmap fmap</code> and my brain exploded. 

+  
+  * When I asked how to do something I thought was really complicated, people started typing things like <code>zip.ap fmap.(id &&& wtf)</code> and the scary thing is that they worked! Anyway, I think those people must actually be robots because there's no way anyone could come up with that in two seconds off the top of their head. 

+  
+  If you have, look no further! You, too, can write and understand concise, elegant, idiomatic Haskell code with the best of them. 

+  
+  There are two keys to an expert Haskell hacker's wisdom: 

+  # Understand the types. 

+  # Gain a deep intuition for each type class and its relationship to other type classes, backed up by familiarity with many examples. 

+  
+  It's impossible to overstate the importance of the first; the patient student of type signatures will uncover many profound secrets. Conversely, anyone ignorant of the types in their code is doomed to eternal uncertainty. “Hmm, it doesn't compile ... maybe I'll stick in an 

+  <code>fmap</code> here ... nope, let's see ... maybe I need another <code>(.)</code> somewhere? ... um ...” 

+  
+  The second key—gaining deep intuition, backed by examples—is also important, but much more difficult to attain. A primary goal of this article is to set you on the road to gaining such intuition. However— 

+  
+  :''There is no royal road to Haskell. {{h:titleWell, he probably would have said it if he knew Haskell.—Euclid}}'' 

+  
+  {{note[http://byorgey.wordpress.com/2009/01/12/abstractionintuitionandthemonadtutorialfallacy/ Abstraction, intuition, and the “monad tutorial fallacy”, by Brent Yorgey]}} This article can only be a starting point, since good intuition comes from hard work, not from learning the right metaphor {{noteref}}. Anyone who reads and understands all of it will still have an arduous journey ahead—but sometimes a good starting point makes a big difference. 

+  
+  It should be noted that this is not a Haskell tutorial; it is assumed that the reader is already familiar with the basics of Haskell, including the standard <code>[http://haskell.org/ghc/docs/latest/html/libraries/base/Prelude.html Prelude]</code>, the type system, data types, and type classes. 

+  
+  {{box 

[[Image:Typeclassopediadiagram.png]] 
[[Image:Typeclassopediadiagram.png]] 

⚫  
+  The type classes we will be discussing and their interrelationships. 

+  * <u>Solid arrows</u> point from the general to the specific; that is, if there is an arrow from Foo to Bar it means that every Bar is (or should be, or can be made into) a Foo. 

+  * <span style{{=}}"borderbottom: 2px dotted black">Dotted arrows</span> indicate some other sort of relationship. 

+  * <code>Monad</code> and <code>ArrowApply</code> are equivalent. 

+  * <code>Pointed</code> and <code>Comonad</code> are greyed out since they are not actually (yet) in the standard Haskell libraries (they are in the [http://hackage.haskell.org/package/categoryextras categoryextras] library).}} 

+  
+  One more note before we begin. I've seen “type class” written as one word, “typeclass,” but let's settle this once and for all: the correct spelling uses two words (the title of this article notwithstanding), as evidenced by, for example, the [http://haskell.org/onlinereport/ Haskell 98 Revised Report], early papers on type classes like [http://citeseer.ist.psu.edu/viewdoc/summary?doi=10.1.1.103.5639 Type classes in Haskell] and [http://research.microsoft.com/enus/um/people/simonpj/papers/typeclassdesignspace/ Type classes: exploring the design space], and [http://citeseer.ist.psu.edu/viewdoc/summary?doi=10.1.1.168.4008 Hudak et al.'s history of Haskell]. 

+  
+  We now begin with the simplest type class of all: <code>Functor</code>. 

+  
+  =Functor= 

+  
+  The <code>Functor</code> class ([http://haskell.org/ghc/docs/latest/html/libraries/base/Prelude.html#t:Functor haddock]) is the most basic and ubiquitous type class in the Haskell libraries. A simple intuition is that a <code>Functor</code> represents a “container” of some sort, along with the ability to apply a function uniformly to every element in the container. For example, a list is a container of elements, and we can apply a function to every element of a list using <code>map</code>. A binary tree is also a container of elements, and it's not hard to come up with a way to recursively apply a function to every element in a tree. 

+  
+  Another intuition is that a <code>Functor</code> represents some sort of “computational context.” This intuition is generally more useful, but is more difficult to explain, precisely because it is so general. Some examples later should help to clarify the <code>Functor</code>ascontext point of view. 

+  
+  In the end, however, a <code>Functor</code> is simply what it is defined to be; doubtless there are many examples of <code>Functor</code> instances that don't exactly fit either of the above intuitions. The wise student will focus their attention on definitions and examples, without leaning too heavily on any particular metaphor. Intuition will come, in time, on its own. 

+  
+  ==Definition== 

+  
+  The type class declaration for <code>Functor</code>: 

+  
+  <haskell> 

+  class Functor f where 

+  fmap :: (a > b) > f a > f b 

+  </haskell> 

+  
+  <code>Functor</code> is exported by the <code>Prelude</code>, so no special imports are needed to use it. 

+  
+  First, the <code>f a</code> and <code>f b</code> in the type signature for <code>fmap</code> tell us that <code>f</code> isn't just a type; it is a ''type constructor'' which takes another type as a parameter. (A more precise way to say this is that the ''kind'' of <code>f</code> must be <code>* > *</code>.) For example, <code>Maybe</code> is such a type constructor: <code>Maybe</code> is not a type in and of itself, but requires another type as a parameter, like <code>Maybe Integer</code>. So it would not make sense to say <code>instance Functor Integer</code>, but it could make sense to say <code>instance Functor Maybe</code>. 

+  
+  Now look at the type of <code>fmap</code>: it takes any function from <code>a</code> to <code>b</code>, and a value of type <code>f a</code>, and outputs a value of type <code>f b</code>. From the container point of view, the intention is that <code>fmap</code> applies a function to each element of a container, without altering the structure of the container. From the context point of view, the intention is that <code>fmap</code> applies a function to a value without altering its context. Let's look at a few specific examples. 

+  
+  ==Instances== 

+  
+  {{noteRecall that <code>[]</code> has two meanings in Haskell: it can either stand for the empty list, or, as here, it can represent the list type constructor (pronounced “listof”). In other words, the type <code>[a]</code> (listof<code>a</code>) can also be written <code>([] a)</code>.}} 

+  
+  {{noteYou might ask why we need a separate <code>map</code> function. Why not just do away with the current listonly <code>map</code> function, and rename <code>fmap</code> to <code>map</code> instead? Well, that's a good question. The usual argument is that someone just learning Haskell, when using <code>map</code> incorrectly, would much rather see an error about lists than about <code>Functor</code>s.}} 

+  
+  As noted before, the list constructor <code>[]</code> is a functor {{noteref}}; we can use the standard list function <code>map</code> to apply a function to each element of a list {{noteref}}. The <code>Maybe</code> type constructor is also a functor, representing a container which might hold a single element. The function <code>fmap g</code> has no effect on <code>Nothing</code> (there are no elements to which <code>g</code> can be applied), and simply applies <code>g</code> to the single element inside a <code>Just</code>. Alternatively, under the context interpretation, the list functor represents a context of nondeterministic choice; that is, a list can be thought of as representing a single value which is nondeterministically chosen from among several possibilities (the elements of the list). Likewise, the <code>Maybe</code> functor represents a context with possible failure. These instances are: 

+  
+  <haskell> 

+  instance Functor [] where 

+  fmap _ [] = [] 

+  fmap g (x:xs) = g x : fmap g xs 

+   or we could just say fmap = map 

+  
+  instance Functor Maybe where 

+  fmap _ Nothing = Nothing 

+  fmap g (Just a) = Just (g a) 

+  </haskell> 

+  
+  As an aside, in idiomatic Haskell code you will often see the letter <code>f</code> used to stand for both an arbitrary <code>Functor</code> and an arbitrary function. In this tutorial, I will use <code>f</code> only to represent <code>Functor</code>s, and <code>g</code> or <code>h</code> to represent functions, but you should be aware of the potential confusion. In practice, what <code>f</code> stands for should always be clear from the context, by noting whether it is part of a type or part of the code. 

+  
+  {{noteNote that some of these instances are not exported by the <code>Prelude</code>; to access them, you can import <code>Control.Monad.Instances</code>.}} There are other <code>Functor</code> instances in the standard libraries; here are a few {{noteref}}: 

+  
+  * <code>Either e</code> is an instance of <code>Functor</code>; <code>Either e a</code> represents a container which can contain either a value of type <code>a</code>, or a value of type <code>e</code> (often representing some sort of error condition). It is similar to <code>Maybe</code> in that it represents possible failure, but it can carry some extra information about the failure as well. 

+  
+  * <code>((,) e)</code> represents a container which holds an “annotation” of type <code>e</code> along with the actual value it holds. 

+  
+  * <code>((>) e)</code>, the type of functions which take a value of type <code>e</code> as a parameter, is a <code>Functor</code>. It would be clearer to write it as <code>(e >)</code>, by analogy with an operator section like <code>(1+)</code>, but that syntax is not allowed. However, you can certainly ''think'' of it as <code>(e >)</code>. As a container, <code>(e > a)</code> represents a (possibly infinite) set of values of <code>a</code>, indexed by values of <code>e</code>. Alternatively, and more usefully, <code>(e >)</code> can be thought of as a context in which a value of type <code>e</code> is available to be consulted in a readonly fashion. This is also why <code>((>) e)</code> is sometimes referred to as the ''reader monad''; more on this later. 

+  
+  * <code>IO</code> is a <code>Functor</code>; a value of type <code>IO a</code> represents a computation producing a value of type <code>a</code> which may have I/O effects. If <code>m</code> computes the value <code>x</code> while producing some I/O effects, then <code>fmap g m</code> will compute the value <code>g x</code> while producing the same I/O effects. 

+  
+  * Many standard types from the [http://hackage.haskell.org/packages/archive/containers/0.2.0.0/doc/html/index.html containers library] (such as <code>Tree</code>, <code>Map</code>, <code>Sequence</code>, and <code>Stream</code>) are instances of <code>Functor</code>. A notable exception is <code>Set</code>, which cannot be made a <code>Functor</code> in Haskell (although it is certainly a mathematical functor) since it requires an <code>Ord</code> constraint on its elements; <code>fmap</code> must be applicable to ''any'' types <code>a</code> and <code>b</code>. 

+  
+  A good exercise is to implement <code>Functor</code> instances for <code>Either e</code>, <code>((,) e)</code>, and <code>((>) e)</code>. 

+  
+  ==Laws== 

+  
+  As far as the Haskell language itself is concerned, the only requirement to be a <code>Functor</code> is an implementation of <code>fmap</code> with the proper type. Any sensible <code>Functor</code> instance, however, will also satisfy the ''functor laws'', which are part of the definition of a mathematical functor. There are two: 

+  
+  <haskell> 

+  fmap id = id 

+  fmap (g . h) = (fmap g) . (fmap h) 

+  </haskell> 

+  
+  {{noteTechnically, these laws make <code>f</code> and <code>fmap</code> together an endofunctor on ''Hask'', the category of Haskell types (ignoring [[Bottom⊥]], which is a party pooper). See [http://en.wikibooks.org/wiki/Haskell/Category_theory Wikibook: Category theory].}} 

+  
+  Together, these laws ensure that <code>fmap g</code> does not change the ''structure'' of a container, only the elements. Equivalently, and more simply, they ensure that <code>fmap g</code> changes a value without altering its context {{noteref}}. 

+  
+  The first law says that mapping the identity function over every item in a container has no effect. The second says that mapping a composition of two functions over every item in a container is the same as first mapping one function, and then mapping the other. 

+  
+  As an example, the following code is a “valid” instance of <code>Functor</code> (it typechecks), but it violates the functor laws. Do you see why? 

+  
+  <haskell> 

+   Evil Functor instance 

+  instance Functor [] where 

+  fmap _ [] = [] 

+  fmap g (x:xs) = g x : g x : fmap g xs 

+  </haskell> 

+  
+  Any Haskeller worth their salt would reject this code as a gruesome abomination. 

+  
+  ==Intuition== 

+  
+  There are two fundamental ways to think about <code>fmap</code>. The first has already been touched on: it takes two parameters, a function and a container, and applies the function “inside” the container, producing a new container. Alternately, we can think of <code>fmap</code> as applying a function to a value in a context (without altering the context). 

+  
+  Just like all other Haskell functions of “more than one parameter,” however, <code>fmap</code> is actually ''curried'': it does not really take two parameters, but takes a single parameter and returns a function. For emphasis, we can write <code>fmap</code>'s type with extra parentheses: <code>fmap :: (a > b) > (f a > f b)</code>. Written in this form, it is apparent that <code>fmap</code> transforms a “normal” function (<code>g :: a > b</code>) into one which operates over containers/contexts (<code>fmap g :: f a > f b</code>). This transformation is often referred to as a ''lift''; <code>fmap</code> “lifts” a function from the “normal world” into the “<code>f</code> world.” 

+  
+  ==Further reading== 

+  
+  A good starting point for reading about the category theory behind the concept of a functor is the excellent [http://en.wikibooks.org/wiki/Haskell/Category_theory Haskell wikibook page on category theory]. 

+  
+  =Pointed= 

+  
+  {{noteIt is, however, a type class in the [http://hackage.haskell.org/package/categoryextras categoryextras] library.}} 

+  
+  The <code>Pointed</code> type class represents ''pointed functors''. It is not actually a type class in the standard libraries {{noteref}}. But it ''could'' be, and it's useful in understanding a few other type classes, notably <code>Applicative</code> and <code>Monad</code>, so let's pretend for a minute. 

+  
+  Given a <code>Functor</code>, the <code>Pointed</code> class represents the additional ability to put a value into a “default context.” Often, this corresponds to creating a container with exactly one element, but it is more general than that. The type class declaration for <code>Pointed</code> is: 

+  
+  <haskell> 

+  class Functor f => Pointed f where 

+  pure :: a > f a  aka singleton, return, unit, point 

+  </haskell> 

+  
+  Most of the standard <code>Functor</code> instances could also be instances of <code>Pointed</code>—for example, the <code>Maybe</code> instance of <code>Pointed</code> is <code>pure = Just</code>; there are many possible implementations for lists, the most natural of which is <code>pure x = [x]</code>; for <code>((>) e)</code> it is ... well, I'll let you work it out. (Just follow the types!) 

+  
+  One example of a <code>Functor</code> which is not <code>Pointed</code> is <code>((,) e)</code>. If you try implementing <code>pure :: a > (e,a)</code> you will quickly see why: since the type <code>e</code> is completely arbitrary, there is no way to generate a value of type <code>e</code> out of thin air! However, as we will see, <code>((,) e)</code> can be made <code>Pointed</code> if we place an additional restriction on <code>e</code> which allows us to generate a default value of type <code>e</code> (the most common solution is to make <code>e</code> an instance of <code>Monoid</code>). 

+  
+  {{noteFor those interested in category theory, this law states precisely that <code>pure</code> is a natural transformation from the identity functor to <code>f</code>.}} The <code>Pointed</code> class has only one law {{noteref}}: 

+  
+  <haskell> 

+  fmap g . pure = pure . g 

+  </haskell> 

+  
+  {{note... modulo ⊥, <code>seq</code>, and assuming a lawful <code>Functor</code> instance.}} 

+  
+  However, you need not worry about it: this law is actually a socalled “free theorem” guaranteed by parametricity (see Wadler's [http://doi.acm.org/10.1145/99370.99404 Theorems for free!]); it's impossible to write an instance of <code>Pointed</code> which does not satisfy it {{noteref}}. 

+  
+  =Applicative= 

+  
+  A somewhat newer addition to the pantheon of standard Haskell type classes, ''applicative functors'' (see [http://haskell.org/ghc/docs/latest/html/libraries/base/ControlApplicative.html their haddock]) represent an abstraction lying exactly in between <code>Functor</code> and <code>Monad</code>, first described by McBride and Paterson. The title of their classic paper, [http://www.soi.city.ac.uk/~ross/papers/Applicative.html Applicative Programming with Effects], gives a hint at the intended intuition behind the <code>Applicative</code> type class. It encapsulates certain sorts of “effectful” computations in a functionally pure way, and encourages an “applicative” programming style. Exactly what these things mean will be seen later. 

+  
+  ==Definition== 

+  
+  The <code>Applicative</code> class adds a single capability to <code>Pointed</code> functors. Recall that <code>Functor</code> allows us to lift a “normal” function to a function on computational contexts. But <code>fmap</code> doesn't allow us to apply a function which is itself in a context to a value in another context. <code>Applicative</code> gives us just such a tool. Here is the type class declaration for <code>Applicative</code>, as defined in <code>Control.Applicative</code>: 

+  
+  <haskell> 

+  class Functor f => Applicative f where 

+  pure :: a > f a 

+  (<*>) :: f (a > b) > f a > f b 

+  </haskell> 

+  
+  Note that every <code>Applicative</code> must also be a <code>Functor</code>. In fact, as we will see, <code>fmap</code> can be implemented using the <code>Applicative</code> methods, so every <code>Applicative</code> is a functor whether we like it or not; the <code>Functor</code> constraint forces us to be honest. 

+  
+  {{noteRecall that <code>($)</code> is just function application: <code>f $ x {{=}} f x</code>.}} 

+  
+  As always, it's crucial to understand the type signature of <code>(<*>)</code>. The best way of thinking about it comes from noting that the type of <code>(<*>)</code> is similar to the type of <code>($)</code> {{noteref}}, but with everything enclosed in an <code>f</code>. In other words, <code>(<*>)</code> is just function application within a computational context. The type of <code>(<*>)</code> is also very similar to the type of <code>fmap</code>; the only difference is that the first parameter is <code>f (a > b)</code>, a function in a context, instead of a “normal” function <code>(a > b)</code>. 

+  
+  Of course, <code>pure</code> looks rather familiar. If we actually had a <code>Pointed</code> type class, <code>Applicative</code> could instead be defined as: 

+  
+  <haskell> 

+  class Pointed f => Applicative' f where 

+  (<*>) :: f (a > b) > f a > f b 

+  </haskell> 

+  
+  ==Laws== 

+  
+  {{noteSee 

+  [http://haskell.org/ghc/docs/latest/html/libraries/base/ControlApplicative.html haddock for Applicative], [http://www.soi.city.ac.uk/~ross/papers/Applicative.html Applicative programming with effects]}} 

+  
+  There are several laws that <code>Applicative</code> instances should satisfy {{noteref}}, but only one is crucial to developing intuition, because it specifies how <code>Applicative</code> should relate to <code>Functor</code> (the other four mostly specify the exact sense in which <code>pure</code> deserves its name). This law is: 

+  
+  <haskell> 

+  fmap g x = pure g <*> x 

+  </haskell> 

+  
+  It says that mapping a pure function <code>g</code> over a context <code>x</code> is the same as first injecting <code>g</code> into a context with <code>pure</code>, and then applying it to <code>x</code> with <code>(<*>)</code>. In other words, we can decompose <code>fmap</code> into two more atomic operations: injection into a context, and application within a context. The <code>Control.Applicative</code> module also defines <code>(<$>)</code> as a synonym for <code>fmap</code>, so the above law can also be expressed as: 

+  
+  <code>g <$> x = pure g <*> x</code>. 

+  
+  ==Instances== 

+  
+  Most of the standard types which are instances of <code>Functor</code> are also instances of <code>Applicative</code>. 

+  
+  <code>Maybe</code> can easily be made an instance of <code>Applicative</code>; writing such an instance is left as an exercise for the reader. 

+  
+  The list type constructor <code>[]</code> can actually be made an instance of <code>Applicative</code> in two ways; essentially, it comes down to whether we want to think of lists as ordered collections of elements, or as contexts representing multiple results of a nondeterministic computation (see Wadler's [http://www.springerlink.com/content/y7450255v2670167/ How to replace failure by a list of successes]). 

+  
+  Let's first consider the collection point of view. Since there can only be one instance of a given type class for any particular type, one or both of the list instances of <code>Applicative</code> need to be defined for a <code>newtype</code> wrapper; as it happens, the nondeterministic computation instance is the default, and the collection instance is defined in terms of a <code>newtype</code> called <code>ZipList</code>. This instance is: 

+  
+  <haskell> 

+  newtype ZipList a = ZipList { getZipList :: [a] } 

+  
+  instance Applicative ZipList where 

+  pure = undefined  exercise 

+  (ZipList gs) <*> (ZipList xs) = ZipList (zipWith ($) gs xs) 

+  </haskell> 

+  
+  To apply a list of functions to a list of inputs with <code>(<*>)</code>, we just match up the functions and inputs elementwise, and produce a list of the resulting outputs. In other words, we “zip” the lists together with function application, <code>($)</code>; hence the name <code>ZipList</code>. As an exercise, determine the correct definition of <code>pure</code>—there is only one implementation that satisfies the law (see section “Laws”). 

+  
+  The other <code>Applicative</code> instance for lists, based on the nondeterministic computation point of view, is: 

+  
+  <haskell> 

+  instance Applicative [] where 

+  pure x = [x] 

+  gs <*> xs = [ g x  g < gs, x < xs ] 

+  </haskell> 

+  
+  Instead of applying functions to inputs pairwise, we apply each function to all the inputs in turn, and collect all the results in a list. 

+  
+  Now we can write nondeterministic computations in a natural style. To add the numbers <code>3</code> and <code>4</code> deterministically, we can of course write <code>(+) 3 4</code>. But suppose instead of <code>3</code> we have a nondeterministic computation that might result in <code>2</code>, <code>3</code>, or <code>4</code>; then we can write 

+  
+  <haskell> 

+  pure (+) <*> [2,3,4] <*> pure 4 

+  </haskell> 

+  
+  or, more idiomatically, 

+  
+  <haskell> 

+  (+) <$> [2,3,4] <*> pure 4. 

+  </haskell> 

+  
+  There are several other <code>Applicative</code> instances as well: 

+  
+  * <code>IO</code> is an instance of <code>Applicative</code>, and behaves exactly as you would think: when <code>g <$> m1 <*> m2 <*> m3</code> is executed, the effects from the <code>mi</code>'s happen in order from left to right. 

+  
+  * <code>((,) a)</code> is an <code>Applicative</code>, as long as <code>a</code> is an instance of <code>Monoid</code> ([[#Monoidsection Monoid]]). The <code>a</code> values are accumulated in parallel with the computation. 

+  
+  * The <code>Applicative</code> module defines the <code>Const</code> type constructor; a value of type <code>Const a b</code> simply contains an <code>a</code>. This is an instance of <code>Applicative</code> for any <code>Monoid a</code>; this instance becomes especially useful in conjunction with things like <code>Foldable</code> ([[#Foldablesection Foldable]]). 

+  
+  * The <code>WrappedMonad</code> and <code>WrappedArrow</code> newtypes make any instances of <code>Monad</code> ([[#Monadsection Monad]]) or <code>Arrow</code> ([[#Arrowsection Arrow]]) respectively into instances of <code>Applicative</code>; as we will see when we study those type classes, both are strictly more expressive than <code>Applicative</code>, in the sense that the <code>Applicative</code> methods can be implemented in terms of their methods. 

+  
+  ==Intuition== 

+  
+  McBride and Paterson's paper introduces the notation <math>g \; x_1 \; x_2 \; \cdots \; x_n</math> to denote function application in a computational context. If each <math>x_i</math> has type <math>f \; t_i</math> for some applicative functor <math>f</math>, and <math>g</math> has type <math>t_1 \to t_2 \to \dots \to t_n \to t</math>, then the entire expression <math>g \; x_1 \; \cdots \; x_n</math> has type <math>f \; t</math>. You can think of this as applying a function to multiple “effectful” arguments. In this sense, the double bracket notation is a generalization of <code>fmap</code>, which allows us to apply a function to a single argument in a context. 

+  
+  Why do we need <code>Applicative</code> to implement this generalization of <code>fmap</code>? Suppose we use <code>fmap</code> to apply <code>g</code> to the first parameter <code>x1</code>. Then we get something of type <code>f (t2 > ... t)</code>, but now we are stuck: we can't apply this functioninacontext to the next argument with <code>fmap</code>. However, this is precisely what <code>(<*>)</code> allows us to do. 

+  
+  This suggests the proper translation of the idealized notation <math>g \; x_1 \; x_2 \; \cdots \; x_n</math> into Haskell, namely 

+  <haskell> 

+  g <$> x1 <*> x2 <*> ... <*> xn, 

+  </haskell> 

+  
+  recalling that <code>Control.Applicative</code> defines <code>(<$>)</code> as convenient infix shorthand for <code>fmap</code>. This is what is meant by an “applicative style”—effectful computations can still be described in terms of function application; the only difference is that we have to use the special operator <code>(<*>)</code> for application instead of simple juxtaposition. 

+  
+  ==Further reading== 

+  
+  There are many other useful combinators in the standard libraries implemented in terms of <code>pure</code> and <code>(<*>)</code>: for example, <code>(*>)</code>, <code>(<*)</code>, <code>(<**>)</code>, <code>(<$)</code>, and so on (see [http://haskell.org/ghc/docs/latest/html/libraries/base/ControlApplicative.html haddock for Applicative]). Judicious use of such secondary combinators can often make code using <code>Applicative</code>s much easier to read. 

+  
+  [http://www.soi.city.ac.uk/~ross/papers/Applicative.html McBride and Paterson's original paper] is a treasuretrove of information and examples, as well as some perspectives on the connection between <code>Applicative</code> and category theory. Beginners will find it difficult to make it through the entire paper, but it is extremely wellmotivated—even beginners will be able to glean something from reading as far as they are able. 

+  
+  Conal Elliott has been one of the biggest proponents of <code>Applicative</code>. For example, the [http://conal.net/papers/functionalimages/ Pan library for functional images] and the [http://conal.net/papers/simplyreactive/ reactive library for functional reactive programming (FRP)] make key use of it; his blog also contains [http://conal.net/blog/tag/applicativefunctor many examples of <code>Applicative</code> in action]. Building on the work of McBride and Paterson, Elliott also built the [[TypeCompose]] library, which embodies the observation (among others) that <code>Applicative</code> types are closed under composition; therefore, <code>Applicative</code> instances can often be automatically derived for complex types built out of simpler ones. 

+  
+  Although the [http://legacy.cs.uu.nl/daan/parsec.html Parsec parsing library] ([http://legacy.cs.uu.nl/daan/download/papers/parsecpaper.pdf paper]) was originally designed for use as a monad, in its most common use cases an <code>Applicative</code> instance can be used to great effect; [http://www.serpentine.com/blog/2008/02/06/thebasicsofapplicativefunctorsputtopracticalwork/ Bryan O'Sullivan's blog post] is a good starting point. If the extra power provided by <code>Monad</code> isn't needed, it's usually a good idea to use <code>Applicative</code> instead. 

+  
+  A couple other nice examples of <code>Applicative</code> in action include the [http://chrisdone.com/blog/html/20090210applicativeconfigfilehsql.html ConfigFile and HSQL libraries] and the [http://groups.inf.ed.ac.uk/links/formlets/ formlets library]. 

+  
+  =Monad= 

+  
+  It's a safe bet that if you're reading this article, you've heard of monads—although it's quite possible you've never heard of <code>Applicative</code> before, or <code>Arrow</code>, or even <code>Monoid</code>. Why are monads such a big deal in Haskell? There are several reasons. 

+  
+  * Haskell does, in fact, single out monads for special attention by making them the framework in which to construct I/O operations. 

+  * Haskell also singles out monads for special attention by providing a special syntactic sugar for monadic expressions: the <code>do</code>notation. 

+  * <code>Monad</code> has been around longer than various other abstract models of computation such as <code>Applicative</code> or <code>Arrow</code>. 

+  * The more monad tutorials there are, the harder people think monads must be, and the more new monad tutorials are written by people who think they finally “get” monads (the [http://byorgey.wordpress.com/2009/01/12/abstractionintuitionandthemonadtutorialfallacy/ monad tutorial fallacy]). 

+  
+  I will let you judge for yourself whether these are good reasons. 

+  
+  In the end, despite all the hoopla, <code>Monad</code> is just another type class. Let's take a look at its definition. 

+  
+  ==Definition== 

+  
+  The type class declaration for <code>Monad</code> ([http://haskell.org/ghc/docs/latest/html/libraries/base/Prelude.html#t:Monad haddock]) is: 

+  
+  <haskell> 

+  class Monad m where 

+  return :: a > m a 

+  (>>=) :: m a > (a > m b) > m b 

+  (>>) :: m a > m b > m b 

+  m >> n = m >>= \_ > n 

+  
+  fail :: String > m a 

+  </haskell> 

+  
+  The <code>Monad</code> type class is exported by the <code>Prelude</code>, along with a few standard instances. However, many utility functions are found in <code>Control.Monad</code>, and there are also several instances (such as <code>((>) e)</code>) defined in <code>Control.Monad.Instances</code>. 

+  
+  Let's examine the methods in the <code>Monad</code> class one by one. The type of <code>return</code> should look familiar; it's the same as <code>pure</code>. Indeed, <code>return</code> ''is'' <code>pure</code>, but with an unfortunate name. (Unfortunate, since someone coming from an imperative programming background might think that <code>return</code> is like the C or Java keyword of the same name, when in fact the similarities are minimal.) From a mathematical point of view, every monad is a pointed functor (indeed, an applicative functor), but for historical reasons, the <code>Monad</code> type class declaration unfortunately does not require this. 

+  
+  We can see that <code>(>>)</code> is a specialized version of <code>(>>=)</code>, with a default implementation given. It is only included in the type class declaration so that specific instances of <code>Monad</code> can override the default implementation of <code>(>>)</code> with a more efficient one, if desired. Also, note that although <code>_ >> n = n</code> would be a typecorrect implementation of <code>(>>)</code>, it would not correspond to the intended semantics: the intention is that <code>m >> n</code> ignores the ''result'' of <code>m</code>, but not its ''effects''. 

+  
+  The <code>fail</code> function is an awful hack that has no place in the <code>Monad</code> class; more on this later. 

+  
+  The only really interesting thing to look at—and what makes <code>Monad</code> strictly more powerful than <code>Pointed</code> or <code>Applicative</code>—is <code>(>>=)</code>, which is often called ''bind''. An alternative definition of <code>Monad</code> could look like: 

+  
+  <haskell> 

+  class Applicative m => Monad' m where 

+  (>>=) :: m a > (a > m b) > m b 

+  </haskell> 

+  
+  We could spend a while talking about the intuition behind <code>(>>=)</code>—and we will. But first, let's look at some examples. 

+  
+  ==Instances== 

+  
+  Even if you don't understand the intuition behind the <code>Monad</code> class, you can still create instances of it by just seeing where the types lead you. You may be surprised to find that this actually gets you a long way towards understanding the intuition; at the very least, it will give you some concrete examples to play with as you read more about the <code>Monad</code> class in general. The first few examples are from the standard <code>Prelude</code>; the remaining examples are from the monad transformer library ([http://hackage.haskell.org/package/mtl mtl]). 

+  
+  * The simplest possible instance of <code>Monad</code> is <code>Identity</code> (see [http://hackage.haskell.org/packages/archive/mtl/1.1.0.2/doc/html/ControlMonadIdentity.html haddock]), which is described in Dan Piponi's highly recommended blog post on [http://blog.sigfpe.com/2007/04/trivialmonad.html The Trivial Monad]. Despite being “trivial,” it is a great introduction to the <code>Monad</code> type class, and contains some good exercises to get your brain working. 

+  * The next simplest instance of <code>Monad</code> is <code>Maybe</code>. We already know how to write <code>return</code>/<code>pure</code> for <code>Maybe</code>. So how do we write <code>(>>=)</code>? Well, let's think about its type. Specializing for <code>Maybe</code>, we have 

+  <haskell> 

+  (>>=) :: Maybe a > (a > Maybe b) > Maybe b. 

+  </haskell> 

+  
+  :If the first argument to <code>(>>=)</code> is <code>Just x</code>, then we have something of type <code>a</code> (namely, <code>x</code>), to which we can apply the second argument—resulting in a <code>Maybe b</code>, which is exactly what we wanted. What if the first argument to <code>(>>=)</code> is <code>Nothing</code>? In that case, we don't have anything to which we can apply the <code>a > Maybe b</code> function, so there's only one thing we can do: yield <code>Nothing</code>. This instance is: 

+  
+  <haskell> 

+  instance Monad Maybe where 

+  return = Just 

+  (Just x) >>= g = g x 

+  Nothing >>= _ = Nothing 

+  </haskell> 

+  
+  :We can already get a bit of intuition as to what is going on here: if we build up a computation by chaining together a bunch of functions with <code>(>>=)</code>, as soon as any one of them fails, the entire computation will fail (because <code>Nothing >>= f</code> is <code>Nothing</code>, no matter what <code>f</code> is). The entire computation succeeds only if all the constituent functions individually succeed. So the <code>Maybe</code> monad models computations which may fail. 

+  
+  * The <code>Monad</code> instance for the list constructor <code>[]</code> is similar to its <code>Applicative</code> instance; I leave its implementation as an exercise. Follow the types! 

+  
+  * Of course, the <code>IO</code> constructor is famously a <code>Monad</code>, but its implementation is somewhat magical, and may in fact differ from compiler to compiler. It is worth emphasizing that the <code>IO</code> monad is the ''only'' monad which is magical. It allows us to build up, in an entirely pure way, values representing possibly effectful computations. The special value <code>main</code>, of type <code>IO ()</code>, is taken by the runtime and actually executed, producing actual effects. Every other monad is functionally pure, and requires no special compiler support. We often speak of monadic values as “effectful computations,” but this is because some monads allow us to write code ''as if'' it has side effects, when in fact the monad is hiding the plumbing which allows these apparent side effects to be implemented in a functionally pure way. 

+  
+  * As mentioned earlier, <code>((>) e)</code> is known as the ''reader monad'', since it describes computations in which a value of type <code>e</code> is available as a readonly environment. It is worth trying to write a <code>Monad</code> instance for <code>((>) e)</code> yourself. 

+  
+  :The <code>Control.Monad.Reader</code> module ([http://hackage.haskell.org/packages/archive/mtl/1.1.0.2/doc/html/ControlMonadReader.html haddock]) provides the <code>Reader e a</code> type, which is just a convenient <code>newtype</code> wrapper around <code>(e > a)</code>, along with an appropriate <code>Monad</code> instance and some <code>Reader</code>specific utility functions such as <code>ask</code> (retrieve the environment), <code>asks</code> (retrieve a function of the environment), and <code>local</code> (run a subcomputation under a different environment). 

+  
+  * The <code>Control.Monad.Writer</code> module ([http://hackage.haskell.org/packages/archive/mtl/1.1.0.2/doc/html/ControlMonadWriterLazy.html haddock]) provides the <code>Writer</code> monad, which allows information to be collected as a computation progresses. <code>Writer w a</code> is isomorphic to <code>(a,w)</code>, where the output value <code>a</code> is carried along with an annotation or “log” of type <code>w</code>, which must be an instance of <code>Monoid</code> ([[#Monoidsection Monoid]]); the special function <code>tell</code> performs logging. 

+  
+  * The <code>Control.Monad.State</code> module ([http://hackage.haskell.org/packages/archive/mtl/1.1.0.2/doc/html/ControlMonadStateLazy.html haddock]) provides the <code>State s a</code> type, a <code>newtype</code> wrapper around <code>s > (a,s)</code>. Something of type <code>State s a</code> represents a stateful computation which produces an <code>a</code> but can access and modify the state of type <code>s</code> along the way. The module also provides <code>State</code>specific utility functions such as <code>get</code> (read the current state), <code>gets</code> (read a function of the current state), <code>put</code> (overwrite the state), and <code>modify</code> (apply a function to the state). 

+  
+  * The <code>Control.Monad.Cont</code> module ([http://hackage.haskell.org/packages/archive/mtl/1.1.0.2/doc/html/ControlMonadCont.html haddock]) provides the <code>Cont</code> monad, which represents computations in continuationpassing style. It can be used to suspend and resume computations, and to implement nonlocal transfers of control, coroutines, other complex control structures—all in a functionally pure way. <code>Cont</code> has been called the [http://blog.sigfpe.com/2008/12/motherofallmonads.html “mother of all monads”] because of its universal properties. 

+  
+  ==Intuition== 

+  
+  Let's look more closely at the type of <code>(>>=)</code>. The basic intuition is that it combines two computations into one larger computation. The first argument, <code>m a</code>, is the first computation. However, it would be boring if the second argument were just an <code>m b</code>; then there would be no way for the computations to interact with one another. So, the second argument to <code>(>>=)</code> has type <code>a > m b</code>: a function of this type, given a ''result'' of the first computation, can produce a second computation to be run. In other words, <code>x >>= k</code> is a computation which runs <code>x</code>, and then uses the result(s) of <code>x</code> to ''decide'' what computation to run second, using the output of the second computation as the result of the entire computation. 

+  
+  Intuitively, it is this ability to use the output from previous computations to decide what computations to run next that makes <code>Monad</code> more powerful than <code>Applicative</code>. The structure of an <code>Applicative</code> computation is fixed, whereas the structure of a <code>Monad</code> computation can change based on intermediate results. 

+  
+  To see the increased power of <code>Monad</code> from a different point of view, let's see what happens if we try to implement <code>(>>=)</code> in terms of <code>fmap</code>, <code>pure</code>, and <code>(<*>)</code>. We are given a value <code>x</code> of type <code>m a</code>, and a function <code>k</code> of type <code>a > m b</code>, so the only thing we can do is apply <code>k</code> to <code>x</code>. We can't apply it directly, of course; we have to use <code>fmap</code> to lift it over the <code>m</code>. But what is the type of <code>fmap k</code>? Well, it's <code>m a > m (m b)</code>. So after we apply it to <code>x</code>, we are left with something of type <code>m (m b)</code>—but now we are stuck; what we really want is an <code>m b</code>, but there's no way to get there from here. We can ''add'' <code>m</code>'s using <code>pure</code>, but we have no way to ''collapse'' multiple <code>m</code>'s into one. 

+  
+  This ability to collapse multiple <code>m</code>'s is exactly the ability provided by the function <code>join :: m (m a) > m a</code>, and it should come as no surprise that an alternative definition of <code>Monad</code> can be given in terms of <code>join</code>: 

+  
+  <haskell> 

+  class Applicative m => Monad'' m where 

+  join :: m (m a) > m a 

+  </haskell> 

+  
+  In fact, monads in category theory are defined in terms of <code>return</code>, <code>fmap</code>, and <code>join</code> (often called <math>\eta</math>, <math>T</math>, and <math>\mu</math> in the mathematical literature). Haskell uses the equivalent formulation in terms of <code>(>>=)</code> instead of <code>join</code> since it is more convenient to use; however, sometimes it can be easier to think about <code>Monad</code> instances in terms of <code>join</code>, since it is a more “atomic” operation. (For example, <code>join</code> for the list monad is just <code>concat</code>.) An excellent exercise is to implement <code>(>>=)</code> in terms of <code>fmap</code> and <code>join</code>, and to implement <code>join</code> in terms of <code>(>>=)</code>. 

+  
+  ==Utility functions== 

+  
+  The <code>Control.Monad</code> module ([http://haskell.org/ghc/docs/latest/html/libraries/base/ControlMonad.html haddock]) provides a large number of convenient utility functions, all of which can be implemented in terms of the basic <code>Monad</code> operations (<code>return</code> and <code>(>>=)</code> in particular). We have already seen one of them, namely, <code>join</code>. We also mention some other noteworthy ones here; implementing these utility functions oneself is a good exercise. For a more detailed guide to these functions, with commentary and example code, see HenkJan van Tuyl's [http://members.chello.nl/hjgtuyl/tourdemonad.html tour]. 

+  
+  * <code>liftM :: Monad m => (a > b) > m a > m b</code>. This should be familiar; of course, it is just <code>fmap</code>. The fact that we have both <code>fmap</code> and <code>liftM</code> is an unfortunate consequence of the fact that the <code>Monad</code> type class does not require a <code>Functor</code> instance, even though mathematically speaking, every monad is a functor. However, <code>fmap</code> and <code>liftM</code> are essentially interchangeable, since it is a bug (in a social rather than technical sense) for any type to be an instance of <code>Monad</code> without also being an instance of <code>Functor</code>. 

+  
+  * <code>ap :: Monad m => m (a > b) > m a > m b</code> should also be familiar: it is equivalent to <code>(<*>)</code>, justifying the claim that the <code>Monad</code> interface is strictly more powerful than <code>Applicative</code>. We can make any <code>Monad</code> into an instance of <code>Applicative</code> by setting <code>pure = return</code> and <code>(<*>) = ap</code>. 

+  
+  * <code>sequence :: Monad m => [m a] > m [a]</code> takes a list of computations and combines them into one computation which collects a list of their results. It is again something of a historical accident that <code>sequence</code> has a <code>Monad</code> constraint, since it can actually be implemented only in terms of <code>Applicative</code>. There is also an additional generalization of <code>sequence</code> to structures other than lists, which will be discussed in the [[#Traversablesection on <code>Traversable</code>]]. 

+  
+  * <code>replicateM :: Monad m => Int > m a > m [a]</code> is simply a combination of <code>replicate</code> and <code>sequence</code>. 

+  
+  * <code>when :: Monad m => Bool > m () > m ()</code> conditionally executes a computation, evaluating to its second argument if the test is <code>True</code>, and to <code>return ()</code> if the test is <code>False</code>. A collection of other sorts of monadic conditionals can be found in the [http://hackage.haskell.org/package/IfElse IfElse package]. 

+  
+  * <code>mapM :: Monad m => (a > m b) > [a] > m [b]</code> maps its first argument over the second, and <code>sequence</code>s the results. The <code>forM</code> function is just <code>mapM</code> with its arguments reversed; it is called <code>forM</code> since it models generalized <code>for</code> loops: the list <code>[a]</code> provides the loop indices, and the function <code>a > m b</code> specifies the “body” of the loop for each index. 

+  
+  * <code>(=<<) :: Monad m => (a > m b) > m a > m b</code> is just <code>(>>=)</code> with its arguments reversed; sometimes this direction is more convenient since it corresponds more closely to function application. 

+  
+  * <code>(>=>) :: Monad m => (a > m b) > (b > m c) > a > m c</code> is sort of like function composition, but with an extra <code>m</code> on the result type of each function, and the arguments swapped. We'll have more to say about this operation later. 

+  
+  * The <code>guard</code> function is for use with instances of <code>MonadPlus</code>, which is discussed at the end of the <code>Monoid</code> section. 

+  
+  Many of these functions also have “underscored” variants, such as <code>sequence_</code> and <code>mapM_</code>; these variants throw away the results of the computations passed to them as arguments, using them only for their side effects. 

+  
+  ==Laws== 

+  
+  There are several laws that instances of <code>Monad</code> should satisfy [[Monad laws]]. The standard presentation is: 

+  
+  <haskell> 

+  return a >>= k = k a 

+  m >>= return = m 

+  m >>= (\x > k x >>= h) = (m >>= k) >>= h 

+  
+  fmap f xs = xs >>= return . f = liftM f xs 

+  </haskell> 

+  
+  The first and second laws express the fact that <code>return</code> behaves nicely: if we inject a value <code>a</code> into a monadic context with <code>return</code>, and then bind to <code>k</code>, it is the same as just applying <code>k</code> to <code>a</code> in the first place; if we bind a computation <code>m</code> to <code>return</code>, nothing changes. The third law essentially says that <code>(>>=)</code> is associative, sort of. The last law ensures that <code>fmap</code> and <code>liftM</code> are the same for types which are instances of both <code>Functor</code> and <code>Monad</code>—which, as already noted, should be every instance of <code>Monad</code>. 

+  
+  {{noteI like to pronounce this operator “fish,” but that's probably not the canonical pronunciation ...}} 

+  
+  However, the presentation of the above laws, especially the third, is marred by the asymmetry of <code>(>>=)</code>. It's hard to look at the laws and see what they're really saying. I prefer a much more elegant version of the laws, which is formulated in terms of <code>(>=>)</code> {{noteref}}. Recall that <code>(>=>)</code> “composes” two functions of type <code>a > m b</code> and <code>b > m c</code>. You can think of something of type <code>a > m b</code> (roughly) as a function from <code>a</code> to <code>b</code> which may also have some sort of effect in the context corresponding to <code>m</code>. (Note that <code>return</code> is such a function.) <code>(>=>)</code> lets us compose these “effectful functions,” and we would like to know what properties <code>(>=>)</code> has. The monad laws reformulated in terms of <code>(>=>)</code> are: 

+  
+  <haskell> 

+  return >=> g = g 

+  g >=> return = g 

+  (g >=> h) >=> k = g >=> (h >=> k) 

+  </haskell> 

+  
+  {{noteAs fans of category theory will note, these laws say precisely that functions of type <code>a > m b</code> are the arrows of a category with <code>(>{{=}}>)</code> as composition! Indeed, this is known as the ''Kleisli category'' of the monad <code>m</code>. It will come up again when we discuss <code>Arrow</code>s.}} 

+  
+  Ah, much better! The laws simply state that <code>return</code> is the identity of <code>(>=>)</code>, and that <code>(>=>)</code> is associative {{noteref}}. Working out the equivalence between these two formulations, given the definition <code>g >=> h = \x > g x >>= h</code>, is left as an exercise. 

+  
+  There is also a formulation of the monad laws in terms of <code>fmap</code>, <code>return</code>, and <code>join</code>; for a discussion of this formulation, see the Haskell [http://en.wikibooks.org/wiki/Haskell/Category_theory wikibook page on category theory]. 

+  
+  ==<code>do</code> notation== 

+  
+  Haskell's special <code>do</code> notation supports an “imperative style” of programming by providing syntactic sugar for chains of monadic expressions. The genesis of the notation lies in realizing that something like <code>a >>= \x > b >> c >>= \y > d </code> can be more readably written by putting successive computations on separate lines: 

+  
+  <haskell> 

+  a >>= \x > 

+  b >> 

+  c >>= \y > 

+  d 

+  </haskell> 

+  
+  This emphasizes that the overall computation consists of four computations <code>a</code>, <code>b</code>, <code>c</code>, and <code>d</code>, and that <code>x</code> is bound to the result of <code>a</code>, and <code>y</code> is bound to the result of <code>c</code> (<code>b</code>, <code>c</code>, and <code>d</code> are allowed to refer to <code>x</code>, and <code>d</code> is allowed to refer to <code>y</code> as well). From here it is not hard to imagine a nicer notation: 

+  
+  <haskell> 

+  do { x < a ; 

+  b ; 

+  y < c ; 

+  d 

+  } 

+  </haskell> 

+  
+  (The curly braces and semicolons may optionally be omitted; the Haskell parser uses layout to determine where they should be inserted.) This discussion should make clear that <code>do</code> notation is just syntactic sugar. In fact, <code>do</code> blocks are recursively translated into monad operations (almost) like this: 

+  
+  <pre> 

+  do e ⇨ e 

+  do { e; stmts } ⇨ e >> do { stmts } 

+  do { v < e; stmts } ⇨ e >>= \v > do { stmts } 

+  do { let decls; stmts} ⇨ let decls in do { stmts } 

+  </pre> 

+  
+  This is not quite the whole story, since <code>v</code> might be a pattern instead of a variable. For example, one can write 

+  
+  <haskell> 

+  do (x:xs) < foo 

+  bar x 

+  </haskell> 

+  
+  but what happens if <code>foo</code> produces an empty list? Well, remember that ugly <code>fail</code> function in the <code>Monad</code> type class declaration? That's what happens. See [http://haskell.org/onlinereport/exps.html#sect3.14 section 3.14 of the Haskell Report] for the full details. See also the discussion of <code>MonadPlus</code> and <code>MonadZero</code> in the [[#Other monoidal classes: Alternative, MonadPlus, ArrowPlussection on other monoidal classes]]. 

+  
+  A final note on intuition: <code>do</code> notation plays very strongly to the “computational context” point of view rather than the “container” point of view, since the binding notation <code>x < m</code> is suggestive of “extracting” a single <code>x</code> from <code>m</code> and doing something with it. But <code>m</code> may represent some sort of a container, such as a list or a tree; the meaning of <code>x < m</code> is entirely dependent on the implementation of <code>(>>=)</code>. For example, if <code>m</code> is a list, <code>x < m</code> actually means that <code>x</code> will take on each value from the list in turn. 

+  
+  ==Monad transformers== 

+  
+  One would often like to be able to combine two monads into one: for example, to have stateful, nondeterministic computations (<code>State</code> + <code>[]</code>), or computations which may fail and can consult a readonly environment (<code>Maybe</code> + <code>Reader</code>), and so on. Unfortunately, monads do not compose as nicely as applicative functors (yet another reason to use <code>Applicative</code> if you don't need the full power that <code>Monad</code> provides), but some monads can be combined in certain ways. 

+  
+  The monad transformer library [http://hackage.haskell.org/package/mtl mtl] provides a number of ''monad transformers'', such as <code>StateT</code>, <code>ReaderT</code>, <code>ErrorT</code> ([http://hackage.haskell.org/packages/archive/mtl/1.1.0.2/doc/html/ControlMonadError.html haddock]), and (soon) <code>MaybeT</code>, which can be applied to other monads to produce a new monad with the effects of both. For example, <code>StateT s Maybe</code> is an instance of <code>Monad</code>; computations of type <code>StateT s Maybe a</code> may fail, and have access to a mutable state of type <code>s</code>. These transformers can be multiply stacked. One thing to keep in mind while using monad transformers is that the order of composition matters. For example, when a <code>StateT s Maybe a</code> computation fails, the state ceases being updated; on the other hand, the state of a <code>MaybeT (State s) a</code> computation may continue to be modified even after the computation has failed. (This may seem backwards, but it is correct. Monad transformers build composite monads “inside out”; for example, <code>MaybeT (State s) a</code> is isomorphic to <code>s > Maybe (a, s)</code>. Lambdabot has an indispensable <code>@unmtl</code> command which you can use to “unpack” a monad transformer stack in this way.) 

+  
+  All monad transformers should implement the <code>MonadTrans</code> type class, defined in <code>Control.Monad.Trans</code>: 

+  
+  <haskell> 

+  class MonadTrans t where 

+  lift :: Monad m => m a > t m a 

+  </haskell> 

+  
+  It allows arbitrary computations in the base monad <code>m</code> to be “lifted” into computations in the transformed monad <code>t m</code>. (Note that type application associates to the left, just like function application, so <code>t m a = (t m) a</code>. As an exercise, you may wish to work out <code>t</code>'s kind, which is rather more interesting than most of the kinds we've seen up to this point.) However, you should only have to think about <code>MonadTrans</code> when defining your own monad transformers, not when using predefined ones. 

+  
+  {{noteThe only problem with this scheme is the quadratic number of instances required as the number of standard monad transformers grows—but as the current set of standard monad transformers seems adequate for most common use cases, this may not be that big of a deal.}} 

+  
+  There are also type classes such as <code>MonadState</code>, which provides statespecific methods like <code>get</code> and <code>put</code>, allowing you to conveniently use these methods not only with <code>State</code>, but with any monad which is an instance of <code>MonadState</code>—including <code>MaybeT (State s)</code>, <code>StateT s (ReaderT r IO)</code>, and so on. Similar type classes exist for <code>Reader</code>, <code>Writer</code>, <code>Cont</code>, <code>IO</code>, and others {{noteref}}. 

+  
+  There are two excellent references on monad transformers. Martin Grabmüller's [http://user.cs.tuberlin.de/~magr/pub/Transformers.en.html Monad Transformers Step by Step] is a thorough description, with running examples, of how to use monad transformers to elegantly build up computations with various effects. [http://cale.yi.org/index.php/How_To_Use_Monad_Transformers Cale Gibbard's article] on how to use monad transformers is more practical, describing how to structure code using monad transformers to make writing it as painless as possible. Another good starting place for learning about monad transformers is a [http://blog.sigfpe.com/2006/05/grokhaskellmonadtransformers.html blog post by Dan Piponi]. 

+  
+  ==MonadFix== 

+  
+  The <code>MonadFix</code> class describes monads which support the special fixpoint operation <code>mfix :: (a > m a) > m a</code>, which allows the output of monadic computations to be defined via recursion. This is supported in GHC and Hugs by a special “recursive do” notation, <code>mdo</code>. For more information, see Levent Erkök's thesis, [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.15.1543&rep=rep1&type=pdf Value Recursion in Monadic Computations]. 

+  
+  ==Further reading== 

+  
+  Philip Wadler was the first to propose using monads to structure functional programs. [http://homepages.inf.ed.ac.uk/wadler/topics/monads.html His paper] is still a readable introduction to the subject. 

+  
+  Much of the monad transformer library [http://hackage.haskell.org/package/mtl mtl], including the <code>Reader</code>, <code>Writer</code>, <code>State</code>, and other monads, as well as the monad transformer framework itself, was inspired by Mark Jones's classic paper [http://web.cecs.pdx.edu/~mpj/pubs/springschool.html Functional Programming with Overloading and HigherOrder Polymorphism]. It's still very much worth a read—and highly readable—after almost fifteen years. 

+  
+  {{note 

+  [http://www.haskell.org/all_about_monads/html/ All About Monads], 

+  [http://haskell.org/haskellwiki/Monads_as_Containers Monads as containers], 

+  [http://en.wikibooks.org/w/index.php?title=Haskell/Understanding_monads&oldid=933545 Understanding monads], 

+  [[[The Monadic Way]], 

+  [http://blog.sigfpe.com/2006/08/youcouldhaveinventedmonadsand.html You Could Have Invented Monads! (And Maybe You Already Have.)], 

+  [http://www.haskell.org/pipermail/haskellcafe/2006November/019190.html there's a monster in my Haskell!], 

+  [http://kawagner.blogspot.com/2007/02/understandingmonadsforreal.html Understanding Monads. For real.], 

+  [http://www.randomhacks.net/articles/2007/03/12/monadsin15minutes Monads in 15 minutes: Backtracking and Maybe], 

+  [http://haskell.org/haskellwiki/Monads_as_computation Monads as computation], 

+  [http://metafoo.co.uk/practicalmonads.txt Practical Monads]}} 

+  
+  There are, of course, numerous monad tutorials of varying quality {{noteref}}. 

+  
+  A few of the best include Cale Gibbard's [http://haskell.org/haskellwiki/Monads_as_Containers Monads as containers] and [http://haskell.org/haskellwiki/Monads_as_computation Monads as computation]; Jeff Newbern's [http://www.haskell.org/all_about_monads/html/ All About Monads], a comprehensive guide with lots of examples; and Dan Piponi's [http://blog.sigfpe.com/2006/08/youcouldhaveinventedmonadsand.html You Could Have Invented Monads!], which features great exercises. If you just want to know how to use <code>IO</code>, you could consult the [[Introduction to IO]]. Even this is just a sampling; the [[monad tutorials timeline]] is a more complete list. (All these monad tutorials have prompted parodies like [http://koweycode.blogspot.com/2007/01/thinkofmonad.html think of a monad ...] as well as other kinds of backlash like [http://ahamsandwich.wordpress.com/2007/07/26/monadsandwhymonadtutorialsareallawful/ Monads! (and Why Monad Tutorials Are All Awful)] or [http://byorgey.wordpress.com/2009/01/12/abstractionintuitionandthemonadtutorialfallacy/ Abstraction, intuition, and the “monad tutorial fallacy”].) 

+  
+  Other good monad references which are not necessarily tutorials include [http://members.chello.nl/hjgtuyl/tourdemonad.html HenkJan van Tuyl's tour] of the functions in <code>Control.Monad</code>, Dan Piponi's [http://blog.sigfpe.com/2006/10/monadsfieldguide.html field guide], and Tim Newsham's [http://www.thenewsh.com/~newsham/haskell/monad.html What's a Monad?]. There are also many blog articles which have been written on various aspects of monads; a collection of links can be found under [[Blog articles/Monads]]. 

+  
+  One of the quirks of the <code>Monad</code> class and the Haskell type system is that it is not possible to straightforwardly declare <code>Monad</code> instances for types which require a class constraint on their data, even if they are monads from a mathematical point of view. For example, <code>Data.Set</code> requires an <code>Ord</code> constraint on its data, so it cannot be easily made an instance of <code>Monad</code>. A solution to this problem was [http://www.randomhacks.net/articles/2007/03/15/datasetmonadhaskellmacros first described by Eric Kidd], and later made into a [http://hackage.haskell.org/cgibin/hackagescripts/package/rmonad library named rmonad] by Ganesh Sittampalam and Peter Gavin. 

+  
+  There are many good reasons for eschewing <code>do</code> notation; some have gone so far as to [[Do_notation_considered_harmfulconsider it harmful]. 

+  
+  Monads can be generalized in various ways; for an exposition of one possibility, see Robert Atkey's paper on [http://homepages.inf.ed.ac.uk/ratkey/paramnotionsjfp.pdf parameterized monads], or Dan Piponi's [http://blog.sigfpe.com/2009/02/beyondmonads.html Beyond Monads]. 

+  
+  For the categorically inclined, monads can be viewed as monoids ([http://blog.sigfpe.com/2008/11/frommonoidstomonads.html From Monoids to Monads]) and also as closure operators [http://blog.plover.com/math/monadclosure.html Triples and Closure]. Derek Elkins's article in [http://www.haskell.org/wikiupload/8/85/TMRIssue13.pdf issue 13 of the Monad.Reader] contains an exposition of the categorytheoretic underpinnings of some of the standard <code>Monad</code> instances, such as <code>State</code> and <code>Cont</code>. There is also an alternative way to compose monads, using coproducts, as described by [http://doi.acm.org/10.1145/583852.581492 Lüth and Ghani], although this method has not (yet?) seen widespread use. 

+  
+  Links to many more research papers related to monads can be found under [[Research papers/Monads and arrows]]. 

+  
+  =Monoid= 

+  
+  A monoid is a set <math>S</math> together with a binary operation <math>\oplus</math> which 

+  combines elements from <math>S</math>. The <math>\oplus</math> operator is required to be associative 

+  (that is, <math>(a \oplus b) \oplus c = a \oplus (b \oplus c)</math>, for any 

+  <math>a,b,c</math> which are elements of <math>S</math>), and there must be some element of 

+  <math>S</math> which is the identity with respect to <math>\oplus</math>. 

+  (If you are familiar with group theory, a monoid is like a 

+  group without the requirement that inverses exist.) For example, the 

+  natural numbers under addition form a monoid: the sum of any two 

+  natural numbers is a natural number; <math>(a+b)+c = a+(b+c)</math> for any 

+  natural numbers <math>a</math>, <math>b</math>, and <math>c\,</math>; and zero is the additive 

+  identity. The integers under multiplication also form a monoid, as do 

+  natural numbers under <math>\max</math>, Boolean values under conjunction and 

+  disjunction, lists under concatenation, functions from a set to itself 

+  under composition ... Monoids show up all over the place, once you 

+  know to look for them. 

+  
+  ==Definition== 

+  
+  The definition of the <code>Monoid</code> type class (defined in 

+  <code>Data.Monoid</code>; [http://haskell.org/ghc/docs/latest/html/libraries/base/DataMonoid.html haddock]) is: 

+  
+  <haskell> 

+  class Monoid a where 

+  mempty :: a 

+  mappend :: a > a > a 

+  
+  mconcat :: [a] > a 

+  mconcat = foldr mappend mempty 

+  </haskell> 

+  
+  The <code>mempty</code> value specifies the identity element of the monoid, and <code>mappend</code> 

+  is the binary operation. The default definition for <code>mconcat</code> 

+  “reduces” a list of elements by combining them all with <code>mappend</code>, 

+  using a right fold. It is only in the <code>Monoid</code> class so that specific 

+  instances have the option of providing an alternative, more efficient 

+  implementation; usually, you can safely ignore <code>mconcat</code> when creating 

+  a <code>Monoid</code> instance, since its default definition will work just fine. 

+  
+  The <code>Monoid</code> methods are rather unfortunately named; they are inspired 

+  by the list instance of <code>Monoid</code>, where indeed <code>mempty = []</code> and <code>mappend = (++)</code>, but this is misleading since many 

+  monoids have little to do with appending (see these [http://thread.gmane.org/gmane.comp.lang.haskell.cafe/50590 Comments from OCaml Hacker Brian Hurt] on the haskellcafe mailing list). 

+  
+  ==Laws== 

+  
+  Of course, every <code>Monoid</code> instance should actually be a monoid in the 

+  mathematical sense, which implies these laws: 

+  
+  <haskell> 

+  mempty `mappend` x = x 

+  x `mappend` mempty = x 

+  (x `mappend` y) `mappend` z = x `mappend` (y `mappend` z) 

+  </haskell> 

+  
+  ==Instances== 

+  
+  There are quite a few interesting <code>Monoid</code> instances defined in 

+  <code>Data.Monoid</code>. 

+  
+  * <code>[a]</code> is a <code>Monoid</code>, with <code>mempty = []</code> and <code>mappend = (++)</code>. 

+  It is not hard to check that 

+  <code>(x ++ y) ++ z = x ++ (y ++ z)</code> for any lists <code>x</code>, <code>y</code>, and <code>z</code>, and 

+  that the empty list is the identity: 

+  <code>[] ++ x = x ++ [] = x</code>. 

+  
+  * As noted previously, we can make a monoid out of any numeric 

+  type under either addition or multiplication. However, since we 

+  can't have two instances for the same type, <code>Data.Monoid</code> provides 

+  two <code>newtype</code> wrappers, <code>Sum</code> and <code>Product</code>, with appropriate 

+  <code>Monoid</code> instances. 

+  
+  <haskell> 

+  > getSum (mconcat . map Sum $ [1..5]) 

+  15 

+  > getProduct (mconcat . map Product $ [1..5]) 

+  120 

+  </haskell> 

+  
+  :This example code is silly, of course; we could just write 

+  <code>sum [1..5]</code> and <code>product [1..5]</code>. Nevertheless, these instances 

+  are useful in more generalized settings, as we will see in the 

+  [[Foldablesection <code>Foldable</code>]]. 

+  
+  * <code>Any</code> and <code>All</code> are <code>newtype</code> wrappers providing <code>Monoid</code> 

+  instances for <code>Bool</code> (under disjunction and conjunction, 

+  respectively). 

+  
+  * There are three instances for <code>Maybe</code>: a basic instance which 

+  lifts a <code>Monoid</code> instance for <code>a</code> to an instance for <code>Maybe a</code>, and 

+  two <code>newtype</code> wrappers <code>First</code> and <code>Last</code> for which <code>mappend</code> 

+  selects the first (respectively last) non<code>Nothing</code> item. 

+  
+  * <code>Endo a</code> is a newtype wrapper for functions <code>a > a</code>, which form 

+  a monoid under composition. 

+  
+  * There are several ways to “lift” <code>Monoid</code> instances to 

+  instances with additional structure. We have already seen that an 

+  instance for <code>a</code> can be lifted to an instance for <code>Maybe a</code>. There 

+  are also tuple instances: if <code>a</code> and <code>b</code> are instances of <code>Monoid</code>, 

+  then so is <code>(a,b)</code>, using the monoid operations for <code>a</code> and <code>b</code> in 

+  the obvious pairwise manner. Finally, if <code>a</code> is a <code>Monoid</code>, then so 

+  is the function type <code>e > a</code> for any <code>e</code>; in particular, 

+  <code>g `mappend` h</code> is the function which applies both <code>g</code> and <code>h</code> to 

+  its argument and then combines the result using the underlying 

+  <code>Monoid</code> instance for <code>a</code>. This can be quite useful and 

+  elegant (see [http://thread.gmane.org/gmane.comp.lang.haskell.cafe/52416 example]). 

+  
+  * The type <code>Ordering = LT  EQ  GT</code> is a <code>Monoid</code>, defined in 

+  such a way that 

+  <code>mconcat (zipWith compare xs ys)</code> computes the 

+  lexicographic ordering of <code>xs</code> and <code>ys</code>. In particular, 

+  <code>mempty = EQ</code>, and <code>mappend</code> evaluates to its leftmost non<code>EQ</code> 

+  argument (or <code>EQ</code> if both arguments are <code>EQ</code>). This can be used 

+  together with the function instance of <code>Monoid</code> to do some clever 

+  things 

+  ([http://www.reddit.com/r/programming/comments/7cf4r/monoids_in_my_programming_language/c06adnx example]). 

+  
+  * There are also <code>Monoid</code> instances for several standard data 

+  structures in the containers library ([http://hackage.haskell.org/packages/archive/containers/0.2.0.0/doc/html/index.html haddock]), 

+  including <code>Map</code>, <code>Set</code>, and <code>Sequence</code>. 

+  
+  <code>Monoid</code> is also used to enable several other type class instances. 

+  As noted previously, we can use <code>Monoid</code> to make <code>((,) e)</code> an instance 

+  of <code>Applicative</code>: 

+  
+  <haskell> 

+  instance Monoid e => Applicative ((,) e) where 

+  pure x = (mempty, x) 

+  (u, f) <*> (v, x) = (u `mappend` v, f x) 

+  </haskell> 

+  
+  <code>Monoid</code> can be similarly used to make <code>((,) e)</code> an instance of 

+  <code>Monad</code> as well; this is known as the ''writer monad''. As we've 

+  already seen, <code>Writer</code> and <code>WriterT</code> are a newtype wrapper and 

+  transformer for this monad, respectively. 

+  
+  <code>Monoid</code> also plays a key role in the <code>Foldable</code> type class 

+  (see section [[#FoldableFoldable]]). 

+  
+  ==Other monoidal classes: Alternative, MonadPlus, ArrowPlus== 

+  
+  The <code>Alternative</code> type class ([http://haskell.org/ghc/docs/latest/html/libraries/base/ControlApplicative.html#g:2 haddock]) 

+  is for <code>Applicative</code> functors which also have 

+  a monoid structure: 

+  
+  <haskell> 

+  class Applicative f => Alternative f where 

+  empty :: f a 

+  (<>) :: f a > f a > f a 

+  </haskell> 

+  
+  Of course, instances of <code>Alternative</code> should satisfy the monoid laws. 

+  
+  Likewise, <code>MonadPlus</code> ([http://haskell.org/ghc/docs/latest/html/libraries/base/ControlMonad.html#t:MonadPlus haddock]) 

+  is for <code>Monad</code>s with a monoid structure: 

+  
+  <haskell> 

+  class Monad m => MonadPlus m where 

+  mzero :: m a 

+  mplus :: m a > m a > m a 

+  </haskell> 

+  
+  The <code>MonadPlus</code> documentation states that it is intended to model 

+  monads which also support “choice and failure”; in addition to the 

+  monoid laws, instances of <code>MonadPlus</code> are expected to satisfy 

+  
+  <haskell> 

+  mzero >>= f = mzero 

+  v >> mzero = mzero 

+  </haskell> 

+  
+  which explains the sense in which <code>mzero</code> denotes failure. Since 

+  <code>mzero</code> should be the identity for <code>mplus</code>, the computation <code>m1 `mplus` m2</code> succeeds (evaluates to something other than <code>mzero</code>) if 

+  either <code>m1</code> or <code>m2</code> does; so <code>mplus</code> represents choice. The <code>guard</code> 

+  function can also be used with instances of <code>MonadPlus</code>; it requires a 

+  condition to be satisfied and fails (using <code>mzero</code>) if it is not. A 

+  simple example of a <code>MonadPlus</code> instance is <code>[]</code>, which is exactly the 

+  same as the <code>Monoid</code> instance for <code>[]</code>: the empty list represents 

+  failure, and list concatenation represents choice. In general, 

+  however, a <code>MonadPlus</code> instance for a type need not be the same as its 

+  <code>Monoid</code> instance; <code>Maybe</code> is an example of such a type. A great 

+  introduction to the <code>MonadPlus</code> type class, with interesting examples 

+  of its use, is Doug Auclair's ''MonadPlus: What a Super Monad!'' in [http://www.haskell.org/sitewiki/images/6/6a/TMRIssue11.pdf the Monad.Reader issue 11]. 

+  
+  There used to be a type class called <code>MonadZero</code> containing only 

+  <code>mzero</code>, representing monads with failure. The <code>do</code>notation requires 

+  some notion of failure to deal with failing pattern matches. 

+  Unfortunately, <code>MonadZero</code> was scrapped in favor of adding the <code>fail</code> 

+  method to the <code>Monad</code> class. If we are lucky, someday <code>MonadZero</code> will 

+  be restored, and <code>fail</code> will be banished to the bit bucket where it 

+  belongs (see [[MonadPlus reform proposal]]). The idea is that any 

+  <code>do</code>block which uses pattern matching (and hence may fail) would require 

+  a <code>MonadZero</code> constraint; otherwise, only a <code>Monad</code> constraint would be 

+  required. 

+  
+  Finally, <code>ArrowZero</code> and <code>ArrowPlus</code> ([http://haskell.org/ghc/docs/latest/html/libraries/base/ControlArrow.html#t:ArrowZero haddock]) 

+  represent <code>Arrow</code>s ([[#Arrowsee below]]) with a 

+  monoid structure: 

+  
+  <haskell> 

+  class Arrow (~>) => ArrowZero (~>) where 

+  zeroArrow :: b ~> c 

+  
+  class ArrowZero (~>) => ArrowPlus (~>) where 

+  (<+>) :: (b ~> c) > (b ~> c) > (b ~> c) 

+  </haskell> 

+  
+  ==Further reading== 

+  
+  Monoids have gotten a fair bit of attention recently, ultimately due 

+  to 

+  [http://enfranchisedmind.com/blog/2009/01/15/randomthoughtsonhaskell/ a blog post by Brian Hurt], in which he 

+  complained about the fact that the names of many Haskell type classes 

+  (<code>Monoid</code> in particular) are taken from abstract mathematics. This 

+  resulted in [http://thread.gmane.org/gmane.comp.lang.haskell.cafe/50590 a long haskellcafe thread] 

+  arguing the point and discussing monoids in general. 

+  
+  {{noteMay its name live forever.}} 

+  
+  However, this was quickly followed by several blog posts about 

+  <code>Monoid</code> {{noteref}}. First, Dan Piponi 

+  wrote a great introductory post, [http://blog.sigfpe.com/2009/01/haskellmonoidsandtheiruses.html Haskell Monoids and their 

+  Uses]. This was quickly followed by 

+  Heinrich Apfelmus's [http://apfelmus.nfshost.com/monoidfingertree.html Monoids and Finger Trees], an accessible exposition of 

+  Hinze and Paterson's [http://www.soi.city.ac.uk/%7Eross/papers/FingerTree.html classic paper on 23 finger trees], which makes very clever 

+  use of <code>Monoid</code> to implement an elegant and generic data structure. 

+  Dan Piponi then wrote two fascinating articles about using <code>Monoids</code> 

+  (and finger trees): [http://blog.sigfpe.com/2009/01/fastincrementalregularexpression.html Fast Incremental Regular Expressions] and [http://blog.sigfpe.com/2009/01/beyondregularexpressionsmore.html Beyond Regular Expressions] 

+  
+  In a similar vein, David Place's article on improving <code>Data.Map</code> in 

+  order to compute incremental folds (see [http://www.haskell.org/sitewiki/images/6/6a/TMRIssue11.pdf the Monad Reader issue 11]) 

+  is also a 

+  good example of using <code>Monoid</code> to generalize a data structure. 

+  
+  Some other interesting examples of <code>Monoid</code> use include [http://www.reddit.com/r/programming/comments/7cf4r/monoids_in_my_programming_language/c06adnx building 

+  elegant list sorting combinators], 

+  [http://byorgey.wordpress.com/2008/04/17/collectingunstructuredinformationwiththemonoidofpartialknowledge/ collecting unstructured information], 

+  and a brilliant series of posts by ChungChieh Shan and Dylan Thurston 

+  using <code>Monoid</code>s to [http://conway.rutgers.edu/~ccshan/wiki/blog/posts/WordNumbers1/ elegantly solve a difficult combinatorial 

+  puzzle] (followed by 

+  [http://conway.rutgers.edu/~ccshan/wiki/blog/posts/WordNumbers2/ part 2], 

+  [http://conway.rutgers.edu/~ccshan/wiki/blog/posts/WordNumbers3/ part 3], 

+  [http://conway.rutgers.edu/~ccshan/wiki/blog/posts/WordNumbers4/ part 4]). 

+  
+  As unlikely as it sounds, monads can actually be viewed as a sort of 

+  monoid, with <code>join</code> playing the role of the binary operation and 

+  <code>return</code> the role of the identity; see [http://blog.sigfpe.com/2008/11/frommonoidstomonads.html Dan Piponi's blog post]. 

+  
+  =Foldable= 

+  
+  The <code>Foldable</code> class, defined in the <code>Data.Foldable</code> 

+  module ([http://haskell.org/ghc/docs/latest/html/libraries/base/DataFoldable.html haddock]), abstracts over containers which can be 

+  “folded” into a summary value. This allows such folding operations 

+  to be written in a containeragnostic way. 

+  
+  ==Definition== 

+  
+  The definition of the <code>Foldable</code> type class is: 

+  
+  <haskell> 

+  class Foldable t where 

+  fold :: Monoid m => t m > m 

+  foldMap :: Monoid m => (a > m) > t a > m 

+  
+  foldr :: (a > b > b) > b > t a > b 

+  foldl :: (a > b > a) > a > t b > a 

+  foldr1 :: (a > a > a) > t a > a 

+  foldl1 :: (a > a > a) > t a > a 

+  </haskell> 

+  
+  This may look complicated, but in fact, to make a <code>Foldable</code> instance 

+  you only need to implement one method: your choice of <code>foldMap</code> or 

+  <code>foldr</code>. All the other methods have default implementations in terms 

+  of these, and are presumably included in the class in case more 

+  efficient implementations can be provided. 

+  
+  ==Instances and examples== 

+  
+  The type of <code>foldMap</code> should make it clear what it is supposed to do: 

+  given a way to convert the data in a container into a <code>Monoid</code> (a 

+  function <code>a > m</code>) and a container of <code>a</code>'s (<code>t a</code>), <code>foldMap</code> 

+  provides a way to iterate over the entire contents of the container, 

+  converting all the <code>a</code>'s to <code>m</code>'s and combining all the <code>m</code>'s with 

+  <code>mappend</code>. The following code shows two examples: a simple 

+  implementation of <code>foldMap</code> for lists, and a binary tree example 

+  provided by the <code>Foldable</code> documentation. 

+  
+  <haskell> 

+  instance Foldable [] where 

+  foldMap g = mconcat . map g 

+  
+  data Tree a = Empty  Leaf a  Node (Tree a) a (Tree a) 

+  
+  instance Foldable Tree where 

+  foldMap f Empty = mempty 

+  foldMap f (Leaf x) = f x 

+  foldMap f (Node l k r) = foldMap f l ++ f k ++ foldMap f r 

+  where (++) = mappend 

+  </haskell> 

+  
+  The <code>foldr</code> function has a type similar to the <code>foldr</code> found in the <code>Prelude</code>, but 

+  more general, since the <code>foldr</code> in the <code>Prelude</code> works only on lists. 

+  
+  The <code>Foldable</code> module also provides instances for <code>Maybe</code> and <code>Array</code>; 

+  additionally, many of the data structures found in the standard [http://hackage.haskell.org/package/containers containers library] (for example, <code>Map</code>, <code>Set</code>, <code>Tree</code>, 

+  and <code>Sequence</code>) provide their own <code>Foldable</code> instances. 

+  
+  ==Derived folds== 

+  
+  Given an instance of <code>Foldable</code>, we can write generic, 

+  containeragnostic functions such as: 

+  
+  <haskell> 

+   Compute the size of any container. 

+  containerSize :: Foldable f => f a > Int 

+  containerSize = getSum . foldMap (const (Sum 1)) 

+  
+   Compute a list of elements of a container satisfying a predicate. 

+  filterF :: Foldable f => (a > Bool) > f a > [a] 

+  filterF p = foldMap (\a > if p a then [a] else []) 

+  
+   Get a list of all the Strings in a container which include the 

+   letter a. 

+  aStrings :: Foldable f => f String > [String] 

+  aStrings = filterF (elem 'a') 

+  </haskell> 

+  
+  The <code>Foldable</code> module also provides a large number of predefined 

+  folds, many of which are generalized versions of <code>Prelude</code> functions of the 

+  same name that only work on lists: <code>concat</code>, <code>concatMap</code>, <code>and</code>, 

+  <code>or</code>, <code>any</code>, <code>all</code>, <code>sum</code>, <code>product</code>, <code>maximum</code>(<code>By</code>), 

+  <code>minimum</code>(<code>By</code>), <code>elem</code>, <code>notElem</code>, and <code>find</code>. The reader may enjoy 

+  coming up with elegant implementations of these functions using <code>fold</code> 

+  or <code>foldMap</code> and appropriate <code>Monoid</code> instances. 

+  
+  There are also generic functions that work with <code>Applicative</code> or 

+  <code>Monad</code> instances to generate some sort of computation from each 

+  element in a container, and then perform all the side effects from 

+  those computations, discarding the results: <code>traverse_</code>, <code>sequenceA_</code>, 

+  and others. The results must be discarded because the <code>Foldable</code> 

+  class is too weak to specify what to do with them: we cannot, in 

+  general, make an arbitrary <code>Applicative</code> or <code>Monad</code> instance into a 

+  <code>Monoid</code>. If we do have an <code>Applicative</code> or <code>Monad</code> with a monoid 

+  structure—that is, an <code>Alternative</code> or a <code>MonadPlus</code>—then we can 

+  use the <code>asum</code> or <code>msum</code> functions, which can combine the results as 

+  well. Consult the [http://haskell.org/ghc/docs/latest/html/libraries/base/DataFoldable.html <code>Foldable</code> documentation] for 

+  more details on any of these functions. 

+  
+  Note that the <code>Foldable</code> operations always forget the structure of 

+  the container being folded. If we start with a container of type <code>t a</code> for some <code>Foldable t</code>, then <code>t</code> will never appear in the output 

+  type of any operations defined in the <code>Foldable</code> module. Many times 

+  this is exactly what we want, but sometimes we would like to be able 

+  to generically traverse a container while preserving its 

+  structure—and this is exactly what the <code>Traversable</code> class provides, 

+  which will be discussed in the next section. 

+  
+  ==Further reading== 

+  
+  The <code>Foldable</code> class had its genesis in [http://www.soi.city.ac.uk/~ross/papers/Applicative.html McBride and Paterson's paper] 

+  introducing <code>Applicative</code>, although it has 

+  been fleshed out quite a bit from the form in the paper. 

+  
+  An interesting use of <code>Foldable</code> (as well as <code>Traversable</code>) can be 

+  found in Janis Voigtländer's paper [http://doi.acm.org/10.1145/1480881.1480904 Bidirectionalization for free!]. 

+  
+  =Traversable= 

+  
+  ==Definition== 

+  
+  The <code>Traversable</code> type class, defined in the <code>Data.Traversable</code> 

+  module ([http://haskell.org/ghc/docs/latest/html/libraries/base/DataTraversable.html haddock]), is: 

+  
+  <haskell> 

+  class (Functor t, Foldable t) => Traversable t where 

+  traverse :: Applicative f => (a > f b) > t a > f (t b) 

+  sequenceA :: Applicative f => t (f a) > f (t a) 

+  mapM :: Monad m => (a > m b) > t a > m (t b) 

+  sequence :: Monad m => t (m a) > m (t a) 

+  </haskell> 

+  
+  As you can see, every <code>Traversable</code> is also a foldable functor. Like 

+  <code>Foldable</code>, there is a lot in this type class, but making instances is 

+  actually rather easy: one need only implement <code>traverse</code> or 

+  <code>sequenceA</code>; the other methods all have default implementations in 

+  terms of these functions. A good exercise is to figure out what the default 

+  implementations should be: given either <code>traverse</code> or <code>sequenceA</code>, how 

+  would you define the other three methods? (Hint for <code>mapM</code>: 

+  <code>Control.Applicative</code> exports the <code>WrapMonad</code> newtype, which makes any 

+  <code>Monad</code> into an <code>Applicative</code>. The <code>sequence</code> function can be implemented in terms 

+  of <code>mapM</code>.) 

+  
+  ==Intuition== 

+  
+  The key method of the <code>Traversable</code> class, and the source of its 

+  unique power, is <code>sequenceA</code>. Consider its type: 

+  <haskell> 

+  sequenceA :: Applicative f => t (f a) > f (t a) 

+  </haskell> 

+  This answers the fundamental question: when can we commute two 

+  functors? For example, can we turn a tree of lists into a list of 

+  trees? (Answer: yes, in two ways. Figuring out what they are, and 

+  why, is left as an exercise. A much more challenging question is 

+  whether a list of trees can be turned into a tree of lists.) 

+  
+  The ability to compose two monads depends crucially on this ability to 

+  commute functors. Intuitively, if we want to build a composed monad 

+  <code>M a = m (n a)</code> out of monads <code>m</code> and <code>n</code>, then to be able to 

+  implement <code>join :: M (M a) > M a</code>, that is, 

+  <code>join :: m (n (m (n a))) > m (n a)</code>, we have to be able to commute 

+  the <code>n</code> past the <code>m</code> to get <code>m (m (n (n a)))</code>, and then we can use the 

+  <code>join</code>s for <code>m</code> and <code>n</code> to produce something of type <code>m (n a)</code>. See 

+  [http://web.cecs.pdx.edu/~mpj/pubs/springschool.html Mark Jones's paper] for more details. 

+  
+  ==Instances and examples== 

+  
+  What's an example of a <code>Traversable</code> instance? 

+  The following code shows an example instance for the same 

+  <code>Tree</code> type used as an example in the previous <code>Foldable</code> section. It 

+  is instructive to compare this instance with a <code>Functor</code> instance for 

+  <code>Tree</code>, which is also shown. 

+  
+  <haskell> 

+  data Tree a = Empty  Leaf a  Node (Tree a) a (Tree a) 

+  
+  instance Traversable Tree where 

+  traverse g Empty = pure Empty 

+  traverse g (Leaf x) = Leaf <$> g x 

+  traverse g (Node l x r) = Node <$> traverse g l 

+  <*> g x 

+  <*> traverse g r 

+  
+  instance Functor Tree where 

+  fmap g Empty = Empty 

+  fmap g (Leaf x) = Leaf $ g x 

+  fmap g (Node l x r) = Node (fmap g l) 

+  (g x) 

+  (fmap g r) 

+  </haskell> 

+  
+  It should be clear that the <code>Traversable</code> and <code>Functor</code> instances for 

+  <code>Tree</code> are almost identical; the only difference is that the <code>Functor</code> 

+  instance involves normal function application, whereas the 

+  applications in the <code>Traversable</code> instance take place within an 

+  <code>Applicative</code> context, using <code>(<$>)</code> and <code>(<*>)</code>. In fact, this will 

+  be 

+  true for any type. 

+  
+  Any <code>Traversable</code> functor is also <code>Foldable</code>, and a <code>Functor</code>. We can see 

+  this not only from the class declaration, but by the fact that we can 

+  implement the methods of both classes given only the <code>Traversable</code> 

+  methods. A good exercise is to implement <code>fmap</code> and <code>foldMap</code> using 

+  only the <code>Traversable</code> methods; the implementations are surprisingly 

+  elegant. The <code>Traversable</code> module provides these 

+  implementations as <code>fmapDefault</code> and <code>foldMapDefault</code>. 

+  
+  The standard libraries provide a number of <code>Traversable</code> instances, 

+  including instances for <code>[]</code>, <code>Maybe</code>, <code>Map</code>, <code>Tree</code>, and <code>Sequence</code>. 

+  Notably, <code>Set</code> is not <code>Traversable</code>, although it is <code>Foldable</code>. 

+  
+  ==Further reading== 

+  
+  The <code>Traversable</code> class also had its genesis in [http://www.soi.city.ac.uk/~ross/papers/Applicative.html McBride and Paterson's 

+  <code>Applicative</code> paper], and is described in 

+  more detail in Gibbons and Oliveira, [http://www.comlab.ox.ac.uk/jeremy.gibbons/publications/iterator.pdf The Essence of the Iterator Pattern], which also contains a wealth of 

+  references to related work. 

+  
+  =Category= 

+  
+  <code>Category</code> is another fairly new addition to the Haskell standard 

+  libraries; you may or may not have it installed depending on the 

+  version of your <code>base</code> package. It generalizes the notion of 

+  function composition to general “morphisms.” 

+  
+  The definition of the <code>Category</code> type class (from 

+  <code>Control.Category</code>—[http://haskell.org/ghc/docs/latest/html/libraries/base/ControlCategory.html haddock]) is shown below. For ease of reading, note that I have used an 

+  infix type constructor <code>(~>)</code>, much like the infix function type 

+  constructor <code>(>)</code>. This syntax is not part of Haskell 98. 

+  The second definition shown is the one used in the standard libraries. 

+  For the remainder of the article, I will use the infix type 

+  constructor <code>(~>)</code> for <code>Category</code> as well as <code>Arrow</code>. 

+  
+  <haskell> 

+  class Category (~>) where 

+  id :: a ~> a 

+  (.) :: (b ~> c) > (a ~> b) > (a ~> c) 

+  
+   The same thing, with a normal (prefix) type constructor 

+  class Category cat where 

+  id :: cat a a 

+  (.) :: cat b c > cat a b > cat a c 

+  </haskell> 

+  
+  Note that an instance of <code>Category</code> should be a type constructor which 

+  takes two type arguments, that is, something of kind <code>* > * > *</code>. It 

+  is instructive to imagine the type constructor variable <code>cat</code> replaced 

+  by the function constructor <code>(>)</code>: indeed, in this case we recover 

+  precisely the familiar identity function <code>id</code> and function composition 

+  operator <code>(.)</code> defined in the standard <code>Prelude</code>. 

+  
+  Of course, the <code>Category</code> module provides exactly such an instance of 

+  <code>Category</code> for <code>(>)</code>. But it also provides one other instance, shown 

+  below, which should be familiar from the 

+  previous discussion of the <code>Monad</code> laws. <code>Kleisli m a b</code>, as defined 

+  in the <code>Control.Arrow</code> module, is just a <code>newtype</code> wrapper around <code>a > m b</code>. 

+  
+  <haskell> 

+  newtype Kleisli m a b = Kleisli { runKleisli :: a > m b } 

+  
+  instance Monad m => Category (Kleisli m) where 

+  id = Kleisli return 

+  Kleisli g . Kleisli h = Kleisli (h >=> g) 

+  </haskell> 

+  
+  The only law that <code>Category</code> instances should satisfy is that <code>id</code> and 

+  <code>(.)</code> should form a monoid—that is, <code>id</code> should be the identity of 

+  <code>(.)</code>, and <code>(.)</code> should be associative. 

+  
+  Finally, the <code>Category</code> module exports two additional operators: 

+  <code>(<<<)</code>, which is just a synonym for <code>(.)</code>, and <code>(>>>)</code>, which is 

+  <code>(.)</code> with its arguments reversed. (In previous versions of the 

+  libraries, these operators were defined as part of the <code>Arrow</code> class.) 

+  
+  ==Further reading== 

+  
+  The name <code>Category</code> is a bit misleading, since the <code>Category</code> class 

+  cannot represent arbitrary categories, but only categories whose 

+  objects are objects of <code>Hask</code>, the category of Haskell types. For a 

+  more general treatment of categories within Haskell, see the 

+  [http://hackage.haskell.org/package/categoryextras categoryextras package]. For more about 

+  category theory in general, see the excellent [http://en.wikibooks.org/wiki/Haskell/Category_theory Haskell wikibook page], 

+  [http://books.google.com/books/about/Category_theory.html?id=MCJ6x2lC7oC Steve Awodey's new book], 

+  Benjamin Pierce's 

+  [http://books.google.com/books/about/Basic_category_theory_for_computer_scien.html?id=ezdeaHfpYPwC Basic category theory for computer scientists], or 

+  [http://folli.loria.fr/cds/1999/esslli99/courses/barrwells.html Barr and Wells's category theory lecture notes]. [http://dekudekuplex.wordpress.com/2009/01/19/motivatinglearningcategorytheoryfornonmathematicians/ Benjamin Russell's blog post] 

+  is another good source of motivation and 

+  category theory links. You certainly don't need to know any category 

+  theory to be a successful and productive Haskell programmer, but it 

+  does lend itself to much deeper appreciation of Haskell's underlying 

+  theory. 

+  
+  =Arrow= 

+  
+  The <code>Arrow</code> class represents another abstraction of computation, in a 

+  similar vein to <code>Monad</code> and <code>Applicative</code>. However, unlike <code>Monad</code> 

+  and <code>Applicative</code>, whose types only reflect their output, the type of 

+  an <code>Arrow</code> computation reflects both its input and output. Arrows 

+  generalize functions: if <code>(~>)</code> is an instance of <code>Arrow</code>, a value of 

+  type <code>b ~> c</code> can be thought of as a computation which takes values of 

+  type <code>b</code> as input, and produces values of type <code>c</code> as output. In the 

+  <code>(>)</code> instance of <code>Arrow</code> this is just a pure function; in general, however, 

+  an arrow may represent some sort of “effectful” computation. 

+  
+  ==Definition== 

+  
+  The definition of the <code>Arrow</code> type class, from 

+  <code>Control.Arrow</code> ([http://haskell.org/ghc/docs/latest/html/libraries/base/ControlArrow.html haddock]), is: 

+  
+  <haskell> 

+  class Category (~>) => Arrow (~>) where 

+  arr :: (b > c) > (b ~> c) 

+  first :: (b ~> c) > ((b, d) ~> (c, d)) 

+  second :: (b ~> c) > ((d, b) ~> (d, c)) 

+  (***) :: (b ~> c) > (b' ~> c') > ((b, b') ~> (c, c')) 

+  (&&&) :: (b ~> c) > (b ~> c') > (b ~> (c, c')) 

+  </haskell> 

+  
+  {{noteIn versions of the <code>base</code> 

+  package prior to version 4, there is no <code>Category</code> class, and the 

+  <code>Arrow</code> class includes the arrow composition operator <code>(>>>)</code>. It 

+  also includes <code>pure</code> as a synonym for <code>arr</code>, but this was removed 

+  since it conflicts with the <code>pure</code> from <code>Applicative</code>.}} 

+  
+  The first thing to note is the <code>Category</code> class constraint, which 

+  means that we get identity arrows and arrow composition for free: 

+  given two arrows <code>g :: b ~> c</code> and <code>h :: c ~> d</code>, we can form their 

+  composition <code>g >>> h :: b ~> d</code> {{noteref}}. 

+  
+  As should be a familiar pattern by now, the only methods which must be 

+  defined when writing a new instance of <code>Arrow</code> are <code>arr</code> and <code>first</code>; 

+  the other methods have default definitions in terms of these, but are 

+  included in the <code>Arrow</code> class so that they can be overridden with more 

+  efficient implementations if desired. 

+  
+  ==Intuition== 

+  
+  Let's look at each of the arrow methods in turn. [http://www.haskell.org/arrows/ Ross Paterson's web page on arrows] has nice diagrams which can help 

+  build intuition. 

+  
+  * The <code>arr</code> function takes any function <code>b > c</code> and turns it into a 

+  generalized arrow <code>b ~> c</code>. The <code>arr</code> method justifies the claim 

+  that arrows generalize functions, since it says that we can treat 

+  any function as an arrow. It is intended that the arrow <code>arr g</code> is 

+  “pure” in the sense that it only computes <code>g</code> and has no 

+  “effects” (whatever that might mean for any particular arrow type). 

+  
+  * The <code>first</code> method turns any arrow from <code>b</code> to <code>c</code> into an arrow 

+  from <code>(b,d)</code> to <code>(c,d)</code>. The idea is that <code>first g</code> uses <code>g</code> to 

+  process the first element of a tuple, and lets the second element pass 

+  through unchanged. For the function instance of <code>Arrow</code>, of course, 

+  <code>first g (x,y) = (g x, y)</code>. 

+  
+  * The <code>second</code> function is similar to <code>first</code>, but with the elements of the 

+  tuples swapped. Indeed, it can be defined in terms of <code>first</code> using 

+  an auxiliary function <code>swap</code>, defined by <code>swap (x,y) = (y,x)</code>. 

+  
+  * The <code>(***)</code> operator is “parallel composition” of arrows: it takes two 

+  arrows and makes them into one arrow on tuples, which has the 

+  behavior of the first arrow on the first element of a tuple, and the 

+  behavior of the second arrow on the second element. The mnemonic 

+  is that <code>g *** h</code> is the ''product'' (hence <code>*</code>) of <code>g</code> and 

+  <code>h</code>. For the function instance of <code>Arrow</code>, 

+  we define <code>(g *** h) (x,y) = (g x, h y)</code>. The default implementation of 

+  <code>(***)</code> is in terms of <code>first</code>, <code>second</code>, and sequential arrow 

+  composition <code>(>>>)</code>. The reader may also wish to think about how to 

+  implement <code>first</code> and <code>second</code> in terms of <code>(***)</code>. 

+  
+  * The <code>(&&&)</code> operator is “fanout composition” of arrows: it takes two arrows 

+  <code>g</code> and <code>h</code> and makes them into a new arrow <code>g &&& h</code> which supplies 

+  its input as the input to both <code>g</code> and <code>h</code>, returning their results 

+  as a tuple. The mnemonic is that <code>g &&& h</code> performs both <code>g</code> 

+  ''and'' <code>h</code> (hence <code>&</code>) on its input. For functions, we define <code>(g &&& h) x = (g x, h x)</code>. 

+  
+  ==Instances== 

+  
+  The <code>Arrow</code> library itself only provides two <code>Arrow</code> instances, both 

+  of which we have already seen: <code>(>)</code>, the normal function 

+  constructor, and <code>Kleisli m</code>, which makes functions of 

+  type <code>a > m b</code> into <code>Arrow</code>s for any <code>Monad m</code>. These instances are: 

+  
+  <haskell> 

+  instance Arrow (>) where 

+  arr g = g 

+  first g (x,y) = (g x, y) 

+  
+  newtype Kleisli m a b = Kleisli { runKleisli :: a > m b } 

+  
+  instance Monad m => Arrow (Kleisli m) where 

+  arr f = Kleisli (return . f) 

+  first (Kleisli f) = Kleisli (\ ~(b,d) > do c < f b 

+  return (c,d) ) 

+  </haskell> 

+  
+  ==Laws== 

+  
+  {{noteSee [http://dx.doi.org/10.1016/S01676423(99)000234 John Hughes: Generalising monads to arrows]; [http://homepages.inf.ed.ac.uk/wadler/papers/arrows/arrows.pdf Sam Lindley, Philip Wadler, Jeremy Yallop: The arrow calculus]; [http://www.soi.city.ac.uk/~ross/papers/fop.html Ross Paterson: Programming with Arrows].}} 

+  
+  There are quite a few laws that instances of <code>Arrow</code> should 

+  satisfy {{noteref}}: 

+  
+  <haskell> 

+  arr id = id 

+  arr (h . g) = arr g >>> arr h 

+  first (arr g) = arr (g *** id) 

+  first (g >>> h) = first g >>> first h 

+  first g >>> arr (id *** h) = arr (id *** h) >>> first g 

+  first g >>> arr fst = arr fst >>> g 

+  first (first g) >>> arr assoc = arr assoc >>> first g 

+  
+  assoc ((x,y),z) = (x,(y,z)) 

+  </haskell> 

+  
+  Note that this version of the laws is slightly different than the laws given in the 

+  first two above references, since several of the laws have now been 

+  subsumed by the <code>Category</code> laws (in particular, the requirements that 

+  <code>id</code> is the identity arrow and that <code>(>>>)</code> is associative). The laws 

+  shown here follow those in Paterson's Programming with Arrows, which uses the 

+  <code>Category</code> class. 

+  
+  {{noteUnless categorytheoryinduced insomnolence is your cup of tea.}} 

+  
+  The reader is advised not to lose too much sleep over the <code>Arrow</code> 

+  laws {{noteref}}, since it is not essential to understand them in order to 

+  program with arrows. There are also laws that <code>ArrowChoice</code>, 

+  <code>ArrowApply</code>, and <code>ArrowLoop</code> instances should satisfy; the interested 

+  reader should consult [http://www.soi.city.ac.uk/~ross/papers/fop.html Paterson: Programming with Arrows]. 

+  
+  ==ArrowChoice== 

+  
+  Computations built using the <code>Arrow</code> class, like those built using 

+  the <code>Applicative</code> class, are rather inflexible: the structure of the computation 

+  is fixed at the outset, and there is no ability to choose between 

+  alternate execution paths based on intermediate results. 

+  The <code>ArrowChoice</code> class provides exactly such an ability: 

+  
+  <haskell> 

+  class Arrow (~>) => ArrowChoice (~>) where 

+  left :: (b ~> c) > (Either b d ~> Either c d) 

+  right :: (b ~> c) > (Either d b ~> Either d c) 

+  (+++) :: (b ~> c) > (b' ~> c') > (Either b b' ~> Either c c') 

+  () :: (b ~> d) > (c ~> d) > (Either b c ~> d) 

+  </haskell> 

+  
+  A comparison of <code>ArrowChoice</code> to <code>Arrow</code> will reveal a striking 

+  parallel between <code>left</code>, <code>right</code>, <code>(+++)</code>, <code>()</code> and <code>first</code>, 

+  <code>second</code>, <code>(***)</code>, <code>(&&&)</code>, respectively. Indeed, they are dual: 

+  <code>first</code>, <code>second</code>, <code>(***)</code>, and <code>(&&&)</code> all operate on product types 

+  (tuples), and <code>left</code>, <code>right</code>, <code>(+++)</code>, and <code>()</code> are the 

+  corresponding operations on sum types. In general, these operations 

+  create arrows whose inputs are tagged with <code>Left</code> or <code>Right</code>, and can 

+  choose how to act based on these tags. 

+  
+  * If <code>g</code> is an arrow from <code>b</code> to <code>c</code>, then <code>left g</code> is an arrow 

+  from <code>Either b d</code> to <code>Either c d</code>. On inputs tagged with <code>Left</code>, 

+  the <code>left g</code> arrow has the behavior of <code>g</code>; on inputs tagged with <code>Right</code>, it 

+  behaves as the identity. 

+  
+  * The <code>right</code> function, of course, is the mirror image of <code>left</code>. The arrow <code>right g</code> 

+  has the behavior of <code>g</code> on inputs tagged with <code>Right</code>. 

+  
+  * The <code>(+++)</code> operator performs “multiplexing”: <code>g +++ h</code> behaves as <code>g</code> 

+  on inputs tagged with <code>Left</code>, and as <code>h</code> on inputs tagged with 

+  <code>Right</code>. The tags are preserved. The <code>(+++)</code> operator is the ''sum'' (hence 

+  <code>+</code>) of two arrows, just as <code>(***)</code> is the product. 

+  
+  * The <code>()</code> operator is “merge” or “fanin”: the arrow <code>g  h</code> 

+  behaves as <code>g</code> on inputs tagged with <code>Left</code>, and <code>h</code> on inputs 

+  tagged with <code>Right</code>, but the tags are discarded (hence, <code>g</code> and <code>h</code> 

+  must have the same output type). The mnemonic is that <code>g  h</code> 

+  performs either <code>g</code> ''or'' <code>h</code> on its input. 

+  
+  The <code>ArrowChoice</code> class allows computations to choose among a finite number of 

+  execution paths, based on intermediate results. The possible 

+  execution paths must be known in advance, and explicitly assembled 

+  with <code>(+++)</code> or <code>()</code>. However, sometimes more flexibility is 

+  needed: we would like to be able to ''compute'' an arrow from 

+  intermediate results, and use this computed arrow to continue the 

+  computation. This is the power given to us by <code>ArrowApply</code>. 

+  
+  ==ArrowApply== 

+  
+  The <code>ArrowApply</code> type class is: 

+  
+  <haskell> 

+  class Arrow (~>) => ArrowApply (~>) where 

+  app :: (b ~> c, b) ~> c 

+  </haskell> 

+  
+  If we have computed an arrow as the output of some previous 

+  computation, then <code>app</code> allows us to apply that arrow to an input, 

+  producing its output as the output of <code>app</code>. As an exercise, the 

+  reader may wish to use <code>app</code> to implement an alternative “curried” 

+  version, <code>app2 :: b ~> ((b ~> c) ~> c)</code>. 

+  
+  This notion of being able to ''compute'' a new computation 

+  may sound familiar: 

+  this is exactly what the monadic bind operator <code>(>>=)</code> does. It 

+  should not particularly come as a surprise that <code>ArrowApply</code> and 

+  <code>Monad</code> are exactly equivalent in expressive power. In particular, 

+  <code>Kleisli m</code> can be made an instance of <code>ArrowApply</code>, and any instance 

+  of <code>ArrowApply</code> can be made a <code>Monad</code> (via the <code>newtype</code> wrapper 

+  <code>ArrowMonad</code>). As an exercise, the reader may wish to try 

+  implementing these instances: 

+  
+  <haskell> 

+  instance Monad m => ArrowApply (Kleisli m) where 

+  app =  exercise 

+  
+  newtype ArrowApply a => ArrowMonad a b = ArrowMonad (a () b) 

+  
+  instance ArrowApply a => Monad (ArrowMonad a) where 

+  return =  exercise 

+  (ArrowMonad a) >>= k =  exercise 

+  </haskell> 

+  
+  ==ArrowLoop== 

+  
+  The <code>ArrowLoop</code> type class is: 

+  
+  <haskell> 

+  class Arrow a => ArrowLoop a where 

+  loop :: a (b, d) (c, d) > a b c 

+  
+  trace :: ((b,d) > (c,d)) > b > c 

+  trace f b = let (c,d) = f (b,d) in c 

+  </haskell> 

+  
+  It describes arrows that can use recursion to compute results, and is 

+  used to desugar the <code>rec</code> construct in arrow notation (described 

+  below). 

+  
+  Taken by itself, the type of the <code>loop</code> method does not seem to tell 

+  us much. Its intention, however, is a generalization of the <code>trace</code> 

+  function which is also shown. The <code>d</code> component of the first arrow's 

+  output is fed back in as its own input. In other words, the arrow 

+  <code>loop g</code> is obtained by recursively “fixing” the second component of 

+  the input to <code>g</code>. 

+  
+  It can be a bit difficult to grok what the <code>trace</code> function is doing. 

+  How can <code>d</code> appear on the left and right sides of the <code>let</code>? Well, 

+  this is Haskell's laziness at work. There is not space here for a 

+  full explanation; the interested reader is encouraged to study the 

+  standard <code>fix</code> function, and to read [http://www.soi.city.ac.uk/~ross/papers/fop.html Paterson's arrow tutorial]. 

+  
+  ==Arrow notation== 

+  
+  Programming directly with the arrow combinators can be painful, 

+  especially when writing complex computations which need to retain 

+  simultaneous reference to a number of intermediate results. With 

+  nothing but the arrow combinators, such intermediate results must be 

+  kept in nested tuples, and it is up to the programmer to remember 

+  which intermediate results are in which components, and to swap, 

+  reassociate, and generally mangle tuples as necessary. This problem 

+  is solved by the special arrow notation supported by GHC, similar to 

+  <code>do</code> notation for monads, that allows names to be assigned to 

+  intermediate results while building up arrow computations. An example 

+  arrow implemented using arrow notation, taken from 

+  Paterson, is: 

+  
+  <haskell> 

+  class ArrowLoop (~>) => ArrowCircuit (~>) where 

+  delay :: b > (b ~> b) 

+  
+  counter :: ArrowCircuit (~>) => Bool ~> Int 

+  counter = proc reset > do 

+  rec output < idA < if reset then 0 else next 

+  next < delay 0 < output + 1 

+  idA < output 

+  </haskell> 

+  
+  This arrow is intended to 

+  represent a recursively defined counter circuit with a reset line. 

+  
+  There is not space here for a full explanation of arrow notation; the 

+  interested reader should consult [http://www.soi.city.ac.uk/~ross/papers/notation.html Paterson's paper introducing the 

+  notation], or his later [http://www.soi.city.ac.uk/~ross/papers/fop.html 

+  tutorial which presents a simplified version]. 

+  
+  ==Further reading== 

+  
+  An excellent starting place for the student of arrows is the [http://www.haskell.org/arrows/ arrows web page], which contains an 

+  introduction and many references. Some key papers on arrows include 

+  Hughes's original paper introducing arrows, [http://dx.doi.org/10.1016/S01676423(99)000234 Generalising monads to arrows], and [http://www.soi.city.ac.uk/~ross/papers/notation.html Paterson's paper on arrow notation]. 

+  
+  Both Hughes and Paterson later wrote accessible tutorials intended for a broader 

+  audience: [http://www.soi.city.ac.uk/~ross/papers/fop.html Paterson: Programming with Arrows] and [http://www.cse.chalmers.se/~rjmh/afparrows.pdf Hughes: Programming with Arrows]. 

+  
+  Although Hughes's goal in defining the <code>Arrow</code> class was to 

+  generalize <code>Monad</code>s, and it has been said that <code>Arrow</code> lies “between 

+  <code>Applicative</code> and <code>Monad</code>” in power, they are not directly 

+  comparable. The precise relationship remained in some confusion until 

+  [http://homepages.inf.ed.ac.uk/wadler/papers/arrowsandidioms/arrowsandidioms.pdf analyzed by Lindley, Wadler, and Yallop], who 

+  also invented a new calculus of arrows, based on the lambda calculus, 

+  which considerably simplifies the presentation of the arrow laws 

+  (see [http://homepages.inf.ed.ac.uk/wadler/papers/arrows/arrows.pdf The arrow calculus]). 

+  
+  Some examples of <code>Arrow</code>s include [http://www.haskell.org/yampa/ Yampa], the 

+  [http://www.fhwedel.de/~si/HXmlToolbox/ Haskell XML Toolkit], and the functional GUI library [[Grapefruit]]. 

+  
+  Some extensions to arrows have been explored; for example, the 

+  [http://www.cs.ru.nl/A.vanWeelden/biarrows/ <code>BiArrow</code>s of Alimarine et al.], for twoway instead of oneway 

+  computation. 

+  
+  The Haskell wiki has [[Research papers/Monads and Arrowslinks to many additional research papers relating to <code>Arrow</code>s]]. 

+  
+  =Comonad= 

+  
+  The final type class we will examine is <code>Comonad</code>. The <code>Comonad</code> class 

+  is the categorical dual of <code>Monad</code>; that is, <code>Comonad</code> is like <code>Monad</code> 

+  but with all the function arrows flipped. It is not actually in the 

+  standard Haskell libraries, but it has seen some interesting uses 

+  recently, so we include it here for completeness. 

+  
+  ==Definition== 

+  
+  The <code>Comonad</code> type class, defined in the <code>Control.Comonad</code> module of 

+  the [http://hackage.haskell.org/package/categoryextras categoryextras library], is: 

+  
+  <haskell> 

+  class Functor f => Copointed f where 

+  extract :: f a > a 

+  
+  class Copointed w => Comonad w where 

+  duplicate :: w a > w (w a) 

+  extend :: (w a > b) > w a > w b 

+  </haskell> 

+  
+  As you can see, <code>extract</code> is the dual of <code>return</code>, <code>duplicate</code> is the 

+  dual of <code>join</code>, and <code>extend</code> is the dual of <code>(>>=)</code> (although its 

+  arguments are in a different order). The definition 

+  of <code>Comonad</code> is a bit redundant (after all, the <code>Monad</code> class does not 

+  need <code>join</code>), but this is so that a <code>Comonad</code> can be defined by <code>fmap</code>, 

+  <code>extract</code>, and ''either'' <code>duplicate</code> or <code>extend</code>. Each has a 

+  default implementation in terms of the other. 

+  
+  A prototypical example of a <code>Comonad</code> instance is: 

+  
+  <haskell> 

+   Infinite lazy streams 

+  data Stream a = Cons a (Stream a) 

+  
+  instance Functor Stream where 

+  fmap g (Cons x xs) = Cons (g x) (fmap g xs) 

+  
+  instance Copointed Stream where 

+  extract (Cons x _) = x 

+  
+   'duplicate' is like the list function 'tails' 

+   'extend' computes a new Stream from an old, where the element 

+   at position n is computed as a function of everything from 

+   position n onwards in the old Stream 

+  instance Comonad Stream where 

+  duplicate s@(Cons x xs) = Cons s (duplicate xs) 

+  extend g s@(Cons x xs) = Cons (g s) (extend g xs) 

+   = fmap g (duplicate s) 

+  </haskell> 

+  
+  ==Further reading== 

+  
+  Dan Piponi explains in a blog post what [http://blog.sigfpe.com/2006/12/evaluatingcellularautomatais.html cellular automata have to do 

+  with comonads]. In another blog post, 

+  Conal Elliott has examined [http://conal.net/blog/posts/functionalinteractivebehavior/ a comonadic formulation of functional 

+  reactive programming]. Sterling 

+  Clover's blog post [http://fmapfixreturn.wordpress.com/2008/07/09/comonadsineverydaylife/ Comonads in everyday life] explains the relationship between 

+  comonads and zippers, and how comonads can be used to design a menu 

+  system for a web site. 

+  
+  Uustalu and Vene have a number of papers exploring ideas related to 

+  comonads and functional programming: 

+  * [http://dx.doi.org/10.1016/j.entcs.2008.05.029 Comonadic Notions of Computation] 

+  * [http://www.cs.ut.ee/~varmo/papers/sfp01book.ps.gz The dual of substitution is redecoration] 

+  * [http://dx.doi.org/10.1016/j.ic.2005.08.005 Recursive coalgebras from comonads] 

+  * [http://www.fing.edu.uy/~pardo/papers/njc01.ps.gz Recursion schemes from comonads] 

+  * [http://cs.ioc.ee/~tarmo/papers/essence.pdf The Essence of Dataflow Programming]. 

+  
+  =Acknowledgements= 

+  
+  A special thanks to all of those who taught me about standard Haskell 

+  type classes and helped me develop good intuition for them, 

+  particularly Jules Bean (quicksilver), Derek Elkins (ddarius), Conal 

+  Elliott (conal), Cale Gibbard (Cale), David House, Dan Piponi 

+  (sigfpe), and Kevin Reid (kpreid). 

+  
+  I also thank the many people who provided a mountain of helpful 

+  feedback and suggestions on a first draft of this article: David Amos, 

+  Kevin Ballard, Reid Barton, Doug Beardsley, Joachim Breitner, Andrew 

+  Cave, David Christiansen, Gregory Collins, Mark Jason Dominus, Conal 

+  Elliott, Yitz Gale, George Giorgidze, Steven Grady, Travis Hartwell, 

+  Steve Hicks, Philip Hölzenspies, Edward Kmett, Eric Kow, Serge Le 

+  Huitouze, Felipe Lessa, Stefan Ljungstrand, Eric Macaulay, Rob MacAulay, Simon Meier, 

+  Eric Mertens, Tim Newsham, Russell O'Connor, Conrad Parker, Walt 

+  RorieBaety, Colin Ross, Tom Schrijvers, Aditya Siram, C. Smith, 

+  Martijn van Steenbergen, Joe Thornber, Jared Updike, Rob Vollmert, 

+  Andrew Wagner, Louis Wasserman, and Ashley Yakeley, as well as a few 

+  only known to me by their IRC nicks: b_jonas, maltem, tehgeekmeister, 

+  and ziman. I have undoubtedly omitted a few inadvertently, which in 

+  no way diminishes my gratitude. 

+  
+  Finally, I would like to thank Wouter Swierstra for his fantastic work 

+  editing the Monad.Reader, and my wife Joyia for her patience during 

+  the process of writing the Typeclassopedia. 

+  
+  =About the author= 

+  
+  Brent Yorgey ([http://byorgey.wordpress.com/ blog], [http://www.cis.upenn.edu/~byorgey/ homepage]) is a firstyear 

+  Ph.D. student in the [http://www.cis.upenn.edu/~plclub/ programming languages group] at the University of 

+  Pennsylvania]. He enjoys teaching, creating EDSLs, 

+  playing Bach fugues, musing upon category theory, and cooking tasty 

+  lambdatreats for the denizens of #haskell. 

+  
+  =Colophon= 

+  
+  The Typeclassopedia was written by Brent Yorgey and initally published in March 2009. Painstakingly converted to wiki syntax by [[User:Geheimdienst]] in November 2011, after asking Brent's permission. If something like this ever needs to be done again, here are some vim commands that helped: 

+  
+  * <nowiki>%s/\\section{\([^}]*\)}/=\1=/gc</nowiki> 

+  * <nowiki>%s/\\subsection{\([^}]*\)}/==\1==/gc</nowiki> 

+  * <nowiki>%s/^ *\\item /\r* /gc</nowiki> 

+  * <nowiki>%s//—/gc</nowiki> 

+  * <nowiki>%s/\$\([^$]*\)\$/<math>\1<\/math>/gc</nowiki> 

+  * <nowiki>%s/\([^]*\)/<code>\1<\/code>/gc</nowiki> 

+  * <nowiki>%s/\\dots/.../gc</nowiki> 

+  * <nowiki>%s/^\\label{.*$//gc</nowiki> 

+  * <nowiki>%s/\\emph{\([^}]*\)}/''\1''/gc</nowiki> 

+  * <nowiki>%s/\\term{\([^}]*\)}/''\1''/gc</nowiki> 

+  
+  To get all the citations into the main text, I first tried processing the source with Tex or Lyx. This didn't work due to missing unfindable packages, syntax errors, and my general ineptitude with Tex. 

+  
+  I then went for the next best solution, which seemed to be extracting all instances of “\cite{something}” from the source and ''in that order'' pulling the referenced entries from the .bib file. This way you can go through the source file and sortedreferences file in parallel, copying over what you need, without searching back and forth in the .bib file. I used: 

+  
+  * egrep o "\cite\{[^\}]*\}" ~/typeclassopedia.lhs  cut c 6  tr "," "\n"  tr d "}" > /tmp/citations 

+  * for i in $(cat /tmp/citations); do grep A99 "$i" ~/typeclassopedia.bibegrep B99 '^\}$' m1 ; done > ~/typeclassorefssorted 

+  
+  [[Category:Applicative Functor]] 

[[Category:Arrow]] 
[[Category:Arrow]] 

+  [[Category:Functor]] 

⚫  
+  [[Category:Standard classes]] 

+  [[Category:Standard libraries]] 

+  [[Category:Standard packages]] 

+  [[Category:Standard types]] 
Revision as of 14:38, 15 November 2011
 By Brent Yorgey, byorgey@cis.upenn.edu
 As published 12 March 2009, issue 13 of the Monad.Reader
 Alternate formats: PDF / tex source / bibliography
The standard Haskell libraries feature a number of type classes with algebraic or categorytheoretic underpinnings. Becoming a fluent Haskell hacker requires intimate familiarity with them all, yet acquiring this familiarity often involves combing through a mountain of tutorials, blog posts, mailing list archives, and IRC logs.
The goal of this article is to serve as a starting point for the student of Haskell wishing to gain a firm grasp of its standard type classes. The essentials of each type class are introduced, with examples, commentary, and extensive references for further reading.
Contents
Introduction
Have you ever had any of the following thoughts?
 What the heck is a monoid, and how is it different from a monad?
 I finally figured out how to use Parsec with donotation, and someone told me I should use something called
Applicative
instead. Um, what?
 Someone in the #haskell IRC channel used
(***)
, and when I asked lambdabot to tell me its type, it printed out scary gobbledygook that didn't even fit on one line! Then someone usedfmap fmap fmap
and my brain exploded.
 When I asked how to do something I thought was really complicated, people started typing things like
zip.ap fmap.(id &&& wtf)
and the scary thing is that they worked! Anyway, I think those people must actually be robots because there's no way anyone could come up with that in two seconds off the top of their head.
If you have, look no further! You, too, can write and understand concise, elegant, idiomatic Haskell code with the best of them.
There are two keys to an expert Haskell hacker's wisdom:
 Understand the types.
 Gain a deep intuition for each type class and its relationship to other type classes, backed up by familiarity with many examples.
It's impossible to overstate the importance of the first; the patient student of type signatures will uncover many profound secrets. Conversely, anyone ignorant of the types in their code is doomed to eternal uncertainty. “Hmm, it doesn't compile ... maybe I'll stick in an
fmap
here ... nope, let's see ... maybe I need another (.)
somewhere? ... um ...”
The second key—gaining deep intuition, backed by examples—is also important, but much more difficult to attain. A primary goal of this article is to set you on the road to gaining such intuition. However—
 There is no royal road to Haskell. —Euclid
∗ Abstraction, intuition, and the “monad tutorial fallacy”, by Brent Yorgey This article can only be a starting point, since good intuition comes from hard work, not from learning the right metaphor ∗. Anyone who reads and understands all of it will still have an arduous journey ahead—but sometimes a good starting point makes a big difference.
It should be noted that this is not a Haskell tutorial; it is assumed that the reader is already familiar with the basics of Haskell, including the standard Prelude
, the type system, data types, and type classes.
The type classes we will be discussing and their interrelationships.
 Solid arrows point from the general to the specific; that is, if there is an arrow from Foo to Bar it means that every Bar is (or should be, or can be made into) a Foo.
 Dotted arrows indicate some other sort of relationship.

Monad
andArrowApply
are equivalent. 
Pointed
andComonad
are greyed out since they are not actually (yet) in the standard Haskell libraries (they are in the categoryextras library).
One more note before we begin. I've seen “type class” written as one word, “typeclass,” but let's settle this once and for all: the correct spelling uses two words (the title of this article notwithstanding), as evidenced by, for example, the Haskell 98 Revised Report, early papers on type classes like Type classes in Haskell and Type classes: exploring the design space, and Hudak et al.'s history of Haskell.
We now begin with the simplest type class of all: Functor
.
Functor
The Functor
class (haddock) is the most basic and ubiquitous type class in the Haskell libraries. A simple intuition is that a Functor
represents a “container” of some sort, along with the ability to apply a function uniformly to every element in the container. For example, a list is a container of elements, and we can apply a function to every element of a list using map
. A binary tree is also a container of elements, and it's not hard to come up with a way to recursively apply a function to every element in a tree.
Another intuition is that a Functor
represents some sort of “computational context.” This intuition is generally more useful, but is more difficult to explain, precisely because it is so general. Some examples later should help to clarify the Functor
ascontext point of view.
In the end, however, a Functor
is simply what it is defined to be; doubtless there are many examples of Functor
instances that don't exactly fit either of the above intuitions. The wise student will focus their attention on definitions and examples, without leaning too heavily on any particular metaphor. Intuition will come, in time, on its own.
Definition
The type class declaration for Functor
:
class Functor f where
fmap :: (a > b) > f a > f b
Functor
is exported by the Prelude
, so no special imports are needed to use it.
First, the f a
and f b
in the type signature for fmap
tell us that f
isn't just a type; it is a type constructor which takes another type as a parameter. (A more precise way to say this is that the kind of f
must be * > *
.) For example, Maybe
is such a type constructor: Maybe
is not a type in and of itself, but requires another type as a parameter, like Maybe Integer
. So it would not make sense to say instance Functor Integer
, but it could make sense to say instance Functor Maybe
.
Now look at the type of fmap
: it takes any function from a
to b
, and a value of type f a
, and outputs a value of type f b
. From the container point of view, the intention is that fmap
applies a function to each element of a container, without altering the structure of the container. From the context point of view, the intention is that fmap
applies a function to a value without altering its context. Let's look at a few specific examples.
Instances
∗ Recall that []
has two meanings in Haskell: it can either stand for the empty list, or, as here, it can represent the list type constructor (pronounced “listof”). In other words, the type [a]
(listofa
) can also be written ([] a)
.
∗ You might ask why we need a separate map
function. Why not just do away with the current listonly map
function, and rename fmap
to map
instead? Well, that's a good question. The usual argument is that someone just learning Haskell, when using map
incorrectly, would much rather see an error about lists than about Functor
s.
As noted before, the list constructor []
is a functor ∗; we can use the standard list function map
to apply a function to each element of a list ∗. The Maybe
type constructor is also a functor, representing a container which might hold a single element. The function fmap g
has no effect on Nothing
(there are no elements to which g
can be applied), and simply applies g
to the single element inside a Just
. Alternatively, under the context interpretation, the list functor represents a context of nondeterministic choice; that is, a list can be thought of as representing a single value which is nondeterministically chosen from among several possibilities (the elements of the list). Likewise, the Maybe
functor represents a context with possible failure. These instances are:
instance Functor [] where
fmap _ [] = []
fmap g (x:xs) = g x : fmap g xs
 or we could just say fmap = map
instance Functor Maybe where
fmap _ Nothing = Nothing
fmap g (Just a) = Just (g a)
As an aside, in idiomatic Haskell code you will often see the letter f
used to stand for both an arbitrary Functor
and an arbitrary function. In this tutorial, I will use f
only to represent Functor
s, and g
or h
to represent functions, but you should be aware of the potential confusion. In practice, what f
stands for should always be clear from the context, by noting whether it is part of a type or part of the code.
∗ Note that some of these instances are not exported by the Prelude
; to access them, you can import Control.Monad.Instances
. There are other Functor
instances in the standard libraries; here are a few ∗:

Either e
is an instance ofFunctor
;Either e a
represents a container which can contain either a value of typea
, or a value of typee
(often representing some sort of error condition). It is similar toMaybe
in that it represents possible failure, but it can carry some extra information about the failure as well.

((,) e)
represents a container which holds an “annotation” of typee
along with the actual value it holds.

((>) e)
, the type of functions which take a value of typee
as a parameter, is aFunctor
. It would be clearer to write it as(e >)
, by analogy with an operator section like(1+)
, but that syntax is not allowed. However, you can certainly think of it as(e >)
. As a container,(e > a)
represents a (possibly infinite) set of values ofa
, indexed by values ofe
. Alternatively, and more usefully,(e >)
can be thought of as a context in which a value of typee
is available to be consulted in a readonly fashion. This is also why((>) e)
is sometimes referred to as the reader monad; more on this later.

IO
is aFunctor
; a value of typeIO a
represents a computation producing a value of typea
which may have I/O effects. Ifm
computes the valuex
while producing some I/O effects, thenfmap g m
will compute the valueg x
while producing the same I/O effects.
 Many standard types from the containers library (such as
Tree
,Map
,Sequence
, andStream
) are instances ofFunctor
. A notable exception isSet
, which cannot be made aFunctor
in Haskell (although it is certainly a mathematical functor) since it requires anOrd
constraint on its elements;fmap
must be applicable to any typesa
andb
.
A good exercise is to implement Functor
instances for Either e
, ((,) e)
, and ((>) e)
.
Laws
As far as the Haskell language itself is concerned, the only requirement to be a Functor
is an implementation of fmap
with the proper type. Any sensible Functor
instance, however, will also satisfy the functor laws, which are part of the definition of a mathematical functor. There are two:
fmap id = id
fmap (g . h) = (fmap g) . (fmap h)
∗ Technically, these laws make f
and fmap
together an endofunctor on Hask, the category of Haskell types (ignoring ⊥, which is a party pooper). See Wikibook: Category theory.
Together, these laws ensure that fmap g
does not change the structure of a container, only the elements. Equivalently, and more simply, they ensure that fmap g
changes a value without altering its context ∗.
The first law says that mapping the identity function over every item in a container has no effect. The second says that mapping a composition of two functions over every item in a container is the same as first mapping one function, and then mapping the other.
As an example, the following code is a “valid” instance of Functor
(it typechecks), but it violates the functor laws. Do you see why?
 Evil Functor instance
instance Functor [] where
fmap _ [] = []
fmap g (x:xs) = g x : g x : fmap g xs
Any Haskeller worth their salt would reject this code as a gruesome abomination.
Intuition
There are two fundamental ways to think about fmap
. The first has already been touched on: it takes two parameters, a function and a container, and applies the function “inside” the container, producing a new container. Alternately, we can think of fmap
as applying a function to a value in a context (without altering the context).
Just like all other Haskell functions of “more than one parameter,” however, fmap
is actually curried: it does not really take two parameters, but takes a single parameter and returns a function. For emphasis, we can write fmap
's type with extra parentheses: fmap :: (a > b) > (f a > f b)
. Written in this form, it is apparent that fmap
transforms a “normal” function (g :: a > b
) into one which operates over containers/contexts (fmap g :: f a > f b
). This transformation is often referred to as a lift; fmap
“lifts” a function from the “normal world” into the “f
world.”
Further reading
A good starting point for reading about the category theory behind the concept of a functor is the excellent Haskell wikibook page on category theory.
Pointed
∗ It is, however, a type class in the categoryextras library.
The Pointed
type class represents pointed functors. It is not actually a type class in the standard libraries ∗. But it could be, and it's useful in understanding a few other type classes, notably Applicative
and Monad
, so let's pretend for a minute.
Given a Functor
, the Pointed
class represents the additional ability to put a value into a “default context.” Often, this corresponds to creating a container with exactly one element, but it is more general than that. The type class declaration for Pointed
is:
class Functor f => Pointed f where
pure :: a > f a  aka singleton, return, unit, point
Most of the standard Functor
instances could also be instances of Pointed
—for example, the Maybe
instance of Pointed
is pure = Just
; there are many possible implementations for lists, the most natural of which is pure x = [x]
; for ((>) e)
it is ... well, I'll let you work it out. (Just follow the types!)
One example of a Functor
which is not Pointed
is ((,) e)
. If you try implementing pure :: a > (e,a)
you will quickly see why: since the type e
is completely arbitrary, there is no way to generate a value of type e
out of thin air! However, as we will see, ((,) e)
can be made Pointed
if we place an additional restriction on e
which allows us to generate a default value of type e
(the most common solution is to make e
an instance of Monoid
).
∗ For those interested in category theory, this law states precisely that pure
is a natural transformation from the identity functor to f
. The Pointed
class has only one law ∗:
fmap g . pure = pure . g
∗ ... modulo ⊥, seq
, and assuming a lawful Functor
instance.
However, you need not worry about it: this law is actually a socalled “free theorem” guaranteed by parametricity (see Wadler's Theorems for free!); it's impossible to write an instance of Pointed
which does not satisfy it ∗.
Applicative
A somewhat newer addition to the pantheon of standard Haskell type classes, applicative functors (see their haddock) represent an abstraction lying exactly in between Functor
and Monad
, first described by McBride and Paterson. The title of their classic paper, Applicative Programming with Effects, gives a hint at the intended intuition behind the Applicative
type class. It encapsulates certain sorts of “effectful” computations in a functionally pure way, and encourages an “applicative” programming style. Exactly what these things mean will be seen later.
Definition
The Applicative
class adds a single capability to Pointed
functors. Recall that Functor
allows us to lift a “normal” function to a function on computational contexts. But fmap
doesn't allow us to apply a function which is itself in a context to a value in another context. Applicative
gives us just such a tool. Here is the type class declaration for Applicative
, as defined in Control.Applicative
:
class Functor f => Applicative f where
pure :: a > f a
(<*>) :: f (a > b) > f a > f b
Note that every Applicative
must also be a Functor
. In fact, as we will see, fmap
can be implemented using the Applicative
methods, so every Applicative
is a functor whether we like it or not; the Functor
constraint forces us to be honest.
∗ Recall that ($)
is just function application: f $ x = f x
.
As always, it's crucial to understand the type signature of (<*>)
. The best way of thinking about it comes from noting that the type of (<*>)
is similar to the type of ($)
∗, but with everything enclosed in an f
. In other words, (<*>)
is just function application within a computational context. The type of (<*>)
is also very similar to the type of fmap
; the only difference is that the first parameter is f (a > b)
, a function in a context, instead of a “normal” function (a > b)
.
Of course, pure
looks rather familiar. If we actually had a Pointed
type class, Applicative
could instead be defined as:
class Pointed f => Applicative' f where
(<*>) :: f (a > b) > f a > f b
Laws
∗ See haddock for Applicative, Applicative programming with effects
There are several laws that Applicative
instances should satisfy ∗, but only one is crucial to developing intuition, because it specifies how Applicative
should relate to Functor
(the other four mostly specify the exact sense in which pure
deserves its name). This law is:
fmap g x = pure g <*> x
It says that mapping a pure function g
over a context x
is the same as first injecting g
into a context with pure
, and then applying it to x
with (<*>)
. In other words, we can decompose fmap
into two more atomic operations: injection into a context, and application within a context. The Control.Applicative
module also defines (<$>)
as a synonym for fmap
, so the above law can also be expressed as:
g <$> x = pure g <*> x
.
Instances
Most of the standard types which are instances of Functor
are also instances of Applicative
.
Maybe
can easily be made an instance of Applicative
; writing such an instance is left as an exercise for the reader.
The list type constructor []
can actually be made an instance of Applicative
in two ways; essentially, it comes down to whether we want to think of lists as ordered collections of elements, or as contexts representing multiple results of a nondeterministic computation (see Wadler's How to replace failure by a list of successes).
Let's first consider the collection point of view. Since there can only be one instance of a given type class for any particular type, one or both of the list instances of Applicative
need to be defined for a newtype
wrapper; as it happens, the nondeterministic computation instance is the default, and the collection instance is defined in terms of a newtype
called ZipList
. This instance is:
newtype ZipList a = ZipList { getZipList :: [a] }
instance Applicative ZipList where
pure = undefined  exercise
(ZipList gs) <*> (ZipList xs) = ZipList (zipWith ($) gs xs)
To apply a list of functions to a list of inputs with (<*>)
, we just match up the functions and inputs elementwise, and produce a list of the resulting outputs. In other words, we “zip” the lists together with function application, ($)
; hence the name ZipList
. As an exercise, determine the correct definition of pure
—there is only one implementation that satisfies the law (see section “Laws”).
The other Applicative
instance for lists, based on the nondeterministic computation point of view, is:
instance Applicative [] where
pure x = [x]
gs <*> xs = [ g x  g < gs, x < xs ]
Instead of applying functions to inputs pairwise, we apply each function to all the inputs in turn, and collect all the results in a list.
Now we can write nondeterministic computations in a natural style. To add the numbers 3
and 4
deterministically, we can of course write (+) 3 4
. But suppose instead of 3
we have a nondeterministic computation that might result in 2
, 3
, or 4
; then we can write
pure (+) <*> [2,3,4] <*> pure 4
or, more idiomatically,
(+) <$> [2,3,4] <*> pure 4.
There are several other Applicative
instances as well:

IO
is an instance ofApplicative
, and behaves exactly as you would think: wheng <$> m1 <*> m2 <*> m3
is executed, the effects from themi
's happen in order from left to right.

((,) a)
is anApplicative
, as long asa
is an instance ofMonoid
(section Monoid). Thea
values are accumulated in parallel with the computation.
 The
Applicative
module defines theConst
type constructor; a value of typeConst a b
simply contains ana
. This is an instance ofApplicative
for anyMonoid a
; this instance becomes especially useful in conjunction with things likeFoldable
(section Foldable).
 The
WrappedMonad
andWrappedArrow
newtypes make any instances ofMonad
(section Monad) orArrow
(section Arrow) respectively into instances ofApplicative
; as we will see when we study those type classes, both are strictly more expressive thanApplicative
, in the sense that theApplicative
methods can be implemented in terms of their methods.
Intuition
McBride and Paterson's paper introduces the notation to denote function application in a computational context. If each has type for some applicative functor , and has type , then the entire expression has type . You can think of this as applying a function to multiple “effectful” arguments. In this sense, the double bracket notation is a generalization of fmap
, which allows us to apply a function to a single argument in a context.
Why do we need Applicative
to implement this generalization of fmap
? Suppose we use fmap
to apply g
to the first parameter x1
. Then we get something of type f (t2 > ... t)
, but now we are stuck: we can't apply this functioninacontext to the next argument with fmap
. However, this is precisely what (<*>)
allows us to do.
This suggests the proper translation of the idealized notation into Haskell, namely
g <$> x1 <*> x2 <*> ... <*> xn,
recalling that Control.Applicative
defines (<$>)
as convenient infix shorthand for fmap
. This is what is meant by an “applicative style”—effectful computations can still be described in terms of function application; the only difference is that we have to use the special operator (<*>)
for application instead of simple juxtaposition.
Further reading
There are many other useful combinators in the standard libraries implemented in terms of pure
and (<*>)
: for example, (*>)
, (<*)
, (<**>)
, (<$)
, and so on (see haddock for Applicative). Judicious use of such secondary combinators can often make code using Applicative
s much easier to read.
McBride and Paterson's original paper is a treasuretrove of information and examples, as well as some perspectives on the connection between Applicative
and category theory. Beginners will find it difficult to make it through the entire paper, but it is extremely wellmotivated—even beginners will be able to glean something from reading as far as they are able.
Conal Elliott has been one of the biggest proponents of Applicative
. For example, the Pan library for functional images and the reactive library for functional reactive programming (FRP) make key use of it; his blog also contains many examples of Applicative
in action. Building on the work of McBride and Paterson, Elliott also built the TypeCompose library, which embodies the observation (among others) that Applicative
types are closed under composition; therefore, Applicative
instances can often be automatically derived for complex types built out of simpler ones.
Although the Parsec parsing library (paper) was originally designed for use as a monad, in its most common use cases an Applicative
instance can be used to great effect; Bryan O'Sullivan's blog post is a good starting point. If the extra power provided by Monad
isn't needed, it's usually a good idea to use Applicative
instead.
A couple other nice examples of Applicative
in action include the ConfigFile and HSQL libraries and the formlets library.
Monad
It's a safe bet that if you're reading this article, you've heard of monads—although it's quite possible you've never heard of Applicative
before, or Arrow
, or even Monoid
. Why are monads such a big deal in Haskell? There are several reasons.
 Haskell does, in fact, single out monads for special attention by making them the framework in which to construct I/O operations.
 Haskell also singles out monads for special attention by providing a special syntactic sugar for monadic expressions: the
do
notation. 
Monad
has been around longer than various other abstract models of computation such asApplicative
orArrow
.  The more monad tutorials there are, the harder people think monads must be, and the more new monad tutorials are written by people who think they finally “get” monads (the monad tutorial fallacy).
I will let you judge for yourself whether these are good reasons.
In the end, despite all the hoopla, Monad
is just another type class. Let's take a look at its definition.
Definition
The type class declaration for Monad
(haddock) is:
class Monad m where
return :: a > m a
(>>=) :: m a > (a > m b) > m b
(>>) :: m a > m b > m b
m >> n = m >>= \_ > n
fail :: String > m a
The Monad
type class is exported by the Prelude
, along with a few standard instances. However, many utility functions are found in Control.Monad
, and there are also several instances (such as ((>) e)
) defined in Control.Monad.Instances
.
Let's examine the methods in the Monad
class one by one. The type of return
should look familiar; it's the same as pure
. Indeed, return
is pure
, but with an unfortunate name. (Unfortunate, since someone coming from an imperative programming background might think that return
is like the C or Java keyword of the same name, when in fact the similarities are minimal.) From a mathematical point of view, every monad is a pointed functor (indeed, an applicative functor), but for historical reasons, the Monad
type class declaration unfortunately does not require this.
We can see that (>>)
is a specialized version of (>>=)
, with a default implementation given. It is only included in the type class declaration so that specific instances of Monad
can override the default implementation of (>>)
with a more efficient one, if desired. Also, note that although _ >> n = n
would be a typecorrect implementation of (>>)
, it would not correspond to the intended semantics: the intention is that m >> n
ignores the result of m
, but not its effects.
The fail
function is an awful hack that has no place in the Monad
class; more on this later.
The only really interesting thing to look at—and what makes Monad
strictly more powerful than Pointed
or Applicative
—is (>>=)
, which is often called bind. An alternative definition of Monad
could look like:
class Applicative m => Monad' m where
(>>=) :: m a > (a > m b) > m b
We could spend a while talking about the intuition behind (>>=)
—and we will. But first, let's look at some examples.
Instances
Even if you don't understand the intuition behind the Monad
class, you can still create instances of it by just seeing where the types lead you. You may be surprised to find that this actually gets you a long way towards understanding the intuition; at the very least, it will give you some concrete examples to play with as you read more about the Monad
class in general. The first few examples are from the standard Prelude
; the remaining examples are from the monad transformer library (mtl).
 The simplest possible instance of
Monad
isIdentity
(see haddock), which is described in Dan Piponi's highly recommended blog post on The Trivial Monad. Despite being “trivial,” it is a great introduction to theMonad
type class, and contains some good exercises to get your brain working.  The next simplest instance of
Monad
isMaybe
. We already know how to writereturn
/pure
forMaybe
. So how do we write(>>=)
? Well, let's think about its type. Specializing forMaybe
, we have
(>>=) :: Maybe a > (a > Maybe b) > Maybe b.
 If the first argument to
(>>=)
isJust x
, then we have something of typea
(namely,x
), to which we can apply the second argument—resulting in aMaybe b
, which is exactly what we wanted. What if the first argument to(>>=)
isNothing
? In that case, we don't have anything to which we can apply thea > Maybe b
function, so there's only one thing we can do: yieldNothing
. This instance is:
instance Monad Maybe where
return = Just
(Just x) >>= g = g x
Nothing >>= _ = Nothing
 We can already get a bit of intuition as to what is going on here: if we build up a computation by chaining together a bunch of functions with
(>>=)
, as soon as any one of them fails, the entire computation will fail (becauseNothing >>= f
isNothing
, no matter whatf
is). The entire computation succeeds only if all the constituent functions individually succeed. So theMaybe
monad models computations which may fail.
 The
Monad
instance for the list constructor[]
is similar to itsApplicative
instance; I leave its implementation as an exercise. Follow the types!
 Of course, the
IO
constructor is famously aMonad
, but its implementation is somewhat magical, and may in fact differ from compiler to compiler. It is worth emphasizing that theIO
monad is the only monad which is magical. It allows us to build up, in an entirely pure way, values representing possibly effectful computations. The special valuemain
, of typeIO ()
, is taken by the runtime and actually executed, producing actual effects. Every other monad is functionally pure, and requires no special compiler support. We often speak of monadic values as “effectful computations,” but this is because some monads allow us to write code as if it has side effects, when in fact the monad is hiding the plumbing which allows these apparent side effects to be implemented in a functionally pure way.
 As mentioned earlier,
((>) e)
is known as the reader monad, since it describes computations in which a value of typee
is available as a readonly environment. It is worth trying to write aMonad
instance for((>) e)
yourself.
 The
Control.Monad.Reader
module (haddock) provides theReader e a
type, which is just a convenientnewtype
wrapper around(e > a)
, along with an appropriateMonad
instance and someReader
specific utility functions such asask
(retrieve the environment),asks
(retrieve a function of the environment), andlocal
(run a subcomputation under a different environment).
 The
Control.Monad.Writer
module (haddock) provides theWriter
monad, which allows information to be collected as a computation progresses.Writer w a
is isomorphic to(a,w)
, where the output valuea
is carried along with an annotation or “log” of typew
, which must be an instance ofMonoid
(section Monoid); the special functiontell
performs logging.
 The
Control.Monad.State
module (haddock) provides theState s a
type, anewtype
wrapper arounds > (a,s)
. Something of typeState s a
represents a stateful computation which produces ana
but can access and modify the state of types
along the way. The module also providesState
specific utility functions such asget
(read the current state),gets
(read a function of the current state),put
(overwrite the state), andmodify
(apply a function to the state).
 The
Control.Monad.Cont
module (haddock) provides theCont
monad, which represents computations in continuationpassing style. It can be used to suspend and resume computations, and to implement nonlocal transfers of control, coroutines, other complex control structures—all in a functionally pure way.Cont
has been called the “mother of all monads” because of its universal properties.
Intuition
Let's look more closely at the type of (>>=)
. The basic intuition is that it combines two computations into one larger computation. The first argument, m a
, is the first computation. However, it would be boring if the second argument were just an m b
; then there would be no way for the computations to interact with one another. So, the second argument to (>>=)
has type a > m b
: a function of this type, given a result of the first computation, can produce a second computation to be run. In other words, x >>= k
is a computation which runs x
, and then uses the result(s) of x
to decide what computation to run second, using the output of the second computation as the result of the entire computation.
Intuitively, it is this ability to use the output from previous computations to decide what computations to run next that makes Monad
more powerful than Applicative
. The structure of an Applicative
computation is fixed, whereas the structure of a Monad
computation can change based on intermediate results.
To see the increased power of Monad
from a different point of view, let's see what happens if we try to implement (>>=)
in terms of fmap
, pure
, and (<*>)
. We are given a value x
of type m a
, and a function k
of type a > m b
, so the only thing we can do is apply k
to x
. We can't apply it directly, of course; we have to use fmap
to lift it over the m
. But what is the type of fmap k
? Well, it's m a > m (m b)
. So after we apply it to x
, we are left with something of type m (m b)
—but now we are stuck; what we really want is an m b
, but there's no way to get there from here. We can add m
's using pure
, but we have no way to collapse multiple m
's into one.
This ability to collapse multiple m
's is exactly the ability provided by the function join :: m (m a) > m a
, and it should come as no surprise that an alternative definition of Monad
can be given in terms of join
:
class Applicative m => Monad'' m where
join :: m (m a) > m a
In fact, monads in category theory are defined in terms of return
, fmap
, and join
(often called , , and in the mathematical literature). Haskell uses the equivalent formulation in terms of (>>=)
instead of join
since it is more convenient to use; however, sometimes it can be easier to think about Monad
instances in terms of join
, since it is a more “atomic” operation. (For example, join
for the list monad is just concat
.) An excellent exercise is to implement (>>=)
in terms of fmap
and join
, and to implement join
in terms of (>>=)
.
Utility functions
The Control.Monad
module (haddock) provides a large number of convenient utility functions, all of which can be implemented in terms of the basic Monad
operations (return
and (>>=)
in particular). We have already seen one of them, namely, join
. We also mention some other noteworthy ones here; implementing these utility functions oneself is a good exercise. For a more detailed guide to these functions, with commentary and example code, see HenkJan van Tuyl's tour.

liftM :: Monad m => (a > b) > m a > m b
. This should be familiar; of course, it is justfmap
. The fact that we have bothfmap
andliftM
is an unfortunate consequence of the fact that theMonad
type class does not require aFunctor
instance, even though mathematically speaking, every monad is a functor. However,fmap
andliftM
are essentially interchangeable, since it is a bug (in a social rather than technical sense) for any type to be an instance ofMonad
without also being an instance ofFunctor
.

ap :: Monad m => m (a > b) > m a > m b
should also be familiar: it is equivalent to(<*>)
, justifying the claim that theMonad
interface is strictly more powerful thanApplicative
. We can make anyMonad
into an instance ofApplicative
by settingpure = return
and(<*>) = ap
.

sequence :: Monad m => [m a] > m [a]
takes a list of computations and combines them into one computation which collects a list of their results. It is again something of a historical accident thatsequence
has aMonad
constraint, since it can actually be implemented only in terms ofApplicative
. There is also an additional generalization ofsequence
to structures other than lists, which will be discussed in the section onTraversable
.

replicateM :: Monad m => Int > m a > m [a]
is simply a combination ofreplicate
andsequence
.

when :: Monad m => Bool > m () > m ()
conditionally executes a computation, evaluating to its second argument if the test isTrue
, and toreturn ()
if the test isFalse
. A collection of other sorts of monadic conditionals can be found in the IfElse package.

mapM :: Monad m => (a > m b) > [a] > m [b]
maps its first argument over the second, andsequence
s the results. TheforM
function is justmapM
with its arguments reversed; it is calledforM
since it models generalizedfor
loops: the list[a]
provides the loop indices, and the functiona > m b
specifies the “body” of the loop for each index.

(=<<) :: Monad m => (a > m b) > m a > m b
is just(>>=)
with its arguments reversed; sometimes this direction is more convenient since it corresponds more closely to function application.

(>=>) :: Monad m => (a > m b) > (b > m c) > a > m c
is sort of like function composition, but with an extram
on the result type of each function, and the arguments swapped. We'll have more to say about this operation later.
 The
guard
function is for use with instances ofMonadPlus
, which is discussed at the end of theMonoid
section.
Many of these functions also have “underscored” variants, such as sequence_
and mapM_
; these variants throw away the results of the computations passed to them as arguments, using them only for their side effects.
Laws
There are several laws that instances of Monad
should satisfy Monad laws. The standard presentation is:
return a >>= k = k a
m >>= return = m
m >>= (\x > k x >>= h) = (m >>= k) >>= h
fmap f xs = xs >>= return . f = liftM f xs
The first and second laws express the fact that return
behaves nicely: if we inject a value a
into a monadic context with return
, and then bind to k
, it is the same as just applying k
to a
in the first place; if we bind a computation m
to return
, nothing changes. The third law essentially says that (>>=)
is associative, sort of. The last law ensures that fmap
and liftM
are the same for types which are instances of both Functor
and Monad
—which, as already noted, should be every instance of Monad
.
∗ I like to pronounce this operator “fish,” but that's probably not the canonical pronunciation ...
However, the presentation of the above laws, especially the third, is marred by the asymmetry of (>>=)
. It's hard to look at the laws and see what they're really saying. I prefer a much more elegant version of the laws, which is formulated in terms of (>=>)
∗. Recall that (>=>)
“composes” two functions of type a > m b
and b > m c
. You can think of something of type a > m b
(roughly) as a function from a
to b
which may also have some sort of effect in the context corresponding to m
. (Note that return
is such a function.) (>=>)
lets us compose these “effectful functions,” and we would like to know what properties (>=>)
has. The monad laws reformulated in terms of (>=>)
are:
return >=> g = g
g >=> return = g
(g >=> h) >=> k = g >=> (h >=> k)
∗ As fans of category theory will note, these laws say precisely that functions of type a > m b
are the arrows of a category with (>=>)
as composition! Indeed, this is known as the Kleisli category of the monad m
. It will come up again when we discuss Arrow
s.
Ah, much better! The laws simply state that return
is the identity of (>=>)
, and that (>=>)
is associative ∗. Working out the equivalence between these two formulations, given the definition g >=> h = \x > g x >>= h
, is left as an exercise.
There is also a formulation of the monad laws in terms of fmap
, return
, and join
; for a discussion of this formulation, see the Haskell wikibook page on category theory.
do
notation
Haskell's special do
notation supports an “imperative style” of programming by providing syntactic sugar for chains of monadic expressions. The genesis of the notation lies in realizing that something like a >>= \x > b >> c >>= \y > d
can be more readably written by putting successive computations on separate lines:
a >>= \x >
b >>
c >>= \y >
d
This emphasizes that the overall computation consists of four computations a
, b
, c
, and d
, and that x
is bound to the result of a
, and y
is bound to the result of c
(b
, c
, and d
are allowed to refer to x
, and d
is allowed to refer to y
as well). From here it is not hard to imagine a nicer notation:
do { x < a ;
b ;
y < c ;
d
}
(The curly braces and semicolons may optionally be omitted; the Haskell parser uses layout to determine where they should be inserted.) This discussion should make clear that do
notation is just syntactic sugar. In fact, do
blocks are recursively translated into monad operations (almost) like this:
do e ⇨ e do { e; stmts } ⇨ e >> do { stmts } do { v < e; stmts } ⇨ e >>= \v > do { stmts } do { let decls; stmts} ⇨ let decls in do { stmts }
This is not quite the whole story, since v
might be a pattern instead of a variable. For example, one can write
do (x:xs) < foo
bar x
but what happens if foo
produces an empty list? Well, remember that ugly fail
function in the Monad
type class declaration? That's what happens. See section 3.14 of the Haskell Report for the full details. See also the discussion of MonadPlus
and MonadZero
in the section on other monoidal classes.
A final note on intuition: do
notation plays very strongly to the “computational context” point of view rather than the “container” point of view, since the binding notation x < m
is suggestive of “extracting” a single x
from m
and doing something with it. But m
may represent some sort of a container, such as a list or a tree; the meaning of x < m
is entirely dependent on the implementation of (>>=)
. For example, if m
is a list, x < m
actually means that x
will take on each value from the list in turn.
Monad transformers
One would often like to be able to combine two monads into one: for example, to have stateful, nondeterministic computations (State
+ []
), or computations which may fail and can consult a readonly environment (Maybe
+ Reader
), and so on. Unfortunately, monads do not compose as nicely as applicative functors (yet another reason to use Applicative
if you don't need the full power that Monad
provides), but some monads can be combined in certain ways.
The monad transformer library mtl provides a number of monad transformers, such as StateT
, ReaderT
, ErrorT
(haddock), and (soon) MaybeT
, which can be applied to other monads to produce a new monad with the effects of both. For example, StateT s Maybe
is an instance of Monad
; computations of type StateT s Maybe a
may fail, and have access to a mutable state of type s
. These transformers can be multiply stacked. One thing to keep in mind while using monad transformers is that the order of composition matters. For example, when a StateT s Maybe a
computation fails, the state ceases being updated; on the other hand, the state of a MaybeT (State s) a
computation may continue to be modified even after the computation has failed. (This may seem backwards, but it is correct. Monad transformers build composite monads “inside out”; for example, MaybeT (State s) a
is isomorphic to s > Maybe (a, s)
. Lambdabot has an indispensable @unmtl
command which you can use to “unpack” a monad transformer stack in this way.)
All monad transformers should implement the MonadTrans
type class, defined in Control.Monad.Trans
:
class MonadTrans t where
lift :: Monad m => m a > t m a
It allows arbitrary computations in the base monad m
to be “lifted” into computations in the transformed monad t m
. (Note that type application associates to the left, just like function application, so t m a = (t m) a
. As an exercise, you may wish to work out t
's kind, which is rather more interesting than most of the kinds we've seen up to this point.) However, you should only have to think about MonadTrans
when defining your own monad transformers, not when using predefined ones.
∗ The only problem with this scheme is the quadratic number of instances required as the number of standard monad transformers grows—but as the current set of standard monad transformers seems adequate for most common use cases, this may not be that big of a deal.
There are also type classes such as MonadState
, which provides statespecific methods like get
and put
, allowing you to conveniently use these methods not only with State
, but with any monad which is an instance of MonadState
—including MaybeT (State s)
, StateT s (ReaderT r IO)
, and so on. Similar type classes exist for Reader
, Writer
, Cont
, IO
, and others ∗.
There are two excellent references on monad transformers. Martin Grabmüller's Monad Transformers Step by Step is a thorough description, with running examples, of how to use monad transformers to elegantly build up computations with various effects. Cale Gibbard's article on how to use monad transformers is more practical, describing how to structure code using monad transformers to make writing it as painless as possible. Another good starting place for learning about monad transformers is a blog post by Dan Piponi.
MonadFix
The MonadFix
class describes monads which support the special fixpoint operation mfix :: (a > m a) > m a
, which allows the output of monadic computations to be defined via recursion. This is supported in GHC and Hugs by a special “recursive do” notation, mdo
. For more information, see Levent Erkök's thesis, Value Recursion in Monadic Computations.
Further reading
Philip Wadler was the first to propose using monads to structure functional programs. His paper is still a readable introduction to the subject.
Much of the monad transformer library mtl, including the Reader
, Writer
, State
, and other monads, as well as the monad transformer framework itself, was inspired by Mark Jones's classic paper Functional Programming with Overloading and HigherOrder Polymorphism. It's still very much worth a read—and highly readable—after almost fifteen years.
∗ {{{1}}}
There are, of course, numerous monad tutorials of varying quality ∗.
A few of the best include Cale Gibbard's Monads as containers and Monads as computation; Jeff Newbern's All About Monads, a comprehensive guide with lots of examples; and Dan Piponi's You Could Have Invented Monads!, which features great exercises. If you just want to know how to use IO
, you could consult the Introduction to IO. Even this is just a sampling; the monad tutorials timeline is a more complete list. (All these monad tutorials have prompted parodies like think of a monad ... as well as other kinds of backlash like Monads! (and Why Monad Tutorials Are All Awful) or Abstraction, intuition, and the “monad tutorial fallacy”.)
Other good monad references which are not necessarily tutorials include HenkJan van Tuyl's tour of the functions in Control.Monad
, Dan Piponi's field guide, and Tim Newsham's What's a Monad?. There are also many blog articles which have been written on various aspects of monads; a collection of links can be found under Blog articles/Monads.
One of the quirks of the Monad
class and the Haskell type system is that it is not possible to straightforwardly declare Monad
instances for types which require a class constraint on their data, even if they are monads from a mathematical point of view. For example, Data.Set
requires an Ord
constraint on its data, so it cannot be easily made an instance of Monad
. A solution to this problem was first described by Eric Kidd, and later made into a library named rmonad by Ganesh Sittampalam and Peter Gavin.
There are many good reasons for eschewing do
notation; some have gone so far as to [[Do_notation_considered_harmfulconsider it harmful].
Monads can be generalized in various ways; for an exposition of one possibility, see Robert Atkey's paper on parameterized monads, or Dan Piponi's Beyond Monads.
For the categorically inclined, monads can be viewed as monoids (From Monoids to Monads) and also as closure operators Triples and Closure. Derek Elkins's article in issue 13 of the Monad.Reader contains an exposition of the categorytheoretic underpinnings of some of the standard Monad
instances, such as State
and Cont
. There is also an alternative way to compose monads, using coproducts, as described by Lüth and Ghani, although this method has not (yet?) seen widespread use.
Links to many more research papers related to monads can be found under Research papers/Monads and arrows.
Monoid
A monoid is a set together with a binary operation which combines elements from . The operator is required to be associative (that is, , for any which are elements of ), and there must be some element of which is the identity with respect to . (If you are familiar with group theory, a monoid is like a group without the requirement that inverses exist.) For example, the natural numbers under addition form a monoid: the sum of any two natural numbers is a natural number; for any natural numbers , , and ; and zero is the additive identity. The integers under multiplication also form a monoid, as do natural numbers under , Boolean values under conjunction and disjunction, lists under concatenation, functions from a set to itself under composition ... Monoids show up all over the place, once you know to look for them.
Definition
The definition of the Monoid
type class (defined in
Data.Monoid
; haddock) is:
class Monoid a where
mempty :: a
mappend :: a > a > a
mconcat :: [a] > a
mconcat = foldr mappend mempty
The mempty
value specifies the identity element of the monoid, and mappend
is the binary operation. The default definition for mconcat
“reduces” a list of elements by combining them all with mappend
,
using a right fold. It is only in the Monoid
class so that specific
instances have the option of providing an alternative, more efficient
implementation; usually, you can safely ignore mconcat
when creating
a Monoid
instance, since its default definition will work just fine.
The Monoid
methods are rather unfortunately named; they are inspired
by the list instance of Monoid
, where indeed mempty = []
and mappend = (++)
, but this is misleading since many
monoids have little to do with appending (see these Comments from OCaml Hacker Brian Hurt on the haskellcafe mailing list).
Laws
Of course, every Monoid
instance should actually be a monoid in the
mathematical sense, which implies these laws:
mempty `mappend` x = x
x `mappend` mempty = x
(x `mappend` y) `mappend` z = x `mappend` (y `mappend` z)
Instances
There are quite a few interesting Monoid
instances defined in
Data.Monoid
.

[a]
is aMonoid
, withmempty = []
andmappend = (++)
.
It is not hard to check that
(x ++ y) ++ z = x ++ (y ++ z)
for any lists x
, y
, and z
, and
that the empty list is the identity:
[] ++ x = x ++ [] = x
.
 As noted previously, we can make a monoid out of any numeric
type under either addition or multiplication. However, since we
can't have two instances for the same type, Data.Monoid
provides
two newtype
wrappers, Sum
and Product
, with appropriate
Monoid
instances.
> getSum (mconcat . map Sum $ [1..5])
15
> getProduct (mconcat . map Product $ [1..5])
120
 This example code is silly, of course; we could just write
sum [1..5]
and product [1..5]
. Nevertheless, these instances
are useful in more generalized settings, as we will see in the
section Foldable
.

Any
andAll
arenewtype
wrappers providingMonoid
instances for Bool
(under disjunction and conjunction,
respectively).
 There are three instances for
Maybe
: a basic instance which
lifts a Monoid
instance for a
to an instance for Maybe a
, and
two newtype
wrappers First
and Last
for which mappend
selects the first (respectively last) nonNothing
item.

Endo a
is a newtype wrapper for functionsa > a
, which form
a monoid under composition.
 There are several ways to “lift”
Monoid
instances to
instances with additional structure. We have already seen that an
instance for a
can be lifted to an instance for Maybe a
. There
are also tuple instances: if a
and b
are instances of Monoid
,
then so is (a,b)
, using the monoid operations for a
and b
in
the obvious pairwise manner. Finally, if a
is a Monoid
, then so
is the function type e > a
for any e
; in particular,
g `mappend` h
is the function which applies both g
and h
to
its argument and then combines the result using the underlying
Monoid
instance for a
. This can be quite useful and
elegant (see example).
 The type
Ordering = LT  EQ  GT
is aMonoid
, defined in
such a way that
mconcat (zipWith compare xs ys)
computes the
lexicographic ordering of xs
and ys
. In particular,
mempty = EQ
, and mappend
evaluates to its leftmost nonEQ
argument (or EQ
if both arguments are EQ
). This can be used
together with the function instance of Monoid
to do some clever
things
(example).
 There are also
Monoid
instances for several standard data
structures in the containers library (haddock),
including Map
, Set
, and Sequence
.
Monoid
is also used to enable several other type class instances.
As noted previously, we can use Monoid
to make ((,) e)
an instance
of Applicative
:
instance Monoid e => Applicative ((,) e) where
pure x = (mempty, x)
(u, f) <*> (v, x) = (u `mappend` v, f x)
Monoid
can be similarly used to make ((,) e)
an instance of
Monad
as well; this is known as the writer monad. As we've
already seen, Writer
and WriterT
are a newtype wrapper and
transformer for this monad, respectively.
Monoid
also plays a key role in the Foldable
type class
(see section Foldable).
Other monoidal classes: Alternative, MonadPlus, ArrowPlus
The Alternative
type class (haddock)
is for Applicative
functors which also have
a monoid structure:
class Applicative f => Alternative f where
empty :: f a
(<>) :: f a > f a > f a
Of course, instances of Alternative
should satisfy the monoid laws.
Likewise, MonadPlus
(haddock)
is for Monad
s with a monoid structure:
class Monad m => MonadPlus m where
mzero :: m a
mplus :: m a > m a > m a
The MonadPlus
documentation states that it is intended to model
monads which also support “choice and failure”; in addition to the
monoid laws, instances of MonadPlus
are expected to satisfy
mzero >>= f = mzero
v >> mzero = mzero
which explains the sense in which mzero
denotes failure. Since
mzero
should be the identity for mplus
, the computation m1 `mplus` m2
succeeds (evaluates to something other than mzero
) if
either m1
or m2
does; so mplus
represents choice. The guard
function can also be used with instances of MonadPlus
; it requires a
condition to be satisfied and fails (using mzero
) if it is not. A
simple example of a MonadPlus
instance is []
, which is exactly the
same as the Monoid
instance for []
: the empty list represents
failure, and list concatenation represents choice. In general,
however, a MonadPlus
instance for a type need not be the same as its
Monoid
instance; Maybe
is an example of such a type. A great
introduction to the MonadPlus
type class, with interesting examples
of its use, is Doug Auclair's MonadPlus: What a Super Monad! in the Monad.Reader issue 11.
There used to be a type class called MonadZero
containing only
mzero
, representing monads with failure. The do
notation requires
some notion of failure to deal with failing pattern matches.
Unfortunately, MonadZero
was scrapped in favor of adding the fail
method to the Monad
class. If we are lucky, someday MonadZero
will
be restored, and fail
will be banished to the bit bucket where it
belongs (see MonadPlus reform proposal). The idea is that any
do
block which uses pattern matching (and hence may fail) would require
a MonadZero
constraint; otherwise, only a Monad
constraint would be
required.
Finally, ArrowZero
and ArrowPlus
(haddock)
represent Arrow
s (see below) with a
monoid structure:
class Arrow (~>) => ArrowZero (~>) where
zeroArrow :: b ~> c
class ArrowZero (~>) => ArrowPlus (~>) where
(<+>) :: (b ~> c) > (b ~> c) > (b ~> c)
Further reading
Monoids have gotten a fair bit of attention recently, ultimately due
to
a blog post by Brian Hurt, in which he
complained about the fact that the names of many Haskell type classes
(Monoid
in particular) are taken from abstract mathematics. This
resulted in a long haskellcafe thread
arguing the point and discussing monoids in general.
∗ May its name live forever.
However, this was quickly followed by several blog posts about
Monoid
∗. First, Dan Piponi
wrote a great introductory post, [http://blog.sigfpe.com/2009/01/haskellmonoidsandtheiruses.html Haskell Monoids and their
Uses]. This was quickly followed by
Heinrich Apfelmus's Monoids and Finger Trees, an accessible exposition of
Hinze and Paterson's classic paper on 23 finger trees, which makes very clever
use of Monoid
to implement an elegant and generic data structure.
Dan Piponi then wrote two fascinating articles about using Monoids
(and finger trees): Fast Incremental Regular Expressions and Beyond Regular Expressions
In a similar vein, David Place's article on improving Data.Map
in
order to compute incremental folds (see the Monad Reader issue 11)
is also a
good example of using Monoid
to generalize a data structure.
Some other interesting examples of Monoid
use include [http://www.reddit.com/r/programming/comments/7cf4r/monoids_in_my_programming_language/c06adnx building
elegant list sorting combinators],
collecting unstructured information,
and a brilliant series of posts by ChungChieh Shan and Dylan Thurston
using Monoid
s to [http://conway.rutgers.edu/~ccshan/wiki/blog/posts/WordNumbers1/ elegantly solve a difficult combinatorial
puzzle] (followed by
part 2,
part 3,
part 4).
As unlikely as it sounds, monads can actually be viewed as a sort of
monoid, with join
playing the role of the binary operation and
return
the role of the identity; see Dan Piponi's blog post.
Foldable
The Foldable
class, defined in the Data.Foldable
module (haddock), abstracts over containers which can be
“folded” into a summary value. This allows such folding operations
to be written in a containeragnostic way.
Definition
The definition of the Foldable
type class is:
class Foldable t where
fold :: Monoid m => t m > m
foldMap :: Monoid m => (a > m) > t a > m
foldr :: (a > b > b) > b > t a > b
foldl :: (a > b > a) > a > t b > a
foldr1 :: (a > a > a) > t a > a
foldl1 :: (a > a > a) > t a > a
This may look complicated, but in fact, to make a Foldable
instance
you only need to implement one method: your choice of foldMap
or
foldr
. All the other methods have default implementations in terms
of these, and are presumably included in the class in case more
efficient implementations can be provided.
Instances and examples
The type of foldMap
should make it clear what it is supposed to do:
given a way to convert the data in a container into a Monoid
(a
function a > m
) and a container of a
's (t a
), foldMap
provides a way to iterate over the entire contents of the container,
converting all the a
's to m
's and combining all the m
's with
mappend
. The following code shows two examples: a simple
implementation of foldMap
for lists, and a binary tree example
provided by the Foldable
documentation.
instance Foldable [] where
foldMap g = mconcat . map g
data Tree a = Empty  Leaf a  Node (Tree a) a (Tree a)
instance Foldable Tree where
foldMap f Empty = mempty
foldMap f (Leaf x) = f x
foldMap f (Node l k r) = foldMap f l ++ f k ++ foldMap f r
where (++) = mappend
The foldr
function has a type similar to the foldr
found in the Prelude
, but
more general, since the foldr
in the Prelude
works only on lists.
The Foldable
module also provides instances for Maybe
and Array
;
additionally, many of the data structures found in the standard containers library (for example, Map
, Set
, Tree
,
and Sequence
) provide their own Foldable
instances.
Derived folds
Given an instance of Foldable
, we can write generic,
containeragnostic functions such as:
 Compute the size of any container.
containerSize :: Foldable f => f a > Int
containerSize = getSum . foldMap (const (Sum 1))
 Compute a list of elements of a container satisfying a predicate.
filterF :: Foldable f => (a > Bool) > f a > [a]
filterF p = foldMap (\a > if p a then [a] else [])
 Get a list of all the Strings in a container which include the
 letter a.
aStrings :: Foldable f => f String > [String]
aStrings = filterF (elem 'a')
The Foldable
module also provides a large number of predefined
folds, many of which are generalized versions of Prelude
functions of the
same name that only work on lists: concat
, concatMap
, and
,
or
, any
, all
, sum
, product
, maximum
(By
),
minimum
(By
), elem
, notElem
, and find
. The reader may enjoy
coming up with elegant implementations of these functions using fold
or foldMap
and appropriate Monoid
instances.
There are also generic functions that work with Applicative
or
Monad
instances to generate some sort of computation from each
element in a container, and then perform all the side effects from
those computations, discarding the results: traverse_
, sequenceA_
,
and others. The results must be discarded because the Foldable
class is too weak to specify what to do with them: we cannot, in
general, make an arbitrary Applicative
or Monad
instance into a
Monoid
. If we do have an Applicative
or Monad
with a monoid
structure—that is, an Alternative
or a MonadPlus
—then we can
use the asum
or msum
functions, which can combine the results as
well. Consult the Foldable
documentation for
more details on any of these functions.
Note that the Foldable
operations always forget the structure of
the container being folded. If we start with a container of type t a
for some Foldable t
, then t
will never appear in the output
type of any operations defined in the Foldable
module. Many times
this is exactly what we want, but sometimes we would like to be able
to generically traverse a container while preserving its
structure—and this is exactly what the Traversable
class provides,
which will be discussed in the next section.
Further reading
The Foldable
class had its genesis in McBride and Paterson's paper
introducing Applicative
, although it has
been fleshed out quite a bit from the form in the paper.
An interesting use of Foldable
(as well as Traversable
) can be
found in Janis Voigtländer's paper Bidirectionalization for free!.
Traversable
Definition
The Traversable
type class, defined in the Data.Traversable
module (haddock), is:
class (Functor t, Foldable t) => Traversable t where
traverse :: Applicative f => (a > f b) > t a > f (t b)
sequenceA :: Applicative f => t (f a) > f (t a)
mapM :: Monad m => (a > m b) > t a > m (t b)
sequence :: Monad m => t (m a) > m (t a)
As you can see, every Traversable
is also a foldable functor. Like
Foldable
, there is a lot in this type class, but making instances is
actually rather easy: one need only implement traverse
or
sequenceA
; the other methods all have default implementations in
terms of these functions. A good exercise is to figure out what the default
implementations should be: given either traverse
or sequenceA
, how
would you define the other three methods? (Hint for mapM
:
Control.Applicative
exports the WrapMonad
newtype, which makes any
Monad
into an Applicative
. The sequence
function can be implemented in terms
of mapM
.)
Intuition
The key method of the Traversable
class, and the source of its
unique power, is sequenceA
. Consider its type:
sequenceA :: Applicative f => t (f a) > f (t a)
This answers the fundamental question: when can we commute two functors? For example, can we turn a tree of lists into a list of trees? (Answer: yes, in two ways. Figuring out what they are, and why, is left as an exercise. A much more challenging question is whether a list of trees can be turned into a tree of lists.)
The ability to compose two monads depends crucially on this ability to
commute functors. Intuitively, if we want to build a composed monad
M a = m (n a)
out of monads m
and n
, then to be able to
implement join :: M (M a) > M a
, that is,
join :: m (n (m (n a))) > m (n a)
, we have to be able to commute
the n
past the m
to get m (m (n (n a)))
, and then we can use the
join
s for m
and n
to produce something of type m (n a)
. See
Mark Jones's paper for more details.
Instances and examples
What's an example of a Traversable
instance?
The following code shows an example instance for the same
Tree
type used as an example in the previous Foldable
section. It
is instructive to compare this instance with a Functor
instance for
Tree
, which is also shown.
data Tree a = Empty  Leaf a  Node (Tree a) a (Tree a)
instance Traversable Tree where
traverse g Empty = pure Empty
traverse g (Leaf x) = Leaf <$> g x
traverse g (Node l x r) = Node <$> traverse g l
<*> g x
<*> traverse g r
instance Functor Tree where
fmap g Empty = Empty
fmap g (Leaf x) = Leaf $ g x
fmap g (Node l x r) = Node (fmap g l)
(g x)
(fmap g r)
It should be clear that the Traversable
and Functor
instances for
Tree
are almost identical; the only difference is that the Functor
instance involves normal function application, whereas the
applications in the Traversable
instance take place within an
Applicative
context, using (<$>)
and (<*>)
. In fact, this will
be
true for any type.
Any Traversable
functor is also Foldable
, and a Functor
. We can see
this not only from the class declaration, but by the fact that we can
implement the methods of both classes given only the Traversable
methods. A good exercise is to implement fmap
and foldMap
using
only the Traversable
methods; the implementations are surprisingly
elegant. The Traversable
module provides these
implementations as fmapDefault
and foldMapDefault
.
The standard libraries provide a number of Traversable
instances,
including instances for []
, Maybe
, Map
, Tree
, and Sequence
.
Notably, Set
is not Traversable
, although it is Foldable
.
Further reading
The Traversable
class also had its genesis in [http://www.soi.city.ac.uk/~ross/papers/Applicative.html McBride and Paterson's
Applicative
paper], and is described in
more detail in Gibbons and Oliveira, The Essence of the Iterator Pattern, which also contains a wealth of
references to related work.
Category
Category
is another fairly new addition to the Haskell standard
libraries; you may or may not have it installed depending on the
version of your base
package. It generalizes the notion of
function composition to general “morphisms.”
The definition of the Category
type class (from
Control.Category
—haddock) is shown below. For ease of reading, note that I have used an
infix type constructor (~>)
, much like the infix function type
constructor (>)
. This syntax is not part of Haskell 98.
The second definition shown is the one used in the standard libraries.
For the remainder of the article, I will use the infix type
constructor (~>)
for Category
as well as Arrow
.
class Category (~>) where
id :: a ~> a
(.) :: (b ~> c) > (a ~> b) > (a ~> c)
 The same thing, with a normal (prefix) type constructor
class Category cat where
id :: cat a a
(.) :: cat b c > cat a b > cat a c
Note that an instance of Category
should be a type constructor which
takes two type arguments, that is, something of kind * > * > *
. It
is instructive to imagine the type constructor variable cat
replaced
by the function constructor (>)
: indeed, in this case we recover
precisely the familiar identity function id
and function composition
operator (.)
defined in the standard Prelude
.
Of course, the Category
module provides exactly such an instance of
Category
for (>)
. But it also provides one other instance, shown
below, which should be familiar from the
previous discussion of the Monad
laws. Kleisli m a b
, as defined
in the Control.Arrow
module, is just a newtype
wrapper around a > m b
.
newtype Kleisli m a b = Kleisli { runKleisli :: a > m b }
instance Monad m => Category (Kleisli m) where
id = Kleisli return
Kleisli g . Kleisli h = Kleisli (h >=> g)
The only law that Category
instances should satisfy is that id
and
(.)
should form a monoid—that is, id
should be the identity of
(.)
, and (.)
should be associative.
Finally, the Category
module exports two additional operators:
(<<<)
, which is just a synonym for (.)
, and (>>>)
, which is
(.)
with its arguments reversed. (In previous versions of the
libraries, these operators were defined as part of the Arrow
class.)
Further reading
The name Category
is a bit misleading, since the Category
class
cannot represent arbitrary categories, but only categories whose
objects are objects of Hask
, the category of Haskell types. For a
more general treatment of categories within Haskell, see the
categoryextras package. For more about
category theory in general, see the excellent Haskell wikibook page,
Steve Awodey's new book,
Benjamin Pierce's
Basic category theory for computer scientists, or
Barr and Wells's category theory lecture notes. Benjamin Russell's blog post
is another good source of motivation and
category theory links. You certainly don't need to know any category
theory to be a successful and productive Haskell programmer, but it
does lend itself to much deeper appreciation of Haskell's underlying
theory.
Arrow
The Arrow
class represents another abstraction of computation, in a
similar vein to Monad
and Applicative
. However, unlike Monad
and Applicative
, whose types only reflect their output, the type of
an Arrow
computation reflects both its input and output. Arrows
generalize functions: if (~>)
is an instance of Arrow
, a value of
type b ~> c
can be thought of as a computation which takes values of
type b
as input, and produces values of type c
as output. In the
(>)
instance of Arrow
this is just a pure function; in general, however,
an arrow may represent some sort of “effectful” computation.
Definition
The definition of the Arrow
type class, from
Control.Arrow
(haddock), is:
class Category (~>) => Arrow (~>) where
arr :: (b > c) > (b ~> c)
first :: (b ~> c) > ((b, d) ~> (c, d))
second :: (b ~> c) > ((d, b) ~> (d, c))
(***) :: (b ~> c) > (b' ~> c') > ((b, b') ~> (c, c'))
(&&&) :: (b ~> c) > (b ~> c') > (b ~> (c, c'))
∗ In versions of the base
package prior to version 4, there is no Category
class, and the
Arrow
class includes the arrow composition operator (>>>)
. It
also includes pure
as a synonym for arr
, but this was removed
since it conflicts with the pure
from Applicative
.
The first thing to note is the Category
class constraint, which
means that we get identity arrows and arrow composition for free:
given two arrows g :: b ~> c
and h :: c ~> d
, we can form their
composition g >>> h :: b ~> d
∗.
As should be a familiar pattern by now, the only methods which must be
defined when writing a new instance of Arrow
are arr
and first
;
the other methods have default definitions in terms of these, but are
included in the Arrow
class so that they can be overridden with more
efficient implementations if desired.
Intuition
Let's look at each of the arrow methods in turn. Ross Paterson's web page on arrows has nice diagrams which can help build intuition.
 The
arr
function takes any functionb > c
and turns it into a
generalized arrow b ~> c
. The arr
method justifies the claim
that arrows generalize functions, since it says that we can treat
any function as an arrow. It is intended that the arrow arr g
is
“pure” in the sense that it only computes g
and has no
“effects” (whatever that might mean for any particular arrow type).
 The
first
method turns any arrow fromb
toc
into an arrow
from (b,d)
to (c,d)
. The idea is that first g
uses g
to
process the first element of a tuple, and lets the second element pass
through unchanged. For the function instance of Arrow
, of course,
first g (x,y) = (g x, y)
.
 The
second
function is similar tofirst
, but with the elements of the
tuples swapped. Indeed, it can be defined in terms of first
using
an auxiliary function swap
, defined by swap (x,y) = (y,x)
.
 The
(***)
operator is “parallel composition” of arrows: it takes two
arrows and makes them into one arrow on tuples, which has the
behavior of the first arrow on the first element of a tuple, and the
behavior of the second arrow on the second element. The mnemonic
is that g *** h
is the product (hence *
) of g
and
h
. For the function instance of Arrow
,
we define (g *** h) (x,y) = (g x, h y)
. The default implementation of
(***)
is in terms of first
, second
, and sequential arrow
composition (>>>)
. The reader may also wish to think about how to
implement first
and second
in terms of (***)
.
 The
(&&&)
operator is “fanout composition” of arrows: it takes two arrows
g
and h
and makes them into a new arrow g &&& h
which supplies
its input as the input to both g
and h
, returning their results
as a tuple. The mnemonic is that g &&& h
performs both g
and h
(hence &
) on its input. For functions, we define (g &&& h) x = (g x, h x)
.
Instances
The Arrow
library itself only provides two Arrow
instances, both
of which we have already seen: (>)
, the normal function
constructor, and Kleisli m
, which makes functions of
type a > m b
into Arrow
s for any Monad m
. These instances are:
instance Arrow (>) where
arr g = g
first g (x,y) = (g x, y)
newtype Kleisli m a b = Kleisli { runKleisli :: a > m b }
instance Monad m => Arrow (Kleisli m) where
arr f = Kleisli (return . f)
first (Kleisli f) = Kleisli (\ ~(b,d) > do c < f b
return (c,d) )
Laws
∗ See John Hughes: Generalising monads to arrows; Sam Lindley, Philip Wadler, Jeremy Yallop: The arrow calculus; Ross Paterson: Programming with Arrows.
There are quite a few laws that instances of Arrow
should
satisfy ∗:
arr id = id
arr (h . g) = arr g >>> arr h
first (arr g) = arr (g *** id)
first (g >>> h) = first g >>> first h
first g >>> arr (id *** h) = arr (id *** h) >>> first g
first g >>> arr fst = arr fst >>> g
first (first g) >>> arr assoc = arr assoc >>> first g
assoc ((x,y),z) = (x,(y,z))
Note that this version of the laws is slightly different than the laws given in the
first two above references, since several of the laws have now been
subsumed by the Category
laws (in particular, the requirements that
id
is the identity arrow and that (>>>)
is associative). The laws
shown here follow those in Paterson's Programming with Arrows, which uses the
Category
class.
∗ Unless categorytheoryinduced insomnolence is your cup of tea.
The reader is advised not to lose too much sleep over the Arrow
laws ∗, since it is not essential to understand them in order to
program with arrows. There are also laws that ArrowChoice
,
ArrowApply
, and ArrowLoop
instances should satisfy; the interested
reader should consult Paterson: Programming with Arrows.
ArrowChoice
Computations built using the Arrow
class, like those built using
the Applicative
class, are rather inflexible: the structure of the computation
is fixed at the outset, and there is no ability to choose between
alternate execution paths based on intermediate results.
The ArrowChoice
class provides exactly such an ability:
class Arrow (~>) => ArrowChoice (~>) where
left :: (b ~> c) > (Either b d ~> Either c d)
right :: (b ~> c) > (Either d b ~> Either d c)
(+++) :: (b ~> c) > (b' ~> c') > (Either b b' ~> Either c c')
() :: (b ~> d) > (c ~> d) > (Either b c ~> d)
A comparison of ArrowChoice
to Arrow
will reveal a striking
parallel between left
, right
, (+++)
, ()
and first
,
second
, (***)
, (&&&)
, respectively. Indeed, they are dual:
first
, second
, (***)
, and (&&&)
all operate on product types
(tuples), and left
, right
, (+++)
, and ()
are the
corresponding operations on sum types. In general, these operations
create arrows whose inputs are tagged with Left
or Right
, and can
choose how to act based on these tags.
 If
g
is an arrow fromb
toc
, thenleft g
is an arrow
from Either b d
to Either c d
. On inputs tagged with Left
,
the left g
arrow has the behavior of g
; on inputs tagged with Right
, it
behaves as the identity.
 The
right
function, of course, is the mirror image ofleft
. The arrowright g
has the behavior of g
on inputs tagged with Right
.
 The
(+++)
operator performs “multiplexing”:g +++ h
behaves asg
on inputs tagged with Left
, and as h
on inputs tagged with
Right
. The tags are preserved. The (+++)
operator is the sum (hence
+
) of two arrows, just as (***)
is the product.
 The
()
operator is “merge” or “fanin”: the arrowg  h
behaves as g
on inputs tagged with Left
, and h
on inputs
tagged with Right
, but the tags are discarded (hence, g
and h
must have the same output type). The mnemonic is that g  h
performs either g
or h
on its input.
The ArrowChoice
class allows computations to choose among a finite number of
execution paths, based on intermediate results. The possible
execution paths must be known in advance, and explicitly assembled
with (+++)
or ()
. However, sometimes more flexibility is
needed: we would like to be able to compute an arrow from
intermediate results, and use this computed arrow to continue the
computation. This is the power given to us by ArrowApply
.
ArrowApply
The ArrowApply
type class is:
class Arrow (~>) => ArrowApply (~>) where
app :: (b ~> c, b) ~> c
If we have computed an arrow as the output of some previous
computation, then app
allows us to apply that arrow to an input,
producing its output as the output of app
. As an exercise, the
reader may wish to use app
to implement an alternative “curried”
version, app2 :: b ~> ((b ~> c) ~> c)
.
This notion of being able to compute a new computation
may sound familiar:
this is exactly what the monadic bind operator (>>=)
does. It
should not particularly come as a surprise that ArrowApply
and
Monad
are exactly equivalent in expressive power. In particular,
Kleisli m
can be made an instance of ArrowApply
, and any instance
of ArrowApply
can be made a Monad
(via the newtype
wrapper
ArrowMonad
). As an exercise, the reader may wish to try
implementing these instances:
instance Monad m => ArrowApply (Kleisli m) where
app =  exercise
newtype ArrowApply a => ArrowMonad a b = ArrowMonad (a () b)
instance ArrowApply a => Monad (ArrowMonad a) where
return =  exercise
(ArrowMonad a) >>= k =  exercise
ArrowLoop
The ArrowLoop
type class is:
class Arrow a => ArrowLoop a where
loop :: a (b, d) (c, d) > a b c
trace :: ((b,d) > (c,d)) > b > c
trace f b = let (c,d) = f (b,d) in c
It describes arrows that can use recursion to compute results, and is
used to desugar the rec
construct in arrow notation (described
below).
Taken by itself, the type of the loop
method does not seem to tell
us much. Its intention, however, is a generalization of the trace
function which is also shown. The d
component of the first arrow's
output is fed back in as its own input. In other words, the arrow
loop g
is obtained by recursively “fixing” the second component of
the input to g
.
It can be a bit difficult to grok what the trace
function is doing.
How can d
appear on the left and right sides of the let
? Well,
this is Haskell's laziness at work. There is not space here for a
full explanation; the interested reader is encouraged to study the
standard fix
function, and to read Paterson's arrow tutorial.
Arrow notation
Programming directly with the arrow combinators can be painful,
especially when writing complex computations which need to retain
simultaneous reference to a number of intermediate results. With
nothing but the arrow combinators, such intermediate results must be
kept in nested tuples, and it is up to the programmer to remember
which intermediate results are in which components, and to swap,
reassociate, and generally mangle tuples as necessary. This problem
is solved by the special arrow notation supported by GHC, similar to
do
notation for monads, that allows names to be assigned to
intermediate results while building up arrow computations. An example
arrow implemented using arrow notation, taken from
Paterson, is:
class ArrowLoop (~>) => ArrowCircuit (~>) where
delay :: b > (b ~> b)
counter :: ArrowCircuit (~>) => Bool ~> Int
counter = proc reset > do
rec output < idA < if reset then 0 else next
next < delay 0 < output + 1
idA < output
This arrow is intended to represent a recursively defined counter circuit with a reset line.
There is not space here for a full explanation of arrow notation; the interested reader should consult [http://www.soi.city.ac.uk/~ross/papers/notation.html Paterson's paper introducing the notation], or his later [http://www.soi.city.ac.uk/~ross/papers/fop.html tutorial which presents a simplified version].
Further reading
An excellent starting place for the student of arrows is the arrows web page, which contains an introduction and many references. Some key papers on arrows include Hughes's original paper introducing arrows, Generalising monads to arrows, and Paterson's paper on arrow notation.
Both Hughes and Paterson later wrote accessible tutorials intended for a broader audience: Paterson: Programming with Arrows and Hughes: Programming with Arrows.
Although Hughes's goal in defining the Arrow
class was to
generalize Monad
s, and it has been said that Arrow
lies “between
Applicative
and Monad
” in power, they are not directly
comparable. The precise relationship remained in some confusion until
analyzed by Lindley, Wadler, and Yallop, who
also invented a new calculus of arrows, based on the lambda calculus,
which considerably simplifies the presentation of the arrow laws
(see The arrow calculus).
Some examples of Arrow
s include Yampa, the
Haskell XML Toolkit, and the functional GUI library Grapefruit.
Some extensions to arrows have been explored; for example, the
BiArrow
s of Alimarine et al., for twoway instead of oneway
computation.
The Haskell wiki has links to many additional research papers relating to Arrow
s.
Comonad
The final type class we will examine is Comonad
. The Comonad
class
is the categorical dual of Monad
; that is, Comonad
is like Monad
but with all the function arrows flipped. It is not actually in the
standard Haskell libraries, but it has seen some interesting uses
recently, so we include it here for completeness.
Definition
The Comonad
type class, defined in the Control.Comonad
module of
the categoryextras library, is:
class Functor f => Copointed f where
extract :: f a > a
class Copointed w => Comonad w where
duplicate :: w a > w (w a)
extend :: (w a > b) > w a > w b
As you can see, extract
is the dual of return
, duplicate
is the
dual of join
, and extend
is the dual of (>>=)
(although its
arguments are in a different order). The definition
of Comonad
is a bit redundant (after all, the Monad
class does not
need join
), but this is so that a Comonad
can be defined by fmap
,
extract
, and either duplicate
or extend
. Each has a
default implementation in terms of the other.
A prototypical example of a Comonad
instance is:
 Infinite lazy streams
data Stream a = Cons a (Stream a)
instance Functor Stream where
fmap g (Cons x xs) = Cons (g x) (fmap g xs)
instance Copointed Stream where
extract (Cons x _) = x
 'duplicate' is like the list function 'tails'
 'extend' computes a new Stream from an old, where the element
 at position n is computed as a function of everything from
 position n onwards in the old Stream
instance Comonad Stream where
duplicate s@(Cons x xs) = Cons s (duplicate xs)
extend g s@(Cons x xs) = Cons (g s) (extend g xs)
 = fmap g (duplicate s)
Further reading
Dan Piponi explains in a blog post what [http://blog.sigfpe.com/2006/12/evaluatingcellularautomatais.html cellular automata have to do with comonads]. In another blog post, Conal Elliott has examined [http://conal.net/blog/posts/functionalinteractivebehavior/ a comonadic formulation of functional reactive programming]. Sterling Clover's blog post Comonads in everyday life explains the relationship between comonads and zippers, and how comonads can be used to design a menu system for a web site.
Uustalu and Vene have a number of papers exploring ideas related to comonads and functional programming:
 Comonadic Notions of Computation
 The dual of substitution is redecoration
 Recursive coalgebras from comonads
 Recursion schemes from comonads
 The Essence of Dataflow Programming.
Acknowledgements
A special thanks to all of those who taught me about standard Haskell type classes and helped me develop good intuition for them, particularly Jules Bean (quicksilver), Derek Elkins (ddarius), Conal Elliott (conal), Cale Gibbard (Cale), David House, Dan Piponi (sigfpe), and Kevin Reid (kpreid).
I also thank the many people who provided a mountain of helpful feedback and suggestions on a first draft of this article: David Amos, Kevin Ballard, Reid Barton, Doug Beardsley, Joachim Breitner, Andrew Cave, David Christiansen, Gregory Collins, Mark Jason Dominus, Conal Elliott, Yitz Gale, George Giorgidze, Steven Grady, Travis Hartwell, Steve Hicks, Philip Hölzenspies, Edward Kmett, Eric Kow, Serge Le Huitouze, Felipe Lessa, Stefan Ljungstrand, Eric Macaulay, Rob MacAulay, Simon Meier, Eric Mertens, Tim Newsham, Russell O'Connor, Conrad Parker, Walt RorieBaety, Colin Ross, Tom Schrijvers, Aditya Siram, C. Smith, Martijn van Steenbergen, Joe Thornber, Jared Updike, Rob Vollmert, Andrew Wagner, Louis Wasserman, and Ashley Yakeley, as well as a few only known to me by their IRC nicks: b_jonas, maltem, tehgeekmeister, and ziman. I have undoubtedly omitted a few inadvertently, which in no way diminishes my gratitude.
Finally, I would like to thank Wouter Swierstra for his fantastic work editing the Monad.Reader, and my wife Joyia for her patience during the process of writing the Typeclassopedia.
About the author
Brent Yorgey (blog, homepage) is a firstyear Ph.D. student in the programming languages group at the University of Pennsylvania]. He enjoys teaching, creating EDSLs, playing Bach fugues, musing upon category theory, and cooking tasty lambdatreats for the denizens of #haskell.
Colophon
The Typeclassopedia was written by Brent Yorgey and initally published in March 2009. Painstakingly converted to wiki syntax by User:Geheimdienst in November 2011, after asking Brent's permission. If something like this ever needs to be done again, here are some vim commands that helped:
 %s/\\section{\([^}]*\)}/=\1=/gc
 %s/\\subsection{\([^}]*\)}/==\1==/gc
 %s/^ *\\item /\r* /gc
 %s//—/gc
 %s/\$\([^$]*\)\$/<math>\1<\/math>/gc
 %s/\([^]*\)/<code>\1<\/code>/gc
 %s/\\dots/.../gc
 %s/^\\label{.*$//gc
 %s/\\emph{\([^}]*\)}/''\1''/gc
 %s/\\term{\([^}]*\)}/''\1''/gc
To get all the citations into the main text, I first tried processing the source with Tex or Lyx. This didn't work due to missing unfindable packages, syntax errors, and my general ineptitude with Tex.
I then went for the next best solution, which seemed to be extracting all instances of “\cite{something}” from the source and in that order pulling the referenced entries from the .bib file. This way you can go through the source file and sortedreferences file in parallel, copying over what you need, without searching back and forth in the .bib file. I used:
 egrep o "\cite\{[^\}]*\}" ~/typeclassopedia.lhs  cut c 6  tr "," "\n"  tr d "}" > /tmp/citations
 for i in $(cat /tmp/citations); do grep A99 "$i" ~/typeclassopedia.bibegrep B99 '^\}$' m1 ; done > ~/typeclassorefssorted