Difference between revisions of "IO inside"

From HaskellWiki
Jump to navigation Jump to search
m (grammar)
m (fixed order of arguments of hSeek function)
(96 intermediate revisions by 22 users not shown)
Line 1: Line 1:
Haskell I/O has always been a source of confusion and surprises for new Haskellers. While simple I/O code in Haskell looks very similar to its equivalents in imperative languages, attempts to write somewhat more complex code often end with a total mess. This is because Haskell I/O is really very different internally. Haskell is a pure language and even the I/O system can't break this purity.
+
Haskell I/O has always been a source of confusion and surprises for new Haskellers. While simple I/O code in Haskell looks very similar to its equivalents in imperative languages, attempts to write somewhat more complex code often result in a total mess. This is because Haskell I/O is really very different internally. Haskell is a pure language and even the I/O system can't break this purity.
   
The following text is an attempt to explain details of Haskell I/O implementations. This explanation should help you eventually master all the smart I/O tricks. Moreover, I added detailed explanation of various traps you might encounter on this way. After reading this text, you will receive a "Master of Haskell I/O" degree that is equal to Bachelor in Computer Science and Mathematics, simultaneously :)
+
The following text is an attempt to explain the details of Haskell I/O implementations. This explanation should help you eventually master all the smart I/O tricks. Moreover, I've added a detailed explanation of various traps you might encounter along the way. After reading this text, you will receive a "Master of Haskell I/O" degree that is equal to a Bachelor in Computer Science and Mathematics, simultaneously.
   
 
If you are new to Haskell I/O you may prefer to start by reading the [[Introduction to IO]] page.
 
If you are new to Haskell I/O you may prefer to start by reading the [[Introduction to IO]] page.
   
   
== Haskell is pure language ==
+
== Haskell is a pure language ==
   
Haskell is pure language, which means that result of any function call is fully determined by its arguments. Pseudo-functions like rand() or getchar() in C which return different results on each call, are just impossible to write in Haskell. Moreover, Haskell functions can't have side effects; they cannot effect any changes in the "real world", like changing files, writing to the screen, printing, sending data over the network, and so on. These two restrictions together mean that any function
+
Haskell is a pure language, which means that the result of any function call is fully determined by its arguments. Pseudo-functions like rand() or getchar() in C, which return different results on each call, are simply impossible to write in Haskell. Moreover, Haskell functions can't have side effects, which means that they can't effect any changes to the "real world", like changing files, writing to the screen, printing, sending data over the network, and so on. These two restrictions together mean that any function call can be replaced by the result of a previous call with the same parameters, and the language '''guarantees''' that all these rearrangements will not change the program result!
call can be omitted, repeated, or replaced by the result of a previous call with the same parameters, and the language '''guarantees''' that all these rearrangements will not change program result!
 
   
Let's compare this to C: optimizing C compilers try to guess which functions don't have side effects or depend on some global mutable variable. If this guess is wrong, an optimization can change program semantics. To avoid this disaster, C optimizers are conservative in their guesses or require hints from the programmer about the purity of variables and functions.
+
Let's compare this to C: optimizing C compilers try to guess which functions have no side effects and don't depend on mutable global variables. If this guess is wrong, an optimization can change the program's semantics! To avoid this kind of disaster, C optimizers are conservative in their guesses or require hints from the programmer about the purity of functions.
 
Comparing to them, Haskell compiler is a set of pure mathematical
 
transformations that can't be wrong by definition - they just
 
translate one abstract data processing algorithm (i.e. some complex
 
function) to another equivalent algorithm, just with better
 
performance. This results in much better high-level optimization
 
facilities comparing to C compilers
 
 
But this purity creates it's own problems. How we can do I/O, work
 
with stateful algorithms and side effects in pure language? This
 
question had many different solutions probed in 18 years of Haskell
 
existence and finally one based on using monads was widely accepted
 
   
  +
Compared to an optimizing C compiler, a Haskell compiler is a set of pure mathematical transformations. This results in much better high-level optimization facilities. Moreover, pure mathematical computations can be much more easily divided into several threads that may be executed in parallel, which is increasingly important in these days of multi-core CPUs. Finally, pure computations are less error-prone and easier to verify, which adds to Haskell's robustness and to the speed of program development using Haskell.
   
  +
Haskell purity allows compiler to call only functions whose results
  +
are really required to calculate final value of high-level function
  +
(i.e., main) - this is called lazy evaluation. It's great thing for
  +
pure mathematical computations, but how about I/O actions? Function
  +
like:
  +
<haskell>
  +
putStrLn "Press any key to begin formatting"
  +
</haskell>
  +
can't return any
  +
meaningful result value, so how can we ensure that compiler will not
  +
omit or reorder its execution? And in general: how we can work with
  +
stateful algorithms and side effects in an entirely lazy language?
  +
This question has had many different solutions proposed in 18 years of
  +
Haskell development (see [[History of Haskell]]), though a solution based on [[monad]]s is now
  +
the standard.
   
== What is the monad? ==
+
== What is a monad? ==
   
What is the monad? It's something from mathematical category theory, i
+
What is a [[monad]]? It's something from mathematical category theory, which I
don't know anymore :) In order to understand how monads are used to
+
don't know anymore. In order to understand how monads are used to
solve problem of I/O and side effects, you don't need to know it. It's
+
solve the problem of I/O and side effects, you don't need to know it. It's
enough to just know elementary mathematics, like I do :)
+
enough to just know elementary mathematics, like I do.
   
Let's imagine that we want to implement in Haskell well-known
+
Let's imagine that we want to implement in Haskell the well-known
'getchar' function. What the type it should have? Let's try:
+
'getchar' function. What type should it have? Let's try:
   
 
<haskell>
 
<haskell>
Line 43: Line 46:
 
</haskell>
 
</haskell>
   
What we will got with 'getchar' having just 'Char' type? You can see
+
What will we get with 'getchar' having just the 'Char' type? You can see
all the possible problems in 'get2chars' definition:
+
all the possible problems in the definition of 'get2chars':
   
# because Haskell compiler treats all functions as pure and not having side effects, it can avoid "excessive" call to 'getchar' and use one returned value two times
+
# Because the Haskell compiler treats all functions as pure (not having side effects), it can avoid "excessive" calls to 'getchar' and use one returned value twice.
# even if it will make two calls, there is no any clue to determine which call should be performed first. Do you want to return chars in the order they read, or in opposite order? Nothing in 'get2chars' definition answers this question.
+
# Even if it does make two calls, there is no way to determine which call should be performed first. Do you want to return the two chars in the order in which they were read, or in the opposite order? Nothing in the definition of 'get2chars' answers this question.
   
  +
How can these problems be solved, from the programmer's viewpoint?
 
  +
Let's introduce a fake parameter of 'getchar' to make each call
How these problems can be solved, from plain programmer's viewpoint?
 
Let's introduce fake parameter of 'getchar' to make each call
 
 
"different" from the compiler's point of view:
 
"different" from the compiler's point of view:
   
Line 60: Line 62:
 
</haskell>
 
</haskell>
   
This right away solved the first problem mentioned above - now
+
Right away, this solves the first problem mentioned above - now the
 
compiler will make two calls because it sees them as having different
 
compiler will make two calls because it sees them as having different
parameters. The whole 'get2chars' function should also had such
+
parameters. The whole 'get2chars' function should also have a
 
fake parameter, otherwise we will have the same problem calling it:
 
fake parameter, otherwise we will have the same problem calling it:
   
Line 73: Line 75:
   
   
Now, we need to give the compiler some clue to determine which function it
+
Now we need to give the compiler some clue to determine which function it
 
should call first. The Haskell language doesn't provide any way to express
 
should call first. The Haskell language doesn't provide any way to express
order of evaluation... except for data dependencies! How about adding
+
order of evaluation... except for data dependencies! How about adding an
artificial data dependency which prevents evaluation of second
+
artificial data dependency which prevents evaluation of the second
 
'getchar' before the first one? In order to achieve this, we will
 
'getchar' before the first one? In order to achieve this, we will
return from 'getchar' additional fake result that will be used as
+
return an additional fake result from 'getchar' that will be used as a
parameter for next 'getchar' call:
+
parameter for the next 'getchar' call:
   
 
<haskell>
 
<haskell>
Line 88: Line 90:
 
</haskell>
 
</haskell>
   
So bad so good - now we can guarantee that 'a' is read before 'b'
+
So far so good - now we can guarantee that 'a' is read before 'b'
because 'b' reading need value (i) that is returned by 'a' reading!
+
because reading 'b' needs the value ('i') that is returned by reading 'a'!
   
We've added fake parameter to 'get2chars' but the problem is that the
+
We've added a fake parameter to 'get2chars' but the problem is that the
 
Haskell compiler is too smart! It can believe that the external 'getchar'
 
Haskell compiler is too smart! It can believe that the external 'getchar'
function is really dependent on it's parameter but for 'get2chars' it
+
function is really dependent on its parameter but for 'get2chars' it
will see that we just cheating and throw it away! Problem? How about
+
will see that we're just cheating because we throw it away! Therefore it won't feel obliged to execute the calls in the order we want. How can we fix this? How about passing this fake parameter to the 'getchar' function?! In this case
  +
the compiler can't guess that it is really unused.
passing this fake parameter to 'getchar' function?! In this case
 
compiler can't guess that it is really unused :)
 
   
 
<haskell>
 
<haskell>
Line 104: Line 105:
   
   
And more - 'get2chars' has all the same purity problems as 'getchar'
+
And more - 'get2chars' has all the same purity problems as the 'getchar'
function. If one need to call it two times, he need a way to describe
+
function. If you need to call it two times, you need a way to describe
order of these calls. Look at:
+
the order of these calls. Look at:
   
 
<haskell>
 
<haskell>
get4chars = [get2chars 1, get2chars 2] -- order of `get2chars` calls isn't defined
+
get4chars = [get2chars 1, get2chars 2] -- order of 'get2chars' calls isn't defined
 
</haskell>
 
</haskell>
   
We already know how to fight with such problem - 'get2chars' should
+
We already know how to deal with these problems - 'get2chars' should
 
also return some fake value that can be used to order calls:
 
also return some fake value that can be used to order calls:
   
Line 123: Line 124:
   
   
But what the fake value it would return? If we will use some integer
+
But what's the fake value 'get2chars' should return? If we use some integer constant, the excessively-smart Haskell compiler will guess that we're cheating again. What about returning the value returned by 'getchar'? See:
constant, too smart Haskell compiler will guess we are cheating, again :)
 
What about returning the value returned by 'getchar'? See:
 
   
 
<haskell>
 
<haskell>
Line 133: Line 132:
 
</haskell>
 
</haskell>
   
Believe you or not, but we just constructed the whole "monadic"
+
Believe it or not, but we've just constructed the whole "monadic"
 
Haskell I/O system.
 
Haskell I/O system.
   
  +
== Welcome to the RealWorld, baby ==
   
  +
Warning: The following story about IO is incorrect in that it cannot actually explain some important aspects of IO (including interaction and concurrency). However, some people find it useful to begin developing an understanding.
 
== Welcome to RealWorld, baby :) ==
 
   
 
The 'main' Haskell function has the type:
 
The 'main' Haskell function has the type:
Line 146: Line 145:
 
</haskell>
 
</haskell>
   
where 'RealWorld' is faking type used instead of our Int. It is something
+
where 'RealWorld' is a fake type used instead of our Int. It's something
like baton passed in relay-race. When 'main' calls some IO function,
+
like the baton passed in a relay race. When 'main' calls some IO function,
it pass the "RealWorld" it received as parameter. All IO functions have
+
it passes the "RealWorld" it received as a parameter. All IO functions have
similar types involving RealWorld as parameter and result. To be
+
similar types involving RealWorld as a parameter and result. To be
 
exact, "IO" is a type synonym defined in the following way:
 
exact, "IO" is a type synonym defined in the following way:
   
Line 157: Line 156:
   
 
So, 'main' just has type "IO ()", 'getChar' has type "IO Char" and so
 
So, 'main' just has type "IO ()", 'getChar' has type "IO Char" and so
on. Let's look at 'main' calling 'getChar' two times:
+
on. You can think of the type "IO Char" as meaning "take the current RealWorld, do something to it, and return a Char and a (possibly changed) RealWorld". Let's look at 'main' calling 'getChar' two times:
   
 
<haskell>
 
<haskell>
Line 169: Line 168:
   
   
Look at this closely: 'main' passes to first 'getChar' the "world" it
+
Look at this closely: 'main' passes the "world" it received to the first 'getChar'. This 'getChar' returns some new value of type RealWorld
  +
that gets used in the next call. Finally, 'main' returns the "world" it got
received. This 'getChar' returns some new value of type RealWorld,
 
  +
from the second 'getChar'.
that is used in next call. Finally, 'main' returns the "world" it got
 
from the second 'getChar':
 
   
# Is it possible here to omit any call of 'getChar' if the char it read is not used? No, because we should return the "world" that is result of second 'getChar' and in turn requires "world" from first 'getChar'.
+
# Is it possible here to omit any call of 'getChar' if the Char it read is not used? No, because we need to return the "world" that is the result of the second 'getChar' and this in turn requires the "world" returned from the first 'getChar'.
# Is it possible to reorder 'getChar' calls? No, second 'getChar' can't be called before first one because it uses "world" it returns.
+
# Is it possible to reorder the 'getChar' calls? No: the second 'getChar' can't be called before the first one because it uses the "world" returned from the first call.
# Is it possible to duplicate calls? In Haskell semantics - yes, but real compilers never duplicate work in such simple cases (otherwise, the programs generated will not have any speed guarantees)
+
# Is it possible to duplicate calls? In Haskell semantics - yes, but real compilers never duplicate work in such simple cases (otherwise, the programs generated will not have any speed guarantees).
   
   
As we already said, RealWorld values are used like baton, passing them
+
As we already said, RealWorld values are used like a baton which gets passed
 
between all routines called by 'main' in strict order. Inside each
 
between all routines called by 'main' in strict order. Inside each
routine called, RealWorld values used in the same way. In whole, in
+
routine called, RealWorld values are used in the same way. Overall, in
order to "compute" world to be returned from 'main', we should perform
+
order to "compute" the world to be returned from 'main', we should perform
 
each IO procedure that is called from 'main', directly or indirectly.
 
each IO procedure that is called from 'main', directly or indirectly.
 
This means that each procedure inserted in the chain will be performed
 
This means that each procedure inserted in the chain will be performed
just at the moment (relative to other IO actions) when we planned it
+
just at the moment (relative to the other IO actions) when we intended it
to be called. Let consider the following program:
+
to be called. Let's consider the following program:
   
 
<haskell>
 
<haskell>
Line 197: Line 195:
 
</haskell>
 
</haskell>
   
Now you have enough knowledge to rewrite it in low-level way and
+
Now you have enough knowledge to rewrite it in a low-level way and
check that each operation what should be performed, will be really
+
check that each operation that should be performed will really be
performed with arguments it should have and in order we expecting
+
performed with the arguments it should have and in the order we expect.
   
   
Line 213: Line 211:
 
</haskell>
 
</haskell>
   
As you can see, we can easily include or exclude from execution chain
+
As you can see, we can easily include or exclude from the execution chain
 
IO procedures (actions) depending on the data values. If 'condition'
 
IO procedures (actions) depending on the data values. If 'condition'
will be False on call of 'when', 'action' will never be called because
+
will be False on the call of 'when', 'action' will never be called because
real Haskell compilers, again, never calls functions whose results
+
real Haskell compilers, again, never call functions whose results
don't required to calculate final result (i.e., here, final "world" value
+
are not required to calculate the final result (''i.e.'', here, the final "world" value of 'main').
of 'main')
 
   
Loops and any more complex control structures can be implemented in
+
Loops and more complex control structures can be implemented in
 
the same way. Try it as an exercise!
 
the same way. Try it as an exercise!
   
   
Finally you may want to know how much costs this passing of RealWorld
+
Finally, you may want to know how much passing these RealWorld
values all around. It's free! These fake values exist for compiler
+
values around the program costs. It's free! These fake values exist solely for the compiler while it analyzes and optimizes the code, but when it gets to assembly code generation, it "suddenly" realize that this type is like "()", so
  +
all these parameters and result values can be omitted from the final generated code. Isn't it beautiful?
only while it analyze and optimize code, but when it goes to assembler
 
code generation, it "suddenly" realize that this type is like "()", so
 
all these parameters and result values can be omitted from generated code.
 
Is it not really beautiful? :)
 
 
 
   
 
== '>>=' and 'do' notation ==
 
== '>>=' and 'do' notation ==
   
All beginners (including me :) start by thinking that 'do' is some
+
All beginners (including me) start by thinking that 'do' is some
magic statement that executes IO actions. It's wrong - 'do' is just a
+
magic statement that executes IO actions. That's wrong - 'do' is just
  +
syntactic sugar that simplifies the writing of procedures that use IO (and also other monads, but that's beyond the scope of this tutorial). 'do' notation eventually gets translated to statements passing "world" values around like we've manually written above and is used to simplify the gluing of several
syntax sugar that simplifies writing of IO procedures. 'do' notation
 
  +
IO actions together. You don't need to use 'do' for just one statement; for instance,
is finally translated to the statements passing "world" values like
 
we manually written above and need only to simplify gluing of several
 
IO actions together. You don't require to use 'do' for just one statement:
 
   
 
<haskell>
 
<haskell>
Line 246: Line 236:
 
</haskell>
 
</haskell>
 
 
is desugared just to:
+
is desugared to:
   
 
<haskell>
 
<haskell>
Line 252: Line 242:
 
</haskell>
 
</haskell>
   
But nevertheless it's a Good Style to use 'do' even for one statement
 
because it simplifies adding new statements in the future.
 
   
   
Let's examine how desugared 'do' with multiple statements on the
+
Let's examine how to desugar a 'do' with multiple statements in the
 
following example:
 
following example:
   
Line 265: Line 253:
 
</haskell>
 
</haskell>
   
'do' statement here just joins several IO actions that should be
+
The 'do' statement here just joins several IO actions that should be
 
performed sequentially. It's translated to sequential applications
 
performed sequentially. It's translated to sequential applications
of so named "binding operator", namely '>>':
+
of one of the so-called "binding operators", namely '>>':
   
 
<haskell>
 
<haskell>
Line 287: Line 275:
 
</haskell>
 
</haskell>
   
If such way to define operator looks strange for you, read this
+
If defining operators this way looks strange to you, read this
definition as the following:
+
definition as follows:
 
 
 
<haskell>
 
<haskell>
Line 298: Line 286:
 
</haskell>
 
</haskell>
   
Now you can substitute definition of '>>' at the places of it's usage
+
Now you can substitute the definition of '>>' at the places of its usage
and check that program constructed by 'do' desugaring is actually the
+
and check that program constructed by the 'do' desugaring is actually the
same as we can write by manually manipulating "world" values.
+
same as we could write by manually manipulating "world" values.
   
   
More complex example involves binding of variable using "<-":
+
A more complex example involves the binding of variables using "<-":
   
 
<haskell>
 
<haskell>
Line 317: Line 305:
 
</haskell>
 
</haskell>
   
As you should remember, '>>' binding operator silently ignores
+
As you should remember, the '>>' binding operator silently ignores
value of it's first action and returns as an overall result just
+
the value of its first action and returns as an overall result
result of second action. On the other side, '>>=' allows to use value
+
the result of its second action only. On the other hand, the '>>=' binding operator (note the extra '=' at the end) allows us to use the result of its first action - it gets passed as an additional parameter to the second one! Look at the definition:
of it's first action - it's passed as additional parameter to the second one!
 
Look at the definition:
 
   
 
<haskell>
 
<haskell>
(>>=) :: IO a -> (a->IO b) -> IO b
+
(>>=) :: IO a -> (a -> IO b) -> IO b
 
(action1 >>= action2) world0 =
 
(action1 >>= action2) world0 =
 
let (a, world1) = action1 world0
 
let (a, world1) = action1 world0
Line 331: Line 317:
 
</haskell>
 
</haskell>
   
First, what means type of second action, namely "a->IO b"? By
+
First, what does the type of the second "action" (more precisely, a function which returns an IO action), namely "a -> IO b", mean? By
 
substituting the "IO" definition, we get "a -> RealWorld -> (b, RealWorld)".
 
substituting the "IO" definition, we get "a -> RealWorld -> (b, RealWorld)".
 
This means that second action actually has two parameters
 
This means that second action actually has two parameters
- of type 'a' actually used inside it, and of type RealWorld used for
+
- the type 'a' actually used inside it, and the value of type RealWorld used for sequencing of IO actions. That's always the case - any IO procedure has one
  +
more parameter compared to what you see in its type signature. This
sequencing of IO actions. That's a destiny - any IO procedure has one
 
more parameter comparing to that you see in it's type signature. This
+
parameter is hidden inside the definition of the type alias "IO".
parameter is hidden inside the definition of type alias "IO".
 
   
 
Second, you can use these '>>' and '>>=' operations to simplify your
 
Second, you can use these '>>' and '>>=' operations to simplify your
 
program. For example, in the code above we don't need to introduce the
 
program. For example, in the code above we don't need to introduce the
variable, because 'readLn' result can be send directly to 'print':
+
variable, because the result of 'readLn' can be send directly to 'print':
   
 
<haskell>
 
<haskell>
Line 356: Line 341:
   
 
where 'action1' has type "IO a" and 'action2' has type "IO b",
 
where 'action1' has type "IO a" and 'action2' has type "IO b",
translated into:
+
translates into:
   
 
<haskell>
 
<haskell>
Line 362: Line 347:
 
</haskell>
 
</haskell>
   
where second argument of '>>=' has the type "a->IO b". It's the way
+
where the second argument of '>>=' has the type "a -> IO b". It's the way
  +
the '<-' binding is processed - the name on the left-hand side of '<-' just becomes a parameter of subsequent operations represented as one large IO action. Note also that if 'action1' has type "IO a" then 'x' will just have type "a"; you can think of the effect of '<-' as "unpacking" the IO value of 'action1' into 'x'. Note also that '<-' is not a true operator; it's pure syntax, just like 'do' itself. Its meaning results only from the way it gets desugared.
how the "<-" binding processed - it just becomes parameter of
 
  +
subsequent operations represented as one large IO action. Look at the
 
next example:
+
Look at the next example:
   
 
<haskell>
 
<haskell>
Line 385: Line 370:
 
</haskell>
 
</haskell>
   
I omitted parentheses here, both '>>' and '>>=' operations are
+
I omitted the parentheses here; both the '>>' and the '>>=' operators are
left-associative that leads to that 'a' and 'b' bindings introduced
+
left-associative, but lambda-bindings always stretches as far to the right as possible, which means that the 'a' and 'b' bindings introduced
here is valid for all remaining actions. As an exercise, add the
+
here are valid for all remaining actions. As an exercise, add the
 
parentheses yourself and translate this procedure into the low-level
 
parentheses yourself and translate this procedure into the low-level
code passing "world" values. I think it should be enough to finally
+
code that explicitly passes "world" values. I think it should be enough to help you finally realize how the 'do' translation and binding operators work.
realize how 'do' translation and binding operators work.
 
   
   
Oh, no. I forgot third monadic operator - 'return'. It just
+
Oh, no! I forgot the third monadic operator - 'return'. It just
combines it's two parameters - value passed and "world":
+
combines its two parameters - the value passed and "world":
   
 
<haskell>
 
<haskell>
Line 401: Line 385:
 
</haskell>
 
</haskell>
   
How about translating some simple example of 'return' usage? Say,
+
How about translating a simple example of 'return' usage? Say,
   
 
<haskell>
 
<haskell>
Line 409: Line 393:
   
   
Programmers with imperative languages background often thinks that
+
Programmers with an imperative language background often think that
'return' in Haskell, like in other languages, immediately returns from
+
'return' in Haskell, as in other languages, immediately returns from
the IO procedure. As you can see in its definition (and even just
+
the IO procedure. As you can see in its definition (and even just from its
type!), such assumption is totally wrong. The only purpose of using
+
type!), such an assumption is totally wrong. The only purpose of using
 
'return' is to "lift" some value (of type 'a') into the result of
 
'return' is to "lift" some value (of type 'a') into the result of
whole action (of type "IO a") and therefore it should be used only as
+
a whole action (of type "IO a") and therefore it should generally be used only as the last executed statement of some IO sequence. For example try to
  +
translate the following procedure into the corresponding low-level code:
last executed statements of some IO sequence. For example try to
 
translate the following procedure into the low-level code:
 
   
 
<haskell>
 
<haskell>
Line 425: Line 408:
 
</haskell>
 
</haskell>
   
and you will realize that 'print' statement is executed anyway. If you
+
and you will realize that the 'print' statement is executed even for non-negative values of 'a'. If you need to escape from the middle of an IO procedure, you can use the 'if' statement:
need to escape from middle of IO procedure, you can use the 'if'
 
statement:
 
   
 
<haskell>
 
<haskell>
Line 446: Line 427:
 
</haskell>
 
</haskell>
   
that may be very useful for escaping from middle of longish 'do' statement.
+
that may be useful for escaping from the middle of a longish 'do' statement.
   
   
Last exercise: implement function 'liftM' that lifts operations on
+
Last exercise: implement a function 'liftM' that lifts operations on
plain values to the operations on monadic ones. It's type signature:
+
plain values to the operations on monadic ones. Its type signature:
   
 
<haskell>
 
<haskell>
liftM :: (a->b) -> (IO a -> IO b)
+
liftM :: (a -> b) -> (IO a -> IO b)
 
</haskell>
 
</haskell>
   
If it's too hard for you, start with the following high-level
+
If that's too hard for you, start with the following high-level
 
definition and rewrite it in low-level fashion:
 
definition and rewrite it in low-level fashion:
   
Line 463: Line 444:
 
return (f x)
 
return (f x)
 
</haskell>
 
</haskell>
 
 
   
 
== Mutable data (references, arrays, hash tables...) ==
 
== Mutable data (references, arrays, hash tables...) ==
   
  +
As you should know, every name in Haskell is bound to one fixed (immutable) value. This greatly simplifies understanding algorithms and code optimization, but it's inappropriate in some cases. As we all know, there are plenty of algorithms that are simpler to implement in terms of updatable
As you should know, all names in Haskell are bind to one fixed value.
 
This greatly simplify understanding of algorithms and optimization of
 
code, but inappropriate for some cases. Yes, there a plenty of
 
algorithms that is simpler to implement in terms of updatable
 
 
variables, arrays and so on. This means that the value associated with
 
variables, arrays and so on. This means that the value associated with
variable, for example, can be different at different execution points,
+
a variable, for example, can be different at different execution points,
so reading it's value can't be considered as pure function. Imagine,
+
so reading its value can't be considered as a pure function. Imagine,
for example the following code:
+
for example, the following code:
   
 
<haskell>
 
<haskell>
Line 481: Line 457:
 
_ = writeVariable varA 1
 
_ = writeVariable varA 1
 
a1 = readVariable varA
 
a1 = readVariable varA
print (a0,a1)
+
print (a0, a1)
 
</haskell>
 
</haskell>
   
Looks strange? First, two calls to 'readVariable' looks the same, so
+
Does this look strange? First, the two calls to 'readVariable' look the same, so the compiler can just reuse the value returned by the first call. Second,
  +
the result of the 'writeVariable' call isn't used so the compiler can (and will!) omit this call completely. To complete the picture, these three calls may be rearranged in any order because they appear to be independent of each
compiler can just reuse the value returned by first call. Second,
 
  +
other. This is obviously not what was intended. What's the solution? You already know this - use IO actions! Using IO actions guarantees that:
result of 'writeVariable' call isn't used so compiler can (and will!)
 
omit this call completely. To finish the picture, these 3 calls may be
 
rearranged to any order because they looking independent on each
 
other. What is the solution? You know - using of IO actions! IO
 
actions guarantees us that:
 
   
# execution order will be retained
+
# the execution order will be retained as written
# each action will be mandatory executed
+
# each action will have to be executed
# result of the "same" action (such as "readVariable varA") will not be reused
+
# the result of the "same" action (such as "readVariable varA") will not be reused
   
 
So, the code above really should be written as:
 
So, the code above really should be written as:
   
 
<haskell>
 
<haskell>
  +
import Data.IORef
main = do varA <- newIORef 0 -- Create and initialize new variable
 
  +
main = do varA <- newIORef 0 -- Create and initialize a new variable
 
a0 <- readIORef varA
 
a0 <- readIORef varA
 
writeIORef varA 1
 
writeIORef varA 1
 
a1 <- readIORef varA
 
a1 <- readIORef varA
print (a0,a1)
+
print (a0, a1)
 
</haskell>
 
</haskell>
   
Here, 'varA' got type "IORef Int" which means "variable (reference) in
+
Here, 'varA' has the type "IORef Int" which means "a variable (reference) in
IO monad holding value of type Int". newIORef creates new variable
+
the IO monad holding a value of type Int". newIORef creates a new variable
 
(reference) and returns it, and then read/write actions use this
 
(reference) and returns it, and then read/write actions use this
reference. Value returned by "readIORef varA" action may depend not
+
reference. The value returned by the "readIORef varA" action depends not
only on variable involved but also on the moment of performing this
+
only on the variable involved but also on the moment this operation is performed so it can return different values on each call.
operation so it can return different values on each call.
 
   
 
Arrays, hash tables and any other _mutable_ data structures are
 
Arrays, hash tables and any other _mutable_ data structures are
defined in the same way - there is operation that creates new "mutable
+
defined in the same way - for each of them, there's an operation that creates new "mutable values" and returns a reference to it. Then special read and write
  +
operations in the IO monad are used. The following code shows an example
value" and returns reference to it. Then special read and write
 
  +
using mutable arrays:
operations in IO monad are used. The following example shows example
 
of using mutable array:
 
   
 
<haskell>
 
<haskell>
import Data.Array.IO
+
import Data.Array.IO
main = do arr <- newArray (1,10) 37 :: IO (IOArray Int Int)
+
main = do arr <- newArray (1,10) 37 :: IO (IOArray Int Int)
a <- readArray arr 1
+
a <- readArray arr 1
writeArray arr 1 64
+
writeArray arr 1 64
b <- readArray arr 1
+
b <- readArray arr 1
print (a,b)
+
print (a, b)
 
</haskell>
 
</haskell>
   
Here, array of 10 elements with 37 as initial values is created. After
+
Here, an array of 10 elements with 37 as the initial value at each location is created. After reading the value of the first element (index 1) into 'a' this element's value is changed to 64 and then read again into 'b'. As you can see by executing this code, 'a' will be set to 37 and 'b' to 64.
reading value of first element to 'a' this element's value is changed
 
to 64 and then read again, to 'b'. As you can see by executing this
 
code, 'a' will be set to 37 and 'b' to 64.
 
 
 
   
   
 
Other state-dependent operations are also often implemented as IO
 
Other state-dependent operations are also often implemented as IO
actions. For example, random numbers generator should return different
+
actions. For example, a random number generator should return a different
values on each call. It looks natural to give it IO-involving type:
+
value on each call. It looks natural to give it a type involving IO:
   
 
<haskell>
 
<haskell>
Line 544: Line 512:
   
 
Moreover, when you import C routines you should be careful - if this
 
Moreover, when you import C routines you should be careful - if this
routine is impure, i.e. it's result depends on something in "real
+
routine is impure, i.e. its result depends on something in the "real
 
world" (file system, memory contents...), internal state and so on,
 
world" (file system, memory contents...), internal state and so on,
you should give it IO-involving type. Otherwise, compiler can
+
you should give it an IO type. Otherwise, the compiler can
"optimize" repetitive calls of this procedure with the same parameters! :)
+
"optimize" repetitive calls of this procedure with the same parameters!
   
For example:
+
For example, we can write a non-IO type for:
   
 
<haskell>
 
<haskell>
Line 556: Line 524:
 
</haskell>
 
</haskell>
   
because 'sin' result depends only on it's argument, but
+
because the result of 'sin' depends only on its argument, but
   
 
<haskell>
 
<haskell>
Line 563: Line 531:
 
</haskell>
 
</haskell>
   
If you will declare 'tell' as pure function (without IO) then you may
+
If you will declare 'tell' as a pure function (without IO) then you may
got the same position on each call! :)
+
get the same position on each call!
   
 
== IO actions as values ==
 
== IO actions as values ==
   
Now you should precisely understand why it's impossible to use IO
+
By this point you should understand why it's impossible to use IO
 
actions inside non-IO (pure) procedures. Such procedures just don't
 
actions inside non-IO (pure) procedures. Such procedures just don't
get a "baton", don't know any "world" value to pass to IO action.
+
get a "baton"; they don't know any "world" value to pass to an IO action.
RealWorld is abstract datatype, so they also can't construct it's
+
The RealWorld type is an abstract datatype, so pure functions also can't construct RealWorld values by themselves, and it's a strict type, so 'undefined' also can't be used. So, the prohibition of using IO actions inside pure procedures is just a type system trick (as it usually is in Haskell).
values by himself, and it's a strict type, so 'undefined' also can't
 
be used. So, prohibition of using IO actions inside pure procedures is
 
just a type trick as it is usual in Haskell :)
 
   
 
But while pure code can't _execute_ IO actions, it can work with them
 
But while pure code can't _execute_ IO actions, it can work with them
 
as with any other functional values - they can be stored in data
 
as with any other functional values - they can be stored in data
structures, passed as parameters and returned as results, collected in
+
structures, passed as parameters, returned as results, collected in
lists, and partially applied. But anyway IO action will remain
+
lists, and partially applied. But an IO action will remain a
 
functional value because we can't apply it to the last argument - of
 
functional value because we can't apply it to the last argument - of
 
type RealWorld.
 
type RealWorld.
   
 
In order to _execute_ the IO action we need to apply it to some
 
In order to _execute_ the IO action we need to apply it to some
RealWorld value that can be done only inside some IO procedure,
+
RealWorld value. That can be done only inside some IO procedure,
in it's "actions chain". And real execution of this action will take
+
in its "actions chain". And real execution of this action will take
place only when this procedure is called as part of process of
+
place only when this procedure is called as part of the process of
"calculating final value of world" for 'main'. Look at this example:
+
"calculating the final value of world" for 'main'. Look at this example:
   
 
<haskell>
 
<haskell>
main = let get2chars = getChar >> getChar
+
main world0 = let get2chars = getChar >> getChar
((), world1) = putStr "Press two keys" world0
+
((), world1) = putStr "Press two keys" world0
(answer, world2) = get2chars world1
+
(answer, world2) = get2chars world1
in ((), world2)
+
in ((), world2)
 
</haskell>
 
</haskell>
 
 
Here we first bind value to 'get2chars' and then write binding
+
Here we first bind a value to 'get2chars' and then write a binding
involving 'putStr'. But what is an execution order? It is not defined
+
involving 'putStr'. But what's the execution order? It's not defined
by order of writing bindings, it is defined by order of processing
+
by the order of the 'let' bindings, it's defined by the order of processing
"world" values! You can arbitrarily reorder binding statements - in
+
"world" values! You can arbitrarily reorder the binding statements - the execution order will be defined by the data dependency with respect to the
  +
"world" values that get passed around. Let's see what this 'main' looks like in the 'do' notation:
any case execution order will be defined by dependence on passing
 
"world" values. Let's see how this 'main' looks in the 'do' notation:
 
   
 
<haskell>
 
<haskell>
Line 610: Line 574:
 
</haskell>
 
</haskell>
   
  +
As you can see, we've eliminated two of the 'let' bindings and left only the one defining 'get2chars'. The non-'let' statements are executed in the exact order in which they're written, because they pass the "world" value from statement to statement as we described above. Thus, this version of the function is much easier to understand because we don't have to mentally figure out the data dependency of the "world" value.
As you can see, the 'let' binding that is not included in IO chain, is
 
translated just to 'let' statement inside the 'do' sequence. And as
 
you now should understand, placement of this 'let' don't has any
 
impact on the evaluation order, which is defined by order of passing
 
"world" values that is, in turn, defined by order of ordinal (non-let)
 
statements inside 'do'!
 
   
Moreover, IO actions like this 'get2chars' can't be executed just
+
Moreover, IO actions like 'get2chars' can't be executed directly
because they are functions with RealWorld parameter. To execute them,
+
because they are functions with a RealWorld parameter. To execute them,
we should supply the RealWorld parameter, i.e. insert them in 'main'
+
we need to supply the RealWorld parameter, i.e. insert them in the 'main'
chain, placing them in some 'do' sequence executed from 'main'. Until
+
chain, placing them in some 'do' sequence executed from 'main' (either directly in the 'main' function, or indirectly in an IO function called from 'main'). Until that's done, they will remain like any function, in partially
that is done, they will be keep as any function, in partially
 
 
evaluated form. And we can work with IO actions as with any other
 
evaluated form. And we can work with IO actions as with any other
functions - bind them to names (like above), save them to data
+
functions - bind them to names (as we did above), save them in data
structures, pass as function parameters and return as results - and
+
structures, pass them as function parameters and return them as results - and
they will not be performed until you give them this magic RealWorld
+
they won't be performed until you give them the magic RealWorld
 
parameter!
 
parameter!
   
   
   
=== Example: list of IO actions ===
+
=== Example: a list of IO actions ===
   
Let's try. How about defining list of IO actions?
+
Let's try defining a list of IO actions:
   
 
<haskell>
 
<haskell>
Line 642: Line 600:
 
</haskell>
 
</haskell>
   
I used additional parentheses around each action, although they are
+
I used additional parentheses around each action, although they aren't really required. If you still can't believe that these actions won't be executed immediately, just recall the real type of this list:
not really required. If you still can't belive that these actions will
 
not be executed until your command, just uncover this list type:
 
 
 
 
<haskell>
 
<haskell>
Line 659: Line 615:
 
</haskell>
 
</haskell>
   
Looks strange, yeah? :) Really, any IO action you write in the 'do'
+
Looks strange, right? Really, any IO action that you write in a 'do'
statement (or use as parameter for '>>'/'>>=') is an expression
+
statement (or use as a parameter for the '>>'/'>>=' operators) is an expression
returning result of type "IO a". Typically, you use some function that
+
returning a result of type 'IO a' for some type 'a'. Typically, you use some function that has the type 'x -> y -> ... -> IO a' and provide all the x, y, etc. parameters. But you're not limited to this standard scenario -
  +
don't forget that Haskell is a functional language and you're free to
has type "x -> y -> ... -> IO a" and provide all these x, y and
 
  +
compute the functional value required (recall that "IO a" is really a function
so on parameters. But you are not limited to this standard scenario -
 
don't forget that Haskell is functional language and you are free to
 
compute the functional value required (recall - "IO a" is a function
 
 
type) in any possible way. Here we just extracted several functions
 
type) in any possible way. Here we just extracted several functions
 
from the list - no problem. This functional value can also be
 
from the list - no problem. This functional value can also be
constructed on-the-fly, as we've done in previous example - it's also
+
constructed on-the-fly, as we've done in the previous example - that's also
ok. Want to see this functional value passed as the parameter - heh,
+
OK. Want to see this functional value passed as a parameter?
just look at the 'when' definition. Hey, we can sell, buy and rent
+
Just look at the definition of 'when'. Hey, we can buy, sell, and rent
these IO actions as any other functional values! For example, let's
+
these IO actions just like we can with any other functional values! For example, let's define a function that executes all the IO actions in the list:
define function that executes all IO actions in the list:
 
   
 
<haskell>
 
<haskell>
Line 681: Line 634:
 
</haskell>
 
</haskell>
   
No black magic - we just extracts IO actions from the list and inserts
+
No black magic - we just extract IO actions from the list and insert
them into chain of IO operations that should be performed to "compute
+
them into a chain of IO operations that should be performed one after another (in the same order that they occurred in the list) to "compute the final world value" of the entire 'sequence_' call.
final world value" of entire 'sequence_' call.
 
   
With help of 'sequence_', we can rewrite our last 'main' as:
+
With the help of 'sequence_', we can rewrite our last 'main' function as:
   
 
<haskell>
 
<haskell>
Line 694: Line 646:
 
Haskell's ability to work with IO actions as with any other
 
Haskell's ability to work with IO actions as with any other
 
(functional and non-functional) values allows us to define control
 
(functional and non-functional) values allows us to define control
structures of any complexity. Try, for example, to define control
+
structures of arbitrary complexity. Try, for example, to define a control
structure that repeats the action until it returns the 'False' result:
+
structure that repeats an action until it returns the 'False' result:
   
 
<haskell>
 
<haskell>
Line 702: Line 654:
 
</haskell>
 
</haskell>
   
  +
Most programming languages don't allow you to define control structures at all, and those that do often require you to use a macro-expansion system. In Haskell, control structures are just trivial functions anyone can write.
   
   
=== Example: returning IO action as a result ===
+
=== Example: returning an IO action as a result ===
   
How about returning IO action as the function result? Well, we done
+
How about returning an IO action as the result of a function? Well, we've done
this each time we defined IO procedure - they all return IO action
+
this each time we've defined an IO procedure - they all return IO actions
that need RealWorld value to be performed. While we most times just
+
that need a RealWorld value to be performed. While we usually just
executed them in chain of higher-level IO procedure, it's also
+
execute them as part of a higher-level IO procedure, it's also
 
possible to just collect them without actual execution:
 
possible to just collect them without actual execution:
   
Line 716: Line 669:
 
b = when True getChar
 
b = when True getChar
 
c = getChar >> getChar
 
c = getChar >> getChar
putStr "'let' statements are not executed!"
+
putStr "These 'let' statements are not executed!"
 
</haskell>
 
</haskell>
   
 
These assigned IO procedures can be used as parameters to other
 
These assigned IO procedures can be used as parameters to other
 
procedures, or written to global variables, or processed in some other
 
procedures, or written to global variables, or processed in some other
way, or just executed later, as we done in example with 'get2chars'.
+
way, or just executed later, as we did in the example with 'get2chars'.
   
But how about returning from IO procedure a parameterized IO action?
+
But how about returning a parameterized IO action from an IO procedure? Let's define a procedure that returns the i'th byte from a file represented as a Handle:
Let's define a procedure that returns i'th byte from file represented
 
as Handle:
 
   
 
<haskell>
 
<haskell>
readi h i = do hSeek h i AbsoluteSeek
+
readi h i = do hSeek h AbsoluteSeek i
 
hGetChar h
 
hGetChar h
 
</haskell>
 
</haskell>
   
So bad so good. But how about procedure that returns i'th byte of file
+
So far so good. But how about a procedure that returns the i'th byte of a file
with given name without reopening it each time?
+
with a given name without reopening it each time?
   
 
<haskell>
 
<haskell>
Line 741: Line 692:
 
</haskell>
 
</haskell>
   
As you can see, it's an IO procedure that opens file and returns...
+
As you can see, it's an IO procedure that opens a file and returns...
another IO procedure that will read byte specified. But we can go
+
another IO procedure that will read the specified byte. But we can go
further and include 'readi' body into 'readfilei':
+
further and include the 'readi' body in 'readfilei':
   
 
<haskell>
 
<haskell>
 
readfilei name = do h <- openFile name ReadMode
 
readfilei name = do h <- openFile name ReadMode
let readi h i = do hSeek h i AbsoluteSeek
+
let readi h i = do hSeek h AbsoluteSeek i
 
hGetChar h
 
hGetChar h
 
return (readi h)
 
return (readi h)
 
</haskell>
 
</haskell>
   
Good? May be better. Why we add 'h' as 'readi' parameter if it can be
+
That's a little better. But why do we add 'h' as a parameter to 'readi' if it can be obtained from the environment where 'readi' is now defined? An even shorter version is this:
got from the environment where 'readi' now defined? Shorter will be:
 
   
 
<haskell>
 
<haskell>
 
readfilei name = do h <- openFile name ReadMode
 
readfilei name = do h <- openFile name ReadMode
let readi i = do hSeek h i AbsoluteSeek
+
let readi i = do hSeek h AbsoluteSeek i
 
hGetChar h
 
hGetChar h
 
return readi
 
return readi
 
</haskell>
 
</haskell>
   
What we've done here? We've build parameterized IO action involving local
+
What have we done here? We've build a parameterized IO action involving local
 
names inside 'readfilei' and returned it as the result. Now it can be
 
names inside 'readfilei' and returned it as the result. Now it can be
used in following way:
+
used in the following way:
   
 
<haskell>
 
<haskell>
Line 774: Line 724:
   
   
Such usage of IO actions is very typical for Haskell programs - you
+
This way of using IO actions is very typical for Haskell programs - you
just construct one or more (using tuple) IO actions that your need,
+
just construct one or more IO actions that you need,
with and/or without parameters, involving the parameters that your
+
with or without parameters, possibly involving the parameters that your
"constructor" received, and return them to caller. Then these IO actions
+
"constructor" received, and return them to the caller. Then these IO actions
can be used in rest of program without any knowledge about your
+
can be used in the rest of the program without any knowledge about your
internal implementation strategies. Actually, this is used to
+
internal implementation strategy. One thing this can be used for is to
partially emulate OOP (to be exact, ADT) programming ideology.
+
partially emulate the OOP (or more precisely, the ADT) programming paradigm.
   
  +
=== Example: a memory allocator generator ===
   
  +
As an example, one of my programs has a module which is a memory suballocator. It receives the address and size of a large memory block and returns two
=== Example: memory allocators generator ===
 
  +
procedures - one to allocate a subblock of a given size and the other to
 
  +
free the allocated subblock:
For example, one of my program's modules is the memory suballocator. It
 
receives address and size of large memory block and returns two
 
procedures - one to allocate subblock of given size and second to
 
return allocated block back:
 
   
 
<haskell>
 
<haskell>
Line 802: Line 750:
 
</haskell>
 
</haskell>
   
How this is implemented? 'alloc' and 'free' works with references
+
How this is implemented? 'alloc' and 'free' work with references
created inside this procedure. Because creation of these references is
+
created inside the memoryAllocator procedure. Because the creation of these references is a part of the memoryAllocator IO actions chain, a new independent set of references will be created for each memory block for which
a part of 'memoryAllocator' IO actions chain, new independent set of
+
memoryAllocator is called:
references will be created for each memory block for which
 
'memoryAllocator' is called:
 
   
 
<haskell>
 
<haskell>
Line 814: Line 760:
 
</haskell>
 
</haskell>
   
These two references (we will implement very simple memory allocator) are
+
These two references are read and written in the 'alloc' and 'free' definitions (we'll implement a very simple memory allocator for this example):
read and written in 'alloc' and 'free' definitions:
 
   
 
<haskell>
 
<haskell>
  +
...
 
let alloc size = do addr <- readIORef start
 
let alloc size = do addr <- readIORef start
 
writeIORef start (addr `plusPtr` size)
 
writeIORef start (addr `plusPtr` size)
Line 825: Line 771:
 
</haskell>
 
</haskell>
   
What we've defined here is just a pair of closures that is using state
+
What we've defined here is just a pair of closures that use state
available on the moment of their definition. As you can see, it's as
+
available at the moment of their definition. As you can see, it's as
easy as in any other functional language, despite the Haskell's lack
+
easy as in any other functional language, despite Haskell's lack
of direct support for non-pure functions.
+
of direct support for impure functions.
 
 
 
The following example uses procedures, returned by memoryAllocator, to
 
The following example uses procedures, returned by memoryAllocator, to
Line 848: Line 794:
   
   
=== Example: emulating OOP with record type ===
+
=== Example: emulating OOP with record types ===
   
 
Let's implement the classical OOP example: drawing figures. There are
 
Let's implement the classical OOP example: drawing figures. There are
 
figures of different types: circles, rectangles and so on. The task is
 
figures of different types: circles, rectangles and so on. The task is
to create heterogeneous figures list. All figures in this list should
+
to create a heterogeneous list of figures. All figures in this list should
 
support the same set of operations: draw, move and so on. We will
 
support the same set of operations: draw, move and so on. We will
represent these operations as IO procedures. Instead of class let's
+
represent these operations as IO procedures. Instead of a "class" let's
 
define a structure containing implementations of all the procedures
 
define a structure containing implementations of all the procedures
 
required:
 
required:
Line 867: Line 813:
   
   
Constructor of each figures' type should just return a Figure record:
+
The constructor of each figure's type should just return a Figure record:
   
 
<haskell>
 
<haskell>
Line 879: Line 825:
   
 
We will "draw" figures by just printing their current parameters.
 
We will "draw" figures by just printing their current parameters.
Let's start with simplified implementation of 'circle' and 'rectangle'
+
Let's start with a simplified implementation of the 'circle' and 'rectangle'
 
constructors, without actual 'move' support:
 
constructors, without actual 'move' support:
   
Line 885: Line 831:
 
circle center radius = do
 
circle center radius = do
 
let description = " Circle at "++show center++" with radius "++show radius
 
let description = " Circle at "++show center++" with radius "++show radius
return $ Figure { draw = putStrLn description}
+
return $ Figure { draw = putStrLn description }
   
 
rectangle from to = do
 
rectangle from to = do
 
let description = " Rectangle "++show from++"-"++show to)
 
let description = " Rectangle "++show from++"-"++show to)
return $ Figure { draw = putStrLn description}
+
return $ Figure { draw = putStrLn description }
 
</haskell>
 
</haskell>
   
   
As you see, they just returns fixed 'draw' procedure that prints
+
As you see, each constructor just returns a fixed 'draw' procedure that prints
parameters with which this concrete figure was created. Let's go to
+
parameters with which the concrete figure was created. Let's test it:
test it:
 
   
 
<haskell>
 
<haskell>
Line 910: Line 855:
   
   
Now let's go to define "full-featured" figures that can be really
+
Now let's define "full-featured" figures that can actually be
 
moved around. In order to achieve this, we should provide each figure
 
moved around. In order to achieve this, we should provide each figure
with mutable variables what holds it's current screen location. Their
+
with a mutable variable that holds each figure's current screen location. The
type will be "IORef Point". These variables should be created in figure
+
type of this variable will be "IORef Point". This variable should be created in the figure constructor and manipulated in IO procedures (closures) enclosed in
  +
the Figure record:
constructor and manipulated in IO procedures (closures) enclosed in
 
Figure record:
 
   
 
<haskell>
 
<haskell>
Line 922: Line 866:
 
 
 
let drawF = do center <- readIORef centerVar
 
let drawF = do center <- readIORef centerVar
putStrLn (" Circle at "++show center++" with radius "++show radius)
+
putStrLn (" Circle at "++show center
  +
++" with radius "++show radius)
 
 
 
let moveF (addX,addY) = do (x,y) <- readIORef centerVar
 
let moveF (addX,addY) = do (x,y) <- readIORef centerVar
Line 947: Line 892:
   
   
Now we can add moving figures around to our test:
+
Now we can test the code which moves figures around:
   
 
<haskell>
 
<haskell>
Line 958: Line 903:
   
   
Let's pay attention that we are not limited to include only IO actions
+
It's important to realize that we are not limited to including only IO actions
in the record that simulates C++/Java interface. The record can include
+
in a record that's intended to simulate a C++/Java-style interface. The record can also include values, IORefs, pure functions - in short, any type of data. For example, we can easily add to the Figure interface fields for area and origin:
values, IORefs, pure functions - in short, any type of data. For
 
example, we can easily add to Figure interface fields for area and
 
origin:
 
   
 
<haskell>
 
<haskell>
Line 974: Line 916:
   
   
  +
== Exception handling (under development) ==
== unsafePerformIO and unsafeInterleaveIO ==
 
   
  +
Although Haskell provides a set of exception raising/handling features comparable to those in popular OOP languages (C++, Java, C#), this part of the language receives much less attention. This is for two reasons. First, you just don't need to worry as much about them - most of the time it just works "behind the scenes". The second reason is that Haskell, lacking OOP inheritance, doesn't allow the programmer to easily subclass exception types, therefore limiting flexibility of exception handling.
Programmers with imperative background often still looks for a ways to
 
  +
execute IO actions inside the pure procedures. But that this means?
 
  +
The Haskell RTS raises more exceptions than traditional languages - pattern match failures, calls with invalid arguments (such as '''head []''') and computations whose results depend on special values '''undefined''' and '''error "...."''' all raise their own exceptions:
Imagine that you try to write procedure that reads contents of file
 
  +
with given name:
 
  +
example 1:
  +
<haskell>
  +
main = print (f 2)
  +
  +
f 0 = "zero"
  +
f 1 = "one"
  +
</haskell>
  +
  +
example 2:
  +
<haskell>
  +
main = print (head [])
  +
</haskell>
  +
  +
example 3:
  +
<haskell>
  +
main = print (1 + (error "Value that wasn't initialized or cannot be computed"))
  +
</haskell>
  +
  +
This allows to write programs in much more error-prone way.
  +
  +
== Interfacing with C/C++ and foreign libraries (under development) ==
  +
  +
While Haskell is great at algorithm development, speed isn't its best side. We can combine the best of both worlds, though, by writing speed-critical parts of program in C and rest in Haskell. We just need a way to call C functions from Haskell and vice versa, and to marshal data between two worlds.
  +
  +
We also need to interact with C world for using Windows/Linux APIs, linking to various libraries and DLLs. Even interfacing with other languages requires to go through C world as "common denominator". [https://www.haskell.org/onlinereport/haskell2010/haskellch8.html Chapter 8 of the Haskell 2010 report] provides a complete description of interfacing with C.
  +
  +
We will learn FFI via a series of examples. These examples include C/C++ code, so they need C/C++ compilers to be installed, the same will be true if you need to include code written in C/C++ in your program (C/C++ compilers are not required when you just need to link with existing libraries providing APIs with C calling convention). On Unix (and Mac OS?) systems, the system-wide default C/C++ compiler is typically used by GHC installation. On Windows, no default compilers exist, so GHC is typically shipped with a C compiler, and you may find on the download page a GHC distribution bundled with C and C++ compilers. Alternatively, you may find and install a GCC/MinGW version compatible with your GHC installation.
  +
  +
If you need to make your C/C++ code as fast as possible, you may compile your code by Intel compilers instead of GCC. However, these compilers are not free, moreover on Windows, code compiled by Intel compilers may not interact correctly with GHC-compiled code, unless one of them is put into DLLs (due to object file incompatibility).
  +
  +
[http://www.haskell.org/haskellwiki/Applications_and_libraries/Interfacing_other_languages More links]:
  +
  +
;[http://www.cse.unsw.edu.au/~chak/haskell/c2hs/ C-&gt;Haskell]
  +
:A lightweight tool for implementing access to C libraries from Haskell.
  +
  +
;[[HSFFIG]]
  +
:Haskell FFI Binding Modules Generator (HSFFIG) is a tool that takes a C library include file (.h) and generates Haskell Foreign Functions Interface import declarations for items (functions, structures, etc.) the header defines.
  +
  +
;[http://quux.org/devel/missingpy MissingPy]
  +
:MissingPy is really two libraries in one. At its lowest level, MissingPy is a library designed to make it easy to call into Python from Haskell. It provides full support for interpreting arbitrary Python code, interfacing with a good part of the Python/C API, and handling Python objects. It also provides tools for converting between Python objects and their Haskell equivalents. Memory management is handled for you, and Python exceptions get mapped to Haskell Dynamic exceptions. At a higher level, MissingPy contains Haskell interfaces to some Python modules.
  +
  +
;[[HsLua]]
  +
:A Haskell interface to the Lua scripting language
  +
  +
  +
=== Calling functions ===
  +
  +
First, we will learn how to call C functions from Haskell and Haskell functions from C. The first example consists of three files:
  +
  +
main.hs:
  +
<haskell>
  +
{-# LANGUAGE ForeignFunctionInterface #-}
  +
  +
main = do print "Hello from main"
  +
c_function
  +
  +
haskell_function = print "Hello from haskell_function"
  +
  +
foreign import ccall safe "prototypes.h"
  +
c_function :: IO ()
  +
  +
foreign export ccall
  +
haskell_function :: IO ()
  +
</haskell>
  +
  +
evil.c:
  +
<haskell>
  +
#include <stdio.h>
  +
#include "prototypes.h"
  +
  +
void c_function (void)
  +
{
  +
printf("Hello from c_function\n");
  +
haskell_function();
  +
}
  +
</haskell>
  +
  +
prototypes.h:
  +
<haskell>
  +
extern void c_function (void);
  +
extern void haskell_function (void);
  +
</haskell>
  +
  +
It may be compiled and linked in one step by ghc:
  +
ghc --make main.hs evil.c
  +
  +
Or, you may compile C module(s) separately and link in .o files (this may be preferable if you use <code>make</code> and don't want to recompile unchanged sources; ghc's --make option provides smart recompilation only for .hs files):
  +
ghc -c evil.c
  +
ghc --make main.hs evil.o
  +
  +
You may use gcc/g++ directly to compile your C/C++ files but I recommend to do linking via ghc because it adds a lot of libraries required for execution of Haskell code. For the same reason, even if your main routine is written in C/C++, I recommend calling it from the Haskell function <hask>main</hask> - otherwise you'll have to explicitly init/shutdown the GHC RTS (run-time system).
  +
  +
We use the "foreign import" specification to import foreign routines into our Haskell world, and "foreign export" to export Haskell routines into the external world. Note that the import statement creates a new Haskell symbol (from the external one), while the export statement uses a Haskell symbol previously defined. Technically speaking, both types of statements create a wrapper that converts the names and calling conventions from C to Haskell or vice versa.
  +
  +
=== All about the "foreign" statement ===
  +
  +
The "ccall" specifier in foreign statements means the use of C (not C++ !) calling convention. This means that if you want to write the external function in C++ (instead of C) you should add '''export "C"''' specification to its declaration - otherwise you'll get linking errors. Let's rewrite our first example to use C++ instead of C:
  +
  +
prototypes.h:
  +
<haskell>
  +
#ifdef __cplusplus
  +
extern "C" {
  +
#endif
  +
  +
extern void c_function (void);
  +
extern void haskell_function (void);
  +
  +
#ifdef __cplusplus
  +
}
  +
#endif
  +
</haskell>
  +
  +
Compile it via:
  +
  +
ghc --make main.hs evil.cpp
  +
  +
where evil.cpp is just a renamed copy of evil.c from the first example. Note that the new prototypes.h is written to allow compiling it both as C and C++ code. When it's included from evil.cpp, it's compiled as C++ code. When GHC compiles main.hs via the C compiler (enabled by -fvia-C option), it also includes prototypes.h but compiles it in C mode. It's why you need to specify .h files in "foreign" declarations - depending on which Haskell compiler you use, these files may be included to check consistency of C and Haskell declarations.
  +
  +
The quoted part of the foreign statement may also be used to import or export a function under another name--for example,
  +
  +
<haskell>
  +
foreign import ccall safe "prototypes.h CFunction"
  +
c_function :: IO ()
  +
  +
foreign export ccall "HaskellFunction"
  +
haskell_function :: IO ()
  +
</haskell>
  +
  +
specifies that the C function called CFunction will become known as the Haskell function c_function, while the Haskell function haskell_function will be known in the C world as HaskellFunction. It's required when the C name doesn't conform to Haskell naming requirements.
  +
  +
Although the Haskell FFI standard tells about many other calling conventions in addition to ccall (e.g. cplusplus, jvm, net) current Haskell implementations support only ccall and stdcall. The latter, also called the "Pascal" calling convention, is used to interface with WinAPI:
  +
  +
<haskell>
  +
foreign import stdcall unsafe "windows.h SetFileApisToOEM"
  +
setFileApisToOEM :: IO ()
  +
</haskell>
  +
  +
And finally, about the safe/unsafe specifier: a C function imported with the "unsafe" keyword is called directly and the Haskell runtime is stopped while the C function is executed (when there are several OS threads executing the Haskell program, only the current OS thread is delayed). This call doesn't allow recursively entering into the Haskell world by calling any Haskell function - the Haskell RTS is just not prepared for such an event. However, unsafe calls are as quick as calls in C world. It's ideal for "momentary" calls that quickly return back to the caller.
  +
  +
When "safe" is specified, the C function is called in safe environment - the Haskell execution context is saved, so it's possible to call back to Haskell and, if the C call takes a long time, another OS thread may be started to execute Haskell code (of course, in threads other than the one that called the C code). This has its own price, though - around 1000 CPU ticks per call.
  +
  +
You can read more about interaction between FFI calls and Haskell concurrency in [7].
  +
  +
=== Marshalling simple types ===
  +
  +
Calling by itself is relatively easy; the real problem of interfacing languages with different data models is passing data between them. In this case, there is no guarantee that Haskell's Int is represented in memory the same way as C's int, nor Haskell's Double the same as C's double and so on. While on *some* platforms they are the same and you can write throw-away programs relying on these, the goal of portability requires you to declare imported and exported functions using special types described in the FFI standard, which are guaranteed to correspond to C types. These are:
  +
  +
<haskell>
  +
import Foreign.C.Types ( -- equivalent to the following C type:
  +
CChar, CUChar, -- char/unsigned char
  +
CShort, CUShort, -- short/unsigned short
  +
CInt, CUInt, CLong, CULong, -- int/unsigned/long/unsigned long
  +
CFloat, CDouble...) -- float/double
  +
</haskell>
  +
  +
Now we can import and export typeful C/Haskell functions:
  +
<haskell>
  +
foreign import ccall unsafe "math.h"
  +
c_sin :: CDouble -> CDouble
  +
</haskell>
  +
  +
Note that pure C functions (those whose results depend only on their arguments) are imported without IO in their return type. The "const" specifier in C is not reflected in Haskell types, so appropriate compiler checks are not performed. <!-- What would these be? -->
  +
  +
All these numeric types are instances of the same classes as their Haskell cousins (Ord, Num, Show and so on), so you may perform calculations on these data directly. Alternatively, you may convert them to native Haskell types. It's very typical to write simple wrappers around imported and exported functions just to provide interfaces having native Haskell types:
  +
  +
<haskell>
  +
-- |Type-conversion wrapper around c_sin
  +
sin :: Double -> Double
  +
sin = fromRational . c_sin . toRational
  +
</haskell>
  +
  +
=== Memory management ===
  +
  +
=== Marshalling strings ===
  +
  +
<haskell>
  +
import Foreign.C.String ( -- representation of strings in C
  +
CString, -- = Ptr CChar
  +
CStringLen) -- = (Ptr CChar, Int)
  +
</haskell>
  +
  +
<haskell>
  +
foreign import ccall unsafe "string.h"
  +
c_strlen :: CString -> IO CSize -- CSize defined in Foreign.C.Types and is equal to size_t
  +
</haskell>
  +
  +
<haskell>
  +
-- |Type-conversion wrapper around c_strlen
  +
strlen :: String -> Int
  +
strlen = ....
  +
</haskell>
  +
  +
=== Marshalling composite types ===
  +
  +
A C array may be manipulated in Haskell as [http://haskell.org/haskellwiki/Arrays#StorableArray_.28module_Data.Array.Storable.29 StorableArray].
  +
  +
There is no built-in support for marshalling C structures and using C constants in Haskell. These are implemented in c2hs preprocessor, though.
  +
  +
Binary marshalling (serializing) of data structures of any complexity is implemented in library Binary.
  +
  +
=== Dynamic calls ===
  +
  +
=== DLLs ===
  +
''because i don't have experience of using DLLs, can someone write into this section? ultimately, we need to consider the following tasks:''
  +
* using DLLs of 3rd-party libraries (such as ziplib)
  +
* putting your own C code into a DLL to use in Haskell
  +
* putting Haskell code into a DLL which may be called from C code
  +
  +
== Dark side of IO monad ==
  +
=== unsafePerformIO ===
  +
  +
Programmers coming from an imperative language background often look for a way to execute IO actions inside a pure procedure. But what does this mean?
  +
Imagine that you're trying to write a procedure that reads the contents of a file with a given name, and you try to write it as a pure (non-IO) function:
   
 
<haskell>
 
<haskell>
Line 985: Line 1,140:
 
</haskell>
 
</haskell>
   
Defining it as pure function will simplify the code that use it, i
+
Defining readContents as a pure function will certainly simplify the code that uses it. But it will also create problems for the compiler:
agree. But this creates troubles for the compiler:
 
   
# This call is not inserted in sequence of "world transformations", so compiler don't get a hint - at what exact moment you want to execute this action. For example, if file contents is one at the program start and another at the end - what contents you want to see? Moment of "consumption" of this value don't make strong guarantees for execution order, because Haskell see all the functions as pure and fell free to reorder their execution as needed.
+
# This call is not inserted in a sequence of "world transformations", so the compiler doesn't know at what exact moment you want to execute this action. For example, if the file has one kind of contents at the beginning of the program and another at the end - which contents do you want to see? You have no idea when (or even if) this function is going to get invoked, because Haskell sees this function as pure and feels free to reorder the execution of any or all pure functions as needed.
# Attempts to read contents of file with the same name can be factorized despite the fact that file (or current directory) can be changed between calls. Again, Haskell looks at all the functions as pure ones and feel free to omit excessive calls with the same parameters.
+
# Attempts to read the contents of files with the same name can be factored (''i.e.'' reduced to a single call) despite the fact that the file (or the current directory) can be changed between calls. Again, Haskell considers all non-IO functions to be pure and feels free to omit multiple calls with the same parameters.
   
So, implementing functions that interacts with Real World as pure ones
+
So, implementing pure functions that interact with the Real World is
considered as a Bad Behavior. Good boys never do it ;)
+
considered to be Bad Behavior. Good boys and girls never do it ;)
   
   
 
Nevertheless, there are (semi-official) ways to use IO actions inside
 
Nevertheless, there are (semi-official) ways to use IO actions inside
 
of pure functions. As you should remember this is prohibited by
 
of pure functions. As you should remember this is prohibited by
  +
requiring the RealWorld "baton" in order to call an IO action. Pure functions don't have the baton, but there is a special "magic" procedure that produces this baton from nowhere, uses it to call an IO action and then throws the resulting "world" away! It's a little low-level magic. This very special (and dangerous) procedure is:
requiring "baton" to call IO action. Pure function don't have the baton,
 
but there is special procedure, that procures this baton from nowhere,
 
uses it to call IO action and then throws resulting "world" away!
 
A little low-level magic :) This very special procedure is:
 
   
 
<haskell>
 
<haskell>
Line 1,006: Line 1,157:
 
</haskell>
 
</haskell>
   
Let's look at it's (possible) definition:
+
Let's look at its (possible) definition:
   
 
<haskell>
 
<haskell>
unsafePerformIO :: (RealWorld -> (a,RealWorld)) -> a
+
unsafePerformIO :: (RealWorld -> (a, RealWorld)) -> a
unsafePerformIO action = let (a,world1) = action createNewWorld
+
unsafePerformIO action = let (a, world1) = action createNewWorld
 
in a
 
in a
 
</haskell>
 
</haskell>
   
where 'createNewWorld' is internal function producing new value of
+
where 'createNewWorld' is an internal function producing a new value of
RealWorld type.
+
the RealWorld type.
   
Using unsafePerformIO, you can easily write pure functions that does
+
Using unsafePerformIO, you can easily write pure functions that do
I/O inside. But don't do this without real need, and remember to
+
I/O inside. But don't do this without a real need, and remember to
follow this rule: compiler don't know that you are cheating, it still
+
follow this rule: the compiler doesn't know that you are cheating; it still
consider each non-IO function as pure one. Therefore, all the usual
+
considers each non-IO function to be a pure one. Therefore, all the usual
optimization rules can (and will!) be applied to it's execution. So
+
optimization rules can (and will!) be applied to its execution. So
 
you must ensure that:
 
you must ensure that:
   
# Result of each call depends only on it's arguments
+
# The result of each call depends only on its arguments.
# You don't rely on side-effects of this function, which may be not executed if it's results are not used
+
# You don't rely on side-effects of this function, which may be not executed if its results are not needed.
   
   
Let's investigate this problem deeper. Function evaluation in Haskell
+
Let's investigate this problem more deeply. Function evaluation in Haskell
  +
is determined by a value's necessity - the language computes only the values that are really required to calculate the final result. But what does this mean with respect to the 'main' function? To "calculate the final world's" value, you need to perform all the intermediate IO actions that are included in the 'main' chain. By using 'unsafePerformIO' we call IO actions outside of this chain. What guarantee do we have that they will be run at all? None. The only time they will be run is if running them is required to compute the overall function result (which in turn should be required to perform some action in the
are ruled by value's necessity - computed only the values that really
 
  +
'main' chain). This is an example of Haskell's evaluation-by-need strategy. Now you should clearly see the difference:
required to calculate final result. But that this means according to
 
'main' function? To "calculate final world's" value, it's required to
 
perform all the intermediate IO actions that included in 'main' chain.
 
By using 'unsafePerformIO' we call IO actions outside of this chain.
 
What can guarantee that they will be run? Nothing. The only case when
 
they will be run is if that is required to compute overall function
 
result (that in turn should be required to perform some action in
 
'main' chain). Here we return to the Haskell-natural
 
evaluation-on-value-need. Now you should clearly see the difference:
 
   
- IO action inside IO procedure guaranteed to execute as long as
+
- An IO action inside an IO procedure is guaranteed to execute as long as
  +
it is (directly or indirectly) inside the 'main' chain - even when its result isn't used (because the implicit "world" value it returns ''will'' be used). You directly specify the order of the action's execution inside the IO procedure. Data dependencies are simulated via the implicit "world" values that are passed from each IO action to the next.
it is inside 'main' chain - even when it's result is not used.
 
You directly specify order of action's execution inside IO procedure.
 
Data dependencies are simulated via "world" values.
 
   
- IO action inside 'unsafePerformIO' will be performed only if
+
- An IO action inside 'unsafePerformIO' will be performed only if
result of this operation is really used. Evaluation order is not
+
result of this operation is really used. The evaluation order is not
guaranteed and you should not rely on it (except when you sure about
+
guaranteed and you should not rely on it (except when you're sure about
  +
whatever data dependencies may exist).
data dependency).
 
   
   
 
I should also say that inside 'unsafePerformIO' call you can organize
 
I should also say that inside 'unsafePerformIO' call you can organize
small internal chain of IO actions with help of the same binding
+
a small internal chain of IO actions with the help of the same binding
  +
operators and/or 'do' syntactic sugar we've seen above. For example, here's a particularly convoluted way to compute the integer that comes after zero:
operators and/or 'do' sugar:
 
   
 
<haskell>
 
<haskell>
  +
one :: Int
 
one = unsafePerformIO $ do var <- newIORef 0
 
one = unsafePerformIO $ do var <- newIORef 0
writeIORef var 1
+
modifyIORef var (+1)
 
readIORef var
 
readIORef var
 
</haskell>
 
</haskell>
   
 
and in this case ALL the operations in this chain will be performed as
 
and in this case ALL the operations in this chain will be performed as
long as 'unsafePerformIO' result will be demanded. To ensure this,
+
long as the result of the 'unsafePerformIO' call is needed. To ensure this,
the actual 'unsafePerformIO' implementation evaluates "world" returned
+
the actual 'unsafePerformIO' implementation evaluates the "world" returned
 
by the 'action':
 
by the 'action':
   
Line 1,071: Line 1,213:
 
</haskell>
 
</haskell>
   
('seq' operation strictly evaluates it's first argument before
+
(The 'seq' operation strictly evaluates its first argument before
returning the value of second one)
+
returning the value of the second one).
   
   
  +
=== inlinePerformIO ===
   
  +
inlinePerformIO has the same definition as unsafePerformIO but with addition of INLINE pragma:
But there is even more strange operation - 'unsafeInterleaveIO' that
 
  +
<haskell>
gets "official baton", makes it's piratical copy, and then run's
 
  +
-- | Just like unsafePerformIO, but we inline it. Big performance gains as
"illegal" relay-race in parallel with main one! I can't further say
 
  +
-- it exposes lots of things to further inlining
about it's behavior without grief and indignation, it's not surprise
 
  +
{-# INLINE inlinePerformIO #-}
that this operation is widely used in such software-piratical
 
  +
inlinePerformIO action = let (a, world1) = action createNewWorld
countries as Russia and China! ;) Don't even ask me - i will say
 
  +
in (world1 `seq` a)
nothing about this dirty trick i using permanently ;)
 
  +
#endif
  +
</haskell>
   
  +
Semantically inlinePerformIO = unsafePerformIO
 
  +
in as much as either of those have any semantics at all.
   
  +
The difference of course is that inlinePerformIO is even less safe than
== fixIO and 'mdo' ==
 
  +
unsafePerformIO. While ghc will try not to duplicate or common up
  +
different uses of unsafePerformIO, we aggressively inline
  +
inlinePerformIO. So you can really only use it where the IO content is
  +
really properly pure, like reading from an immutable memory buffer (as
  +
in the case of ByteStrings). However things like allocating new buffers
  +
should not be done inside inlinePerformIO since that can easily be
  +
floated out and performed just once for the whole program, so you end up
  +
with many things sharing the same buffer, which would be bad.
   
  +
So the rule of thumb is that IO things wrapped in unsafePerformIO have
== ST monad ==
 
  +
to be externally pure while with inlinePerformIO it has to be really
  +
really pure or it'll all go horribly wrong.
   
  +
That said, here's some really hairy code. This should frighten any pure
== Q monad ==
 
  +
functional programmer...
   
  +
<haskell>
== Welcome to machine: actual [[GHC]] implementation ==
 
  +
write :: Int -> (Ptr Word8 -> IO ()) -> Put ()
  +
write !n body = Put $ \c buf@(Buffer fp o u l) ->
  +
if n <= l
  +
then write' c fp o u l
  +
else write' (flushOld c n fp o u) (newBuffer c n) 0 0 0
  +
  +
where {-# NOINLINE write' #-}
  +
write' c !fp !o !u !l =
  +
-- warning: this is a tad hardcore
  +
inlinePerformIO
  +
(withForeignPtr fp
  +
(\p -> body $! (p `plusPtr` (o+u))))
  +
`seq` c () (Buffer fp o (u+n) (l-n))
  +
</haskell>
  +
  +
it's used like:
  +
<haskell>
  +
word8 w = write 1 (\p -> poke p w)
  +
</haskell>
  +
  +
This does not adhere to my rule of thumb above. Don't ask exactly why we
  +
claim it's safe :-) (and if anyone really wants to know, ask Ross
  +
Paterson who did it first in the Builder monoid)
  +
  +
=== unsafeInterleaveIO ===
  +
  +
But there is an even stranger operation called 'unsafeInterleaveIO' that
  +
gets the "official baton", makes its own pirate copy, and then runs
  +
an "illegal" relay-race in parallel with the main one! I can't talk further
  +
about its behavior without causing grief and indignation, so it's no surprise
  +
that this operation is widely used in countries that are hotbeds of software piracy such as Russia and China! ;) Don't even ask me - I won't say anything more about this dirty trick I use all the time ;)
  +
  +
One can use unsafePerformIO (not unsafeInterleaveIO) to perform I/O
  +
operations not in predefined order but by demand. For example, the
  +
following code:
  +
  +
<haskell>
  +
do let c = unsafePerformIO getChar
  +
do_proc c
  +
</haskell>
  +
  +
will perform getChar I/O call only when value of c is really required
  +
by code, i.e. it this call will be performed lazily as any usual
  +
Haskell computation.
  +
  +
Now imagine the following code:
  +
  +
<haskell>
  +
do let s = [unsafePerformIO getChar, unsafePerformIO getChar, unsafePerformIO getChar]
  +
do_proc s
  +
</haskell>
  +
  +
Three chars inside this list will be computed on demand too, and this
  +
means that their values will depend on the order they are consumed. It
  +
is not that we usually need.
  +
  +
  +
unsafeInterleaveIO solves this problem - it performs I/O only on
  +
demand but allows to define exact *internal* execution order for parts
  +
of your datastructure. It is why I wrote that unsafeInterleaveIO makes
  +
illegal copy of baton.
  +
  +
First, unsafeInterleaveIO has (IO a) action as a parameter and returns
  +
value of type 'a':
  +
  +
<haskell>
  +
do str <- unsafeInterleaveIO myGetContents
  +
</haskell>
  +
  +
Second, unsafeInterleaveIO don't perform any action immediately, it
  +
only creates a box of type 'a' which on requesting this value will
  +
perform action specified as a parameter.
  +
  +
Third, this action by itself may compute the whole value immediately
  +
or... use unsafeInterleaveIO again to defer calculation of some
  +
sub-components:
  +
  +
<haskell>
  +
myGetContents = do
  +
c <- getChar
  +
s <- unsafeInterleaveIO myGetContents
  +
return (c:s)
  +
</haskell>
  +
  +
This code will be executed only at the moment when value of str is
  +
really demanded. In this moment, getChar will be performed (with
  +
result assigned to c) and one more lazy IO box will be created - for s.
  +
This box again contains link to the myGetContents call
  +
  +
Then, list cell returned that contains one char read and link to
  +
myGetContents call as a way to compute rest of the list. Only at the
  +
moment when next value in list required, this operation will be
  +
performed again
  +
  +
As a final result, we get inability to read second char in list before
  +
first one, but lazy character of reading in whole. bingo!
  +
  +
  +
PS: of course, actual code should include EOF checking. also note that
  +
you can read many chars/records at each call:
  +
  +
<haskell>
  +
myGetContents = do
  +
c <- replicateM 512 getChar
  +
s <- unsafeInterleaveIO myGetContents
  +
return (c++s)
  +
</haskell>
  +
  +
== A safer approach: the ST monad ==
  +
  +
We said earlier that we can use unsafePerformIO to perform computations that are totally pure but nevertheless interact with the Real World in some way. There is, however, a better way! One that remains totally pure and yet allows the use of references, arrays, and so on -- and it's done using, you guessed it, type magic. This is the ST monad.
  +
  +
The ST monad's version of unsafePerformIO is called runST, and it has a very unusual type.
  +
<haskell>
  +
runST :: (forall s . ST s a) -> a
  +
</haskell>
  +
  +
The s variable in the ST monad is the state type. Moreover, all the fun mutable stuff available in the ST monad is quantified over s:
  +
<haskell>
  +
newSTRef :: a -> ST s (STRef s a)
  +
newArray_ :: Ix i => (i, i) -> ST s (STArray s i e)
  +
</haskell>
  +
  +
So why does runST have such a funky type? Let's see what would happen if we wrote
  +
<haskell>
  +
makeSTRef :: a -> STRef s a
  +
makeSTRef a = runST (newSTRef a)
  +
</haskell>
  +
This fails, because newSTRef a doesn't work for all state types s -- it only works for the s from the return type STRef s a.
  +
  +
This is all sort of wacky, but the result is that you can only run an ST computation where the output type is functionally pure, and makes no references to the internal mutable state of the computation. The ST monad doesn't have access to I/O operations like writing to the console, either -- only references, arrays, and suchlike that come in handy for pure computations.
  +
  +
Important note -- the state type doesn't actually mean anything. We never have a value of type s, for instance. It's just a way of getting the type system to do the work of ensuring purity for us, with smoke and mirrors.
  +
  +
It's really just type system magic: secretly, on the inside, runST runs a computation with the real world baton just like unsafePerformIO. Their internal implementations are almost identical: in fact, there's a function
  +
<haskell>
  +
stToIO :: ST RealWorld a -> IO a
  +
</haskell>
  +
  +
The difference is that ST uses type system magic to forbid unsafe behavior like extracting mutable objects from their safe ST wrapping, but allowing purely functional outputs to be performed with all the handy access to mutable references and arrays.
  +
  +
So here's how we'd rewrite our function using unsafePerformIO from above:
  +
  +
<haskell>
  +
oneST :: ST s Int -- note that this works correctly for any s
  +
oneST = do var <- newSTRef 0
  +
modifySTRef var (+1)
  +
readSTRef var
  +
  +
one :: Int
  +
one = runST oneST
  +
</haskell>
  +
  +
== Welcome to the machine: the actual [[GHC]] implementation ==
   
A little disclaimer: after all, i should say that i don't described
+
A little disclaimer: I should say that I'm not describing
here what is a monad (i even don't know it myself) and what my
+
here exactly what a monad is (I don't even completely understand it myself) and my explanation shows only one _possible_ way to implement the IO monad in
  +
Haskell. For example, the hbc Haskell compiler and the Hugs interpreter
explanations shows only the one _possible_ way to implement them in
 
  +
implements the IO monad via continuations. I also haven't said anything about
Haskell. For example, hbc Haskell compiler implements monads via
 
  +
exception handling, which is a natural part of the "monad" concept. You can
continuations. I also don't said anything about exception handling
 
  +
read the "All About Monads" guide to learn more about these topics.
that is natural part of "monad" concept. You can read "All about
 
monads" guide to learn more on these topics.
 
   
But there are a good news: first, monad understanding you've build
+
But there is some good news: first, the IO monad understanding you've just acquired will work with any implementation and with many other monads. You just can't work with RealWorld
will work with any implementation. You just can't work with RealWorld
 
 
values directly.
 
values directly.
   
Second, IO monad implementation described here is really used in GHC,
+
Second, the IO monad implementation described here is really used in the GHC,
yhc/nhc (Hugs/jhc, too?) compilers. It is the really real IO definition
+
yhc/nhc (jhc, too?) compilers. Here is the actual IO definition
from GHC sources:
+
from the GHC sources:
   
 
<haskell>
 
<haskell>
Line 1,114: Line 1,423:
 
</haskell>
 
</haskell>
   
It uses "State# RealWorld" type instead of our RealWorld, it uses "(# #)"
+
It uses the "State# RealWorld" type instead of our RealWorld, it uses the "(# #)" strict tuple for optimization, and it adds an IO data constructor
  +
around the type. Nevertheless, there are no significant changes from the standpoint of our explanation. Knowing the principle of "chaining" IO actions via fake "state of the world" values, you can now easily understand and write low-level implementations of GHC I/O operations.
strict tuple for optimization, and it adds IO data constructor
 
around the type. Nevertheless, there are no principal changes. Knowing
 
the principle of "chaining" IO actions via fake "state of world"
 
values, now you can easily understand and write low-level
 
implementations of GHC I/O operations.
 
   
   
Line 1,130: Line 1,435:
   
 
This implementation makes the "World" disappear somewhat, and returns Either a
 
This implementation makes the "World" disappear somewhat, and returns Either a
result "a", or if an error occurs then "IOError". The lack of the World on the
+
result of type "a", or if an error occurs then "IOError". The lack of the World on the right-hand side of the function can only be done because the compiler knows special things about the IO type, and won't overoptimise it.
right hand side of the function can only be done because the compiler knows
 
special things about the IO type, and will not over optimise it.
 
   
   
 
== Further reading ==
 
== Further reading ==
   
This tutorial is largely based on the Simon Peyton Jones' paper [http://research.microsoft.com/%7Esimonpj/Papers/marktoberdorf Tackling the awkward squad: monadic input/output, concurrency, exceptions, and foreign-language calls in Haskell]. You sould proceed to read it if you want to learn about concurrency, exceptions and FFI in Haskell/GHC.
+
[1] This tutorial is largely based on the Simon Peyton Jones' paper [http://research.microsoft.com/%7Esimonpj/Papers/marktoberdorf Tackling the awkward squad: monadic input/output, concurrency, exceptions, and foreign-language calls in Haskell]. I hope that my tutorial improves his original explanation of the Haskell I/O system and brings it closer to the point of view of beginning Haskell programmers. But if you need to learn about concurrency, exceptions and FFI in Haskell/GHC, the original paper is the best source of information.
   
  +
[2] You can find more information about concurrency, FFI and STM at the [[GHC/Concurrency#Starting points]] page.
Look also at the [[Books and tutorials#Using Monads]] page, which contains tutorials and papers really describing these mysterious monads :)
 
   
  +
[3] The [[Arrays]] page contains exhaustive explanations about using mutable arrays.
Are you have more questions? Ask in the [http://www.haskell.org/mailman/listinfo/haskell-cafe haskell-cafe mailing list].
 
  +
  +
[4] Look also at the [[Tutorials#Using_monads|Using monads]] page, which contains tutorials and papers really describing these mysterious monads.
  +
  +
[5] An explanation of the basic monad functions, with examples, can be found in the reference guide [http://members.chello.nl/hjgtuyl/tourdemonad.html A tour of the Haskell Monad functions], by Henk-Jan van Tuyl.
  +
  +
[6] Official FFI specifications can be found on the page [http://www.cse.unsw.edu.au/~chak/haskell/ffi/ The Haskell 98 Foreign Function Interface 1.0: An Addendum to the Haskell 98 Report]
  +
  +
[7] Using FFI in multithreaded programs described in paper [http://www.haskell.org/~simonmar/bib/concffi04_abstract.html Extending the Haskell Foreign Function Interface with Concurrency]
  +
  +
Do you have more questions? Ask in the [http://www.haskell.org/mailman/listinfo/haskell-cafe haskell-cafe mailing list].
  +
  +
== To-do list ==
  +
  +
If you are interested in adding more information to this manual, please add your questions/topics here.
  +
  +
Topics:
  +
* fixIO and 'mdo'
  +
* Q monad
  +
  +
Questions:
  +
* split '>>='/'>>'/return section and 'do' section, more examples of using binding operators
  +
* IORef detailed explanation (==const*), usage examples, syntax sugar, unboxed refs
  +
* explanation of how the actual data "in" mutable references are inside 'RealWorld', rather than inside the references themselves ('IORef','IOArray',&c.)
  +
* control structures developing - much more examples
  +
* unsafePerformIO usage examples: global variable, ByteString, other examples
  +
* how 'unsafeInterLeaveIO' can be seen as a kind of concurrency, and therefore isn't so unsafe (unlike 'unsafeInterleaveST' which really is unsafe)
  +
* discussion about different senses of "safe"/"unsafe" (like breaking equational reasoning vs. invoking undefined behaviour (so can corrupt the run-time system))
  +
* actual GHC implementation - how to write low-level routines on example of newIORef implementation
  +
  +
This manual is collective work, so feel free to add more information to it yourself. The final goal is to collectively develop a comprehensive manual for using the IO monad.
   
 
----
 
----
  +
  +
[[Category:Tutorials]]

Revision as of 07:24, 11 May 2015

Haskell I/O has always been a source of confusion and surprises for new Haskellers. While simple I/O code in Haskell looks very similar to its equivalents in imperative languages, attempts to write somewhat more complex code often result in a total mess. This is because Haskell I/O is really very different internally. Haskell is a pure language and even the I/O system can't break this purity.

The following text is an attempt to explain the details of Haskell I/O implementations. This explanation should help you eventually master all the smart I/O tricks. Moreover, I've added a detailed explanation of various traps you might encounter along the way. After reading this text, you will receive a "Master of Haskell I/O" degree that is equal to a Bachelor in Computer Science and Mathematics, simultaneously.

If you are new to Haskell I/O you may prefer to start by reading the Introduction to IO page.


Haskell is a pure language

Haskell is a pure language, which means that the result of any function call is fully determined by its arguments. Pseudo-functions like rand() or getchar() in C, which return different results on each call, are simply impossible to write in Haskell. Moreover, Haskell functions can't have side effects, which means that they can't effect any changes to the "real world", like changing files, writing to the screen, printing, sending data over the network, and so on. These two restrictions together mean that any function call can be replaced by the result of a previous call with the same parameters, and the language guarantees that all these rearrangements will not change the program result!

Let's compare this to C: optimizing C compilers try to guess which functions have no side effects and don't depend on mutable global variables. If this guess is wrong, an optimization can change the program's semantics! To avoid this kind of disaster, C optimizers are conservative in their guesses or require hints from the programmer about the purity of functions.

Compared to an optimizing C compiler, a Haskell compiler is a set of pure mathematical transformations. This results in much better high-level optimization facilities. Moreover, pure mathematical computations can be much more easily divided into several threads that may be executed in parallel, which is increasingly important in these days of multi-core CPUs. Finally, pure computations are less error-prone and easier to verify, which adds to Haskell's robustness and to the speed of program development using Haskell.

Haskell purity allows compiler to call only functions whose results are really required to calculate final value of high-level function (i.e., main) - this is called lazy evaluation. It's great thing for pure mathematical computations, but how about I/O actions? Function like:

putStrLn "Press any key to begin formatting"

can't return any meaningful result value, so how can we ensure that compiler will not omit or reorder its execution? And in general: how we can work with stateful algorithms and side effects in an entirely lazy language? This question has had many different solutions proposed in 18 years of Haskell development (see History of Haskell), though a solution based on monads is now the standard.

What is a monad?

What is a monad? It's something from mathematical category theory, which I don't know anymore. In order to understand how monads are used to solve the problem of I/O and side effects, you don't need to know it. It's enough to just know elementary mathematics, like I do.

Let's imagine that we want to implement in Haskell the well-known 'getchar' function. What type should it have? Let's try:

getchar :: Char

get2chars = [getchar,getchar]

What will we get with 'getchar' having just the 'Char' type? You can see all the possible problems in the definition of 'get2chars':

  1. Because the Haskell compiler treats all functions as pure (not having side effects), it can avoid "excessive" calls to 'getchar' and use one returned value twice.
  2. Even if it does make two calls, there is no way to determine which call should be performed first. Do you want to return the two chars in the order in which they were read, or in the opposite order? Nothing in the definition of 'get2chars' answers this question.

How can these problems be solved, from the programmer's viewpoint? Let's introduce a fake parameter of 'getchar' to make each call "different" from the compiler's point of view:

getchar :: Int -> Char

get2chars = [getchar 1, getchar 2]

Right away, this solves the first problem mentioned above - now the compiler will make two calls because it sees them as having different parameters. The whole 'get2chars' function should also have a fake parameter, otherwise we will have the same problem calling it:

getchar   :: Int -> Char
get2chars :: Int -> String

get2chars _ = [getchar 1, getchar 2]


Now we need to give the compiler some clue to determine which function it should call first. The Haskell language doesn't provide any way to express order of evaluation... except for data dependencies! How about adding an artificial data dependency which prevents evaluation of the second 'getchar' before the first one? In order to achieve this, we will return an additional fake result from 'getchar' that will be used as a parameter for the next 'getchar' call:

getchar :: Int -> (Char, Int)

get2chars _ = [a,b]  where (a,i) = getchar 1
                           (b,_) = getchar i

So far so good - now we can guarantee that 'a' is read before 'b' because reading 'b' needs the value ('i') that is returned by reading 'a'!

We've added a fake parameter to 'get2chars' but the problem is that the Haskell compiler is too smart! It can believe that the external 'getchar' function is really dependent on its parameter but for 'get2chars' it will see that we're just cheating because we throw it away! Therefore it won't feel obliged to execute the calls in the order we want. How can we fix this? How about passing this fake parameter to the 'getchar' function?! In this case the compiler can't guess that it is really unused.

get2chars i0 = [a,b]  where (a,i1) = getchar i0
                            (b,i2) = getchar i1


And more - 'get2chars' has all the same purity problems as the 'getchar' function. If you need to call it two times, you need a way to describe the order of these calls. Look at:

get4chars = [get2chars 1, get2chars 2]  -- order of 'get2chars' calls isn't defined

We already know how to deal with these problems - 'get2chars' should also return some fake value that can be used to order calls:

get2chars :: Int -> (String, Int)

get4chars i0 = (a++b)  where (a,i1) = get2chars i0
                             (b,i2) = get2chars i1


But what's the fake value 'get2chars' should return? If we use some integer constant, the excessively-smart Haskell compiler will guess that we're cheating again. What about returning the value returned by 'getchar'? See:

get2chars :: Int -> (String, Int)
get2chars i0 = ([a,b], i2)  where (a,i1) = getchar i0
                                  (b,i2) = getchar i1

Believe it or not, but we've just constructed the whole "monadic" Haskell I/O system.

Welcome to the RealWorld, baby

Warning: The following story about IO is incorrect in that it cannot actually explain some important aspects of IO (including interaction and concurrency). However, some people find it useful to begin developing an understanding.

The 'main' Haskell function has the type:

main :: RealWorld -> ((), RealWorld)

where 'RealWorld' is a fake type used instead of our Int. It's something like the baton passed in a relay race. When 'main' calls some IO function, it passes the "RealWorld" it received as a parameter. All IO functions have similar types involving RealWorld as a parameter and result. To be exact, "IO" is a type synonym defined in the following way:

type IO a  =  RealWorld -> (a, RealWorld)

So, 'main' just has type "IO ()", 'getChar' has type "IO Char" and so on. You can think of the type "IO Char" as meaning "take the current RealWorld, do something to it, and return a Char and a (possibly changed) RealWorld". Let's look at 'main' calling 'getChar' two times:

getChar :: RealWorld -> (Char, RealWorld)

main :: RealWorld -> ((), RealWorld)
main world0 = let (a, world1) = getChar world0
                  (b, world2) = getChar world1
              in ((), world2)


Look at this closely: 'main' passes the "world" it received to the first 'getChar'. This 'getChar' returns some new value of type RealWorld that gets used in the next call. Finally, 'main' returns the "world" it got from the second 'getChar'.

  1. Is it possible here to omit any call of 'getChar' if the Char it read is not used? No, because we need to return the "world" that is the result of the second 'getChar' and this in turn requires the "world" returned from the first 'getChar'.
  2. Is it possible to reorder the 'getChar' calls? No: the second 'getChar' can't be called before the first one because it uses the "world" returned from the first call.
  3. Is it possible to duplicate calls? In Haskell semantics - yes, but real compilers never duplicate work in such simple cases (otherwise, the programs generated will not have any speed guarantees).


As we already said, RealWorld values are used like a baton which gets passed between all routines called by 'main' in strict order. Inside each routine called, RealWorld values are used in the same way. Overall, in order to "compute" the world to be returned from 'main', we should perform each IO procedure that is called from 'main', directly or indirectly. This means that each procedure inserted in the chain will be performed just at the moment (relative to the other IO actions) when we intended it to be called. Let's consider the following program:

main = do a <- ask "What is your name?"
          b <- ask "How old are you?"
          return ()

ask s = do putStr s
           readLn

Now you have enough knowledge to rewrite it in a low-level way and check that each operation that should be performed will really be performed with the arguments it should have and in the order we expect.


But what about conditional execution? No problem. Let's define the well-known 'when' operation:

when :: Bool -> IO () -> IO ()
when condition action world =
    if condition
      then action world
      else ((), world)

As you can see, we can easily include or exclude from the execution chain IO procedures (actions) depending on the data values. If 'condition' will be False on the call of 'when', 'action' will never be called because real Haskell compilers, again, never call functions whose results are not required to calculate the final result (i.e., here, the final "world" value of 'main').

Loops and more complex control structures can be implemented in the same way. Try it as an exercise!


Finally, you may want to know how much passing these RealWorld values around the program costs. It's free! These fake values exist solely for the compiler while it analyzes and optimizes the code, but when it gets to assembly code generation, it "suddenly" realize that this type is like "()", so all these parameters and result values can be omitted from the final generated code. Isn't it beautiful?

'>>=' and 'do' notation

All beginners (including me) start by thinking that 'do' is some magic statement that executes IO actions. That's wrong - 'do' is just syntactic sugar that simplifies the writing of procedures that use IO (and also other monads, but that's beyond the scope of this tutorial). 'do' notation eventually gets translated to statements passing "world" values around like we've manually written above and is used to simplify the gluing of several IO actions together. You don't need to use 'do' for just one statement; for instance,

  main = do putStr "Hello!"

is desugared to:

  main = putStr "Hello!"


Let's examine how to desugar a 'do' with multiple statements in the following example:

main = do putStr "What is your name?"
          putStr "How old are you?"
          putStr "Nice day!"

The 'do' statement here just joins several IO actions that should be performed sequentially. It's translated to sequential applications of one of the so-called "binding operators", namely '>>':

main = (putStr "What is your name?")
       >> ( (putStr "How old are you?")
            >> (putStr "Nice day!")
          )

This binding operator just combines two IO actions, executing them sequentially by passing the "world" between them:

(>>) :: IO a -> IO b -> IO b
(action1 >> action2) world0 =
   let (a, world1) = action1 world0
       (b, world2) = action2 world1
   in (b, world2)

If defining operators this way looks strange to you, read this definition as follows:

action1 >> action2 = action
  where
    action world0 = let (a, world1) = action1 world0
                        (b, world2) = action2 world1
                    in (b, world2)

Now you can substitute the definition of '>>' at the places of its usage and check that program constructed by the 'do' desugaring is actually the same as we could write by manually manipulating "world" values.


A more complex example involves the binding of variables using "<-":

main = do a <- readLn
          print a

This code is desugared into:

main = readLn
       >>= (\a -> print a)

As you should remember, the '>>' binding operator silently ignores the value of its first action and returns as an overall result the result of its second action only. On the other hand, the '>>=' binding operator (note the extra '=' at the end) allows us to use the result of its first action - it gets passed as an additional parameter to the second one! Look at the definition:

(>>=) :: IO a -> (a -> IO b) -> IO b
(action1 >>= action2) world0 =
   let (a, world1) = action1 world0
       (b, world2) = action2 a world1
   in (b, world2)

First, what does the type of the second "action" (more precisely, a function which returns an IO action), namely "a -> IO b", mean? By substituting the "IO" definition, we get "a -> RealWorld -> (b, RealWorld)". This means that second action actually has two parameters - the type 'a' actually used inside it, and the value of type RealWorld used for sequencing of IO actions. That's always the case - any IO procedure has one more parameter compared to what you see in its type signature. This parameter is hidden inside the definition of the type alias "IO".

Second, you can use these '>>' and '>>=' operations to simplify your program. For example, in the code above we don't need to introduce the variable, because the result of 'readLn' can be send directly to 'print':

main = readLn >>= print


And third - as you see, the notation:

 do x <- action1
    action2

where 'action1' has type "IO a" and 'action2' has type "IO b", translates into:

 action1 >>= (\x -> action2)

where the second argument of '>>=' has the type "a -> IO b". It's the way the '<-' binding is processed - the name on the left-hand side of '<-' just becomes a parameter of subsequent operations represented as one large IO action. Note also that if 'action1' has type "IO a" then 'x' will just have type "a"; you can think of the effect of '<-' as "unpacking" the IO value of 'action1' into 'x'. Note also that '<-' is not a true operator; it's pure syntax, just like 'do' itself. Its meaning results only from the way it gets desugared.

Look at the next example:

main = do putStr "What is your name?"
          a <- readLn
          putStr "How old are you?"
          b <- readLn
          print (a,b)

This code is desugared into:

main = putStr "What is your name?"
       >> readLn
       >>= \a -> putStr "How old are you?"
       >> readLn
       >>= \b -> print (a,b)

I omitted the parentheses here; both the '>>' and the '>>=' operators are left-associative, but lambda-bindings always stretches as far to the right as possible, which means that the 'a' and 'b' bindings introduced here are valid for all remaining actions. As an exercise, add the parentheses yourself and translate this procedure into the low-level code that explicitly passes "world" values. I think it should be enough to help you finally realize how the 'do' translation and binding operators work.


Oh, no! I forgot the third monadic operator - 'return'. It just combines its two parameters - the value passed and "world":

return :: a -> IO a
return a world0  =  (a, world0)

How about translating a simple example of 'return' usage? Say,

main = do a <- readLn
          return (a*2)


Programmers with an imperative language background often think that 'return' in Haskell, as in other languages, immediately returns from the IO procedure. As you can see in its definition (and even just from its type!), such an assumption is totally wrong. The only purpose of using 'return' is to "lift" some value (of type 'a') into the result of a whole action (of type "IO a") and therefore it should generally be used only as the last executed statement of some IO sequence. For example try to translate the following procedure into the corresponding low-level code:

main = do a <- readLn
          when (a>=0) $ do
              return ()
          print "a is negative"

and you will realize that the 'print' statement is executed even for non-negative values of 'a'. If you need to escape from the middle of an IO procedure, you can use the 'if' statement:

main = do a <- readLn
          if (a>=0)
            then return ()
            else print "a is negative"

Moreover, Haskell layout rules allow us to use the following layout:

main = do a <- readLn
          if (a>=0) then return ()
            else do
          print "a is negative"
          ...

that may be useful for escaping from the middle of a longish 'do' statement.


Last exercise: implement a function 'liftM' that lifts operations on plain values to the operations on monadic ones. Its type signature:

liftM :: (a -> b) -> (IO a -> IO b)

If that's too hard for you, start with the following high-level definition and rewrite it in low-level fashion:

liftM f action = do x <- action
                    return (f x)

Mutable data (references, arrays, hash tables...)

As you should know, every name in Haskell is bound to one fixed (immutable) value. This greatly simplifies understanding algorithms and code optimization, but it's inappropriate in some cases. As we all know, there are plenty of algorithms that are simpler to implement in terms of updatable variables, arrays and so on. This means that the value associated with a variable, for example, can be different at different execution points, so reading its value can't be considered as a pure function. Imagine, for example, the following code:

main = do let a0 = readVariable varA
              _  = writeVariable varA 1
              a1 = readVariable varA
          print (a0, a1)

Does this look strange? First, the two calls to 'readVariable' look the same, so the compiler can just reuse the value returned by the first call. Second, the result of the 'writeVariable' call isn't used so the compiler can (and will!) omit this call completely. To complete the picture, these three calls may be rearranged in any order because they appear to be independent of each other. This is obviously not what was intended. What's the solution? You already know this - use IO actions! Using IO actions guarantees that:

  1. the execution order will be retained as written
  2. each action will have to be executed
  3. the result of the "same" action (such as "readVariable varA") will not be reused

So, the code above really should be written as:

import Data.IORef
main = do varA <- newIORef 0  -- Create and initialize a new variable
          a0 <- readIORef varA
          writeIORef varA 1
          a1 <- readIORef varA
          print (a0, a1)

Here, 'varA' has the type "IORef Int" which means "a variable (reference) in the IO monad holding a value of type Int". newIORef creates a new variable (reference) and returns it, and then read/write actions use this reference. The value returned by the "readIORef varA" action depends not only on the variable involved but also on the moment this operation is performed so it can return different values on each call.

Arrays, hash tables and any other _mutable_ data structures are defined in the same way - for each of them, there's an operation that creates new "mutable values" and returns a reference to it. Then special read and write operations in the IO monad are used. The following code shows an example using mutable arrays:

import Data.Array.IO
main = do arr <- newArray (1,10) 37 :: IO (IOArray Int Int)
          a <- readArray arr 1
          writeArray arr 1 64
          b <- readArray arr 1
          print (a, b)

Here, an array of 10 elements with 37 as the initial value at each location is created. After reading the value of the first element (index 1) into 'a' this element's value is changed to 64 and then read again into 'b'. As you can see by executing this code, 'a' will be set to 37 and 'b' to 64.


Other state-dependent operations are also often implemented as IO actions. For example, a random number generator should return a different value on each call. It looks natural to give it a type involving IO:

rand :: IO Int

Moreover, when you import C routines you should be careful - if this routine is impure, i.e. its result depends on something in the "real world" (file system, memory contents...), internal state and so on, you should give it an IO type. Otherwise, the compiler can "optimize" repetitive calls of this procedure with the same parameters!

For example, we can write a non-IO type for:

foreign import ccall
   sin :: Double -> Double

because the result of 'sin' depends only on its argument, but

foreign import ccall
   tell :: Int -> IO Int

If you will declare 'tell' as a pure function (without IO) then you may get the same position on each call!

IO actions as values

By this point you should understand why it's impossible to use IO actions inside non-IO (pure) procedures. Such procedures just don't get a "baton"; they don't know any "world" value to pass to an IO action. The RealWorld type is an abstract datatype, so pure functions also can't construct RealWorld values by themselves, and it's a strict type, so 'undefined' also can't be used. So, the prohibition of using IO actions inside pure procedures is just a type system trick (as it usually is in Haskell).

But while pure code can't _execute_ IO actions, it can work with them as with any other functional values - they can be stored in data structures, passed as parameters, returned as results, collected in lists, and partially applied. But an IO action will remain a functional value because we can't apply it to the last argument - of type RealWorld.

In order to _execute_ the IO action we need to apply it to some RealWorld value. That can be done only inside some IO procedure, in its "actions chain". And real execution of this action will take place only when this procedure is called as part of the process of "calculating the final value of world" for 'main'. Look at this example:

main world0 = let get2chars = getChar >> getChar
                  ((), world1) = putStr "Press two keys" world0
                  (answer, world2) = get2chars world1
              in ((), world2)

Here we first bind a value to 'get2chars' and then write a binding involving 'putStr'. But what's the execution order? It's not defined by the order of the 'let' bindings, it's defined by the order of processing "world" values! You can arbitrarily reorder the binding statements - the execution order will be defined by the data dependency with respect to the "world" values that get passed around. Let's see what this 'main' looks like in the 'do' notation:

main = do let get2chars = getChar >> getChar
          putStr "Press two keys"
          get2chars
          return ()

As you can see, we've eliminated two of the 'let' bindings and left only the one defining 'get2chars'. The non-'let' statements are executed in the exact order in which they're written, because they pass the "world" value from statement to statement as we described above. Thus, this version of the function is much easier to understand because we don't have to mentally figure out the data dependency of the "world" value.

Moreover, IO actions like 'get2chars' can't be executed directly because they are functions with a RealWorld parameter. To execute them, we need to supply the RealWorld parameter, i.e. insert them in the 'main' chain, placing them in some 'do' sequence executed from 'main' (either directly in the 'main' function, or indirectly in an IO function called from 'main'). Until that's done, they will remain like any function, in partially evaluated form. And we can work with IO actions as with any other functions - bind them to names (as we did above), save them in data structures, pass them as function parameters and return them as results - and they won't be performed until you give them the magic RealWorld parameter!


Example: a list of IO actions

Let's try defining a list of IO actions:

ioActions :: [IO ()]
ioActions = [(print "Hello!"),
             (putStr "just kidding"),
             (getChar >> return ())
            ]

I used additional parentheses around each action, although they aren't really required. If you still can't believe that these actions won't be executed immediately, just recall the real type of this list:

ioActions :: [RealWorld -> ((), RealWorld)]

Well, now we want to execute some of these actions. No problem, just insert them into the 'main' chain:

main = do head ioActions
          ioActions !! 1
          last ioActions

Looks strange, right? Really, any IO action that you write in a 'do' statement (or use as a parameter for the '>>'/'>>=' operators) is an expression returning a result of type 'IO a' for some type 'a'. Typically, you use some function that has the type 'x -> y -> ... -> IO a' and provide all the x, y, etc. parameters. But you're not limited to this standard scenario - don't forget that Haskell is a functional language and you're free to compute the functional value required (recall that "IO a" is really a function type) in any possible way. Here we just extracted several functions from the list - no problem. This functional value can also be constructed on-the-fly, as we've done in the previous example - that's also OK. Want to see this functional value passed as a parameter? Just look at the definition of 'when'. Hey, we can buy, sell, and rent these IO actions just like we can with any other functional values! For example, let's define a function that executes all the IO actions in the list:

sequence_ :: [IO a] -> IO ()
sequence_ [] = return ()
sequence_ (x:xs) = do x
                      sequence_ xs

No black magic - we just extract IO actions from the list and insert them into a chain of IO operations that should be performed one after another (in the same order that they occurred in the list) to "compute the final world value" of the entire 'sequence_' call.

With the help of 'sequence_', we can rewrite our last 'main' function as:

main = sequence_ ioActions


Haskell's ability to work with IO actions as with any other (functional and non-functional) values allows us to define control structures of arbitrary complexity. Try, for example, to define a control structure that repeats an action until it returns the 'False' result:

while :: IO Bool -> IO ()
while action = ???

Most programming languages don't allow you to define control structures at all, and those that do often require you to use a macro-expansion system. In Haskell, control structures are just trivial functions anyone can write.


Example: returning an IO action as a result

How about returning an IO action as the result of a function? Well, we've done this each time we've defined an IO procedure - they all return IO actions that need a RealWorld value to be performed. While we usually just execute them as part of a higher-level IO procedure, it's also possible to just collect them without actual execution:

main = do let a = sequence ioActions
              b = when True getChar
              c = getChar >> getChar
          putStr "These 'let' statements are not executed!"

These assigned IO procedures can be used as parameters to other procedures, or written to global variables, or processed in some other way, or just executed later, as we did in the example with 'get2chars'.

But how about returning a parameterized IO action from an IO procedure? Let's define a procedure that returns the i'th byte from a file represented as a Handle:

readi h i = do hSeek h AbsoluteSeek i
               hGetChar h

So far so good. But how about a procedure that returns the i'th byte of a file with a given name without reopening it each time?

readfilei :: String -> IO (Integer -> IO Char)
readfilei name = do h <- openFile name ReadMode
                    return (readi h)

As you can see, it's an IO procedure that opens a file and returns... another IO procedure that will read the specified byte. But we can go further and include the 'readi' body in 'readfilei':

readfilei name = do h <- openFile name ReadMode
                    let readi h i = do hSeek h AbsoluteSeek i
                                       hGetChar h
                    return (readi h)

That's a little better. But why do we add 'h' as a parameter to 'readi' if it can be obtained from the environment where 'readi' is now defined? An even shorter version is this:

readfilei name = do h <- openFile name ReadMode
                    let readi i = do hSeek h AbsoluteSeek i
                                     hGetChar h
                    return readi

What have we done here? We've build a parameterized IO action involving local names inside 'readfilei' and returned it as the result. Now it can be used in the following way:

main = do myfile <- readfilei "test"
          a <- myfile 0
          b <- myfile 1
          print (a,b)


This way of using IO actions is very typical for Haskell programs - you just construct one or more IO actions that you need, with or without parameters, possibly involving the parameters that your "constructor" received, and return them to the caller. Then these IO actions can be used in the rest of the program without any knowledge about your internal implementation strategy. One thing this can be used for is to partially emulate the OOP (or more precisely, the ADT) programming paradigm.

Example: a memory allocator generator

As an example, one of my programs has a module which is a memory suballocator. It receives the address and size of a large memory block and returns two procedures - one to allocate a subblock of a given size and the other to free the allocated subblock:

memoryAllocator :: Ptr a -> Int -> IO (Int -> IO (Ptr b),
                                       Ptr c -> IO ())

memoryAllocator buf size = do ......
                              let alloc size = do ...
                                                  ...
                                  free ptr = do ...
                                                ...
                              return (alloc, free)

How this is implemented? 'alloc' and 'free' work with references created inside the memoryAllocator procedure. Because the creation of these references is a part of the memoryAllocator IO actions chain, a new independent set of references will be created for each memory block for which memoryAllocator is called:

memoryAllocator buf size = do start <- newIORef buf
                              end <- newIORef (buf `plusPtr` size)
                              ...

These two references are read and written in the 'alloc' and 'free' definitions (we'll implement a very simple memory allocator for this example):

      ...
      let alloc size = do addr <- readIORef start
                          writeIORef start (addr `plusPtr` size)
                          return addr
                          
      let free ptr = do writeIORef start ptr

What we've defined here is just a pair of closures that use state available at the moment of their definition. As you can see, it's as easy as in any other functional language, despite Haskell's lack of direct support for impure functions.

The following example uses procedures, returned by memoryAllocator, to simultaneously allocate/free blocks in two independent memory buffers:

main = do buf1 <- mallocBytes (2^16)
          buf2 <- mallocBytes (2^20)
          (alloc1, free1) <- memoryAllocator buf1 (2^16)
          (alloc2, free2) <- memoryAllocator buf2 (2^20)
          ptr11 <- alloc1 100
          ptr21 <- alloc2 1000
          free1 ptr11
          free2 ptr21
          ptr12 <- alloc1 100
          ptr22 <- alloc2 1000


Example: emulating OOP with record types

Let's implement the classical OOP example: drawing figures. There are figures of different types: circles, rectangles and so on. The task is to create a heterogeneous list of figures. All figures in this list should support the same set of operations: draw, move and so on. We will represent these operations as IO procedures. Instead of a "class" let's define a structure containing implementations of all the procedures required:

data Figure = Figure { draw :: IO (),
                       move :: Displacement -> IO ()
                     }

type Displacement = (Int, Int)  -- horizontal and vertical displacement in points


The constructor of each figure's type should just return a Figure record:

circle    :: Point -> Radius -> IO Figure
rectangle :: Point -> Point -> IO Figure

type Point = (Int, Int)  -- point coordinates
type Radius = Int        -- circle radius in points


We will "draw" figures by just printing their current parameters. Let's start with a simplified implementation of the 'circle' and 'rectangle' constructors, without actual 'move' support:

circle center radius = do
    let description = "  Circle at "++show center++" with radius "++show radius
    return $ Figure { draw = putStrLn description }

rectangle from to = do
    let description = "  Rectangle "++show from++"-"++show to)
    return $ Figure { draw = putStrLn description }


As you see, each constructor just returns a fixed 'draw' procedure that prints parameters with which the concrete figure was created. Let's test it:

drawAll :: [Figure] -> IO ()
drawAll figures = do putStrLn "Drawing figures:"
                     mapM_ draw figures

main = do figures <- sequence [circle (10,10) 5,
                               circle (20,20) 3,
                               rectangle (10,10) (20,20),
                               rectangle (15,15) (40,40)]
          drawAll figures


Now let's define "full-featured" figures that can actually be moved around. In order to achieve this, we should provide each figure with a mutable variable that holds each figure's current screen location. The type of this variable will be "IORef Point". This variable should be created in the figure constructor and manipulated in IO procedures (closures) enclosed in the Figure record:

circle center radius = do
    centerVar <- newIORef center
    
    let drawF = do center <- readIORef centerVar
                   putStrLn ("  Circle at "++show center
                             ++" with radius "++show radius)
                   
    let moveF (addX,addY) = do (x,y) <- readIORef centerVar
                               writeIORef centerVar (x+addX, y+addY)
                               
    return $ Figure { draw=drawF, move=moveF }

    
rectangle from to = do
    fromVar <- newIORef from
    toVar   <- newIORef to

    let drawF = do from <- readIORef fromVar
                   to   <- readIORef toVar
                   putStrLn ("  Rectangle "++show from++"-"++show to)
                   
    let moveF (addX,addY) = do (fromX,fromY) <- readIORef fromVar
                               (toX,toY)     <- readIORef toVar
                               writeIORef fromVar (fromX+addX, fromY+addY)
                               writeIORef toVar   (toX+addX, toY+addY)

    return $ Figure { draw=drawF, move=moveF }


Now we can test the code which moves figures around:

main = do figures <- sequence [circle (10,10) 5,
                               rectangle (10,10) (20,20)]
          drawAll figures
          mapM_ (\fig -> move fig (10,10)) figures
          drawAll figures


It's important to realize that we are not limited to including only IO actions in a record that's intended to simulate a C++/Java-style interface. The record can also include values, IORefs, pure functions - in short, any type of data. For example, we can easily add to the Figure interface fields for area and origin:

data Figure = Figure { draw :: IO (),
                       move :: Displacement -> IO (),
                       area :: Double,
                       origin :: IORef Point
                     }


Exception handling (under development)

Although Haskell provides a set of exception raising/handling features comparable to those in popular OOP languages (C++, Java, C#), this part of the language receives much less attention. This is for two reasons. First, you just don't need to worry as much about them - most of the time it just works "behind the scenes". The second reason is that Haskell, lacking OOP inheritance, doesn't allow the programmer to easily subclass exception types, therefore limiting flexibility of exception handling.

The Haskell RTS raises more exceptions than traditional languages - pattern match failures, calls with invalid arguments (such as head []) and computations whose results depend on special values undefined and error "...." all raise their own exceptions:

example 1:

main = print (f 2)

f 0 = "zero"
f 1 = "one"

example 2:

main = print (head [])

example 3:

main = print (1 + (error "Value that wasn't initialized or cannot be computed"))

This allows to write programs in much more error-prone way.

Interfacing with C/C++ and foreign libraries (under development)

While Haskell is great at algorithm development, speed isn't its best side. We can combine the best of both worlds, though, by writing speed-critical parts of program in C and rest in Haskell. We just need a way to call C functions from Haskell and vice versa, and to marshal data between two worlds.

We also need to interact with C world for using Windows/Linux APIs, linking to various libraries and DLLs. Even interfacing with other languages requires to go through C world as "common denominator". Chapter 8 of the Haskell 2010 report provides a complete description of interfacing with C.

We will learn FFI via a series of examples. These examples include C/C++ code, so they need C/C++ compilers to be installed, the same will be true if you need to include code written in C/C++ in your program (C/C++ compilers are not required when you just need to link with existing libraries providing APIs with C calling convention). On Unix (and Mac OS?) systems, the system-wide default C/C++ compiler is typically used by GHC installation. On Windows, no default compilers exist, so GHC is typically shipped with a C compiler, and you may find on the download page a GHC distribution bundled with C and C++ compilers. Alternatively, you may find and install a GCC/MinGW version compatible with your GHC installation.

If you need to make your C/C++ code as fast as possible, you may compile your code by Intel compilers instead of GCC. However, these compilers are not free, moreover on Windows, code compiled by Intel compilers may not interact correctly with GHC-compiled code, unless one of them is put into DLLs (due to object file incompatibility).

More links:

C->Haskell
A lightweight tool for implementing access to C libraries from Haskell.
HSFFIG
Haskell FFI Binding Modules Generator (HSFFIG) is a tool that takes a C library include file (.h) and generates Haskell Foreign Functions Interface import declarations for items (functions, structures, etc.) the header defines.
MissingPy
MissingPy is really two libraries in one. At its lowest level, MissingPy is a library designed to make it easy to call into Python from Haskell. It provides full support for interpreting arbitrary Python code, interfacing with a good part of the Python/C API, and handling Python objects. It also provides tools for converting between Python objects and their Haskell equivalents. Memory management is handled for you, and Python exceptions get mapped to Haskell Dynamic exceptions. At a higher level, MissingPy contains Haskell interfaces to some Python modules.
HsLua
A Haskell interface to the Lua scripting language


Calling functions

First, we will learn how to call C functions from Haskell and Haskell functions from C. The first example consists of three files:

main.hs:

{-# LANGUAGE ForeignFunctionInterface #-}

main = do print "Hello from main"
          c_function

haskell_function = print "Hello from haskell_function"

foreign import ccall safe "prototypes.h"
    c_function :: IO ()

foreign export ccall
    haskell_function :: IO ()

evil.c:

#include <stdio.h>
#include "prototypes.h"

void c_function (void)
{
  printf("Hello from c_function\n");
  haskell_function();
}

prototypes.h:

extern void c_function (void);
extern void haskell_function (void);

It may be compiled and linked in one step by ghc:

 ghc --make main.hs evil.c

Or, you may compile C module(s) separately and link in .o files (this may be preferable if you use make and don't want to recompile unchanged sources; ghc's --make option provides smart recompilation only for .hs files):

 ghc -c evil.c
 ghc --make main.hs evil.o

You may use gcc/g++ directly to compile your C/C++ files but I recommend to do linking via ghc because it adds a lot of libraries required for execution of Haskell code. For the same reason, even if your main routine is written in C/C++, I recommend calling it from the Haskell function main - otherwise you'll have to explicitly init/shutdown the GHC RTS (run-time system).

We use the "foreign import" specification to import foreign routines into our Haskell world, and "foreign export" to export Haskell routines into the external world. Note that the import statement creates a new Haskell symbol (from the external one), while the export statement uses a Haskell symbol previously defined. Technically speaking, both types of statements create a wrapper that converts the names and calling conventions from C to Haskell or vice versa.

All about the "foreign" statement

The "ccall" specifier in foreign statements means the use of C (not C++ !) calling convention. This means that if you want to write the external function in C++ (instead of C) you should add export "C" specification to its declaration - otherwise you'll get linking errors. Let's rewrite our first example to use C++ instead of C:

prototypes.h:

#ifdef __cplusplus
extern "C" {
#endif

extern void c_function (void);
extern void haskell_function (void);

#ifdef __cplusplus
}
#endif

Compile it via:

 ghc --make main.hs evil.cpp

where evil.cpp is just a renamed copy of evil.c from the first example. Note that the new prototypes.h is written to allow compiling it both as C and C++ code. When it's included from evil.cpp, it's compiled as C++ code. When GHC compiles main.hs via the C compiler (enabled by -fvia-C option), it also includes prototypes.h but compiles it in C mode. It's why you need to specify .h files in "foreign" declarations - depending on which Haskell compiler you use, these files may be included to check consistency of C and Haskell declarations.

The quoted part of the foreign statement may also be used to import or export a function under another name--for example,

foreign import ccall safe "prototypes.h CFunction"
    c_function :: IO ()

foreign export ccall "HaskellFunction"
    haskell_function :: IO ()

specifies that the C function called CFunction will become known as the Haskell function c_function, while the Haskell function haskell_function will be known in the C world as HaskellFunction. It's required when the C name doesn't conform to Haskell naming requirements.

Although the Haskell FFI standard tells about many other calling conventions in addition to ccall (e.g. cplusplus, jvm, net) current Haskell implementations support only ccall and stdcall. The latter, also called the "Pascal" calling convention, is used to interface with WinAPI:

foreign import stdcall unsafe "windows.h SetFileApisToOEM"
  setFileApisToOEM :: IO ()

And finally, about the safe/unsafe specifier: a C function imported with the "unsafe" keyword is called directly and the Haskell runtime is stopped while the C function is executed (when there are several OS threads executing the Haskell program, only the current OS thread is delayed). This call doesn't allow recursively entering into the Haskell world by calling any Haskell function - the Haskell RTS is just not prepared for such an event. However, unsafe calls are as quick as calls in C world. It's ideal for "momentary" calls that quickly return back to the caller.

When "safe" is specified, the C function is called in safe environment - the Haskell execution context is saved, so it's possible to call back to Haskell and, if the C call takes a long time, another OS thread may be started to execute Haskell code (of course, in threads other than the one that called the C code). This has its own price, though - around 1000 CPU ticks per call.

You can read more about interaction between FFI calls and Haskell concurrency in [7].

Marshalling simple types

Calling by itself is relatively easy; the real problem of interfacing languages with different data models is passing data between them. In this case, there is no guarantee that Haskell's Int is represented in memory the same way as C's int, nor Haskell's Double the same as C's double and so on. While on *some* platforms they are the same and you can write throw-away programs relying on these, the goal of portability requires you to declare imported and exported functions using special types described in the FFI standard, which are guaranteed to correspond to C types. These are:

import Foreign.C.Types (               -- equivalent to the following C type:
         CChar, CUChar,                --  char/unsigned char
         CShort, CUShort,              --  short/unsigned short
         CInt, CUInt, CLong, CULong,   --  int/unsigned/long/unsigned long
         CFloat, CDouble...)           --  float/double

Now we can import and export typeful C/Haskell functions:

foreign import ccall unsafe "math.h"
    c_sin :: CDouble -> CDouble

Note that pure C functions (those whose results depend only on their arguments) are imported without IO in their return type. The "const" specifier in C is not reflected in Haskell types, so appropriate compiler checks are not performed.

All these numeric types are instances of the same classes as their Haskell cousins (Ord, Num, Show and so on), so you may perform calculations on these data directly. Alternatively, you may convert them to native Haskell types. It's very typical to write simple wrappers around imported and exported functions just to provide interfaces having native Haskell types:

-- |Type-conversion wrapper around c_sin
sin :: Double -> Double
sin = fromRational . c_sin . toRational

Memory management

Marshalling strings

import Foreign.C.String (   -- representation of strings in C
         CString,           -- = Ptr CChar
         CStringLen)        -- = (Ptr CChar, Int)
foreign import ccall unsafe "string.h"
    c_strlen :: CString -> IO CSize     -- CSize defined in Foreign.C.Types and is equal to size_t
-- |Type-conversion wrapper around c_strlen 
strlen :: String -> Int
strlen = ....

Marshalling composite types

A C array may be manipulated in Haskell as StorableArray.

There is no built-in support for marshalling C structures and using C constants in Haskell. These are implemented in c2hs preprocessor, though.

Binary marshalling (serializing) of data structures of any complexity is implemented in library Binary.

Dynamic calls

DLLs

because i don't have experience of using DLLs, can someone write into this section? ultimately, we need to consider the following tasks:

  • using DLLs of 3rd-party libraries (such as ziplib)
  • putting your own C code into a DLL to use in Haskell
  • putting Haskell code into a DLL which may be called from C code

Dark side of IO monad

unsafePerformIO

Programmers coming from an imperative language background often look for a way to execute IO actions inside a pure procedure. But what does this mean? Imagine that you're trying to write a procedure that reads the contents of a file with a given name, and you try to write it as a pure (non-IO) function:

readContents :: Filename -> String

Defining readContents as a pure function will certainly simplify the code that uses it. But it will also create problems for the compiler:

  1. This call is not inserted in a sequence of "world transformations", so the compiler doesn't know at what exact moment you want to execute this action. For example, if the file has one kind of contents at the beginning of the program and another at the end - which contents do you want to see? You have no idea when (or even if) this function is going to get invoked, because Haskell sees this function as pure and feels free to reorder the execution of any or all pure functions as needed.
  2. Attempts to read the contents of files with the same name can be factored (i.e. reduced to a single call) despite the fact that the file (or the current directory) can be changed between calls. Again, Haskell considers all non-IO functions to be pure and feels free to omit multiple calls with the same parameters.

So, implementing pure functions that interact with the Real World is considered to be Bad Behavior. Good boys and girls never do it ;)


Nevertheless, there are (semi-official) ways to use IO actions inside of pure functions. As you should remember this is prohibited by requiring the RealWorld "baton" in order to call an IO action. Pure functions don't have the baton, but there is a special "magic" procedure that produces this baton from nowhere, uses it to call an IO action and then throws the resulting "world" away! It's a little low-level magic. This very special (and dangerous) procedure is:

unsafePerformIO :: IO a -> a

Let's look at its (possible) definition:

unsafePerformIO :: (RealWorld -> (a, RealWorld)) -> a
unsafePerformIO action = let (a, world1) = action createNewWorld
                         in a

where 'createNewWorld' is an internal function producing a new value of the RealWorld type.

Using unsafePerformIO, you can easily write pure functions that do I/O inside. But don't do this without a real need, and remember to follow this rule: the compiler doesn't know that you are cheating; it still considers each non-IO function to be a pure one. Therefore, all the usual optimization rules can (and will!) be applied to its execution. So you must ensure that:

  1. The result of each call depends only on its arguments.
  2. You don't rely on side-effects of this function, which may be not executed if its results are not needed.


Let's investigate this problem more deeply. Function evaluation in Haskell is determined by a value's necessity - the language computes only the values that are really required to calculate the final result. But what does this mean with respect to the 'main' function? To "calculate the final world's" value, you need to perform all the intermediate IO actions that are included in the 'main' chain. By using 'unsafePerformIO' we call IO actions outside of this chain. What guarantee do we have that they will be run at all? None. The only time they will be run is if running them is required to compute the overall function result (which in turn should be required to perform some action in the 'main' chain). This is an example of Haskell's evaluation-by-need strategy. Now you should clearly see the difference:

- An IO action inside an IO procedure is guaranteed to execute as long as it is (directly or indirectly) inside the 'main' chain - even when its result isn't used (because the implicit "world" value it returns will be used). You directly specify the order of the action's execution inside the IO procedure. Data dependencies are simulated via the implicit "world" values that are passed from each IO action to the next.

- An IO action inside 'unsafePerformIO' will be performed only if result of this operation is really used. The evaluation order is not guaranteed and you should not rely on it (except when you're sure about whatever data dependencies may exist).


I should also say that inside 'unsafePerformIO' call you can organize a small internal chain of IO actions with the help of the same binding operators and/or 'do' syntactic sugar we've seen above. For example, here's a particularly convoluted way to compute the integer that comes after zero:

one :: Int
one = unsafePerformIO $ do var <- newIORef 0
                           modifyIORef var (+1)
                           readIORef var

and in this case ALL the operations in this chain will be performed as long as the result of the 'unsafePerformIO' call is needed. To ensure this, the actual 'unsafePerformIO' implementation evaluates the "world" returned by the 'action':

unsafePerformIO action = let (a,world1) = action createNewWorld
                         in (world1 `seq` a)

(The 'seq' operation strictly evaluates its first argument before returning the value of the second one).


inlinePerformIO

inlinePerformIO has the same definition as unsafePerformIO but with addition of INLINE pragma:

-- | Just like unsafePerformIO, but we inline it. Big performance gains as
-- it exposes lots of things to further inlining
{-# INLINE inlinePerformIO #-}
inlinePerformIO action = let (a, world1) = action createNewWorld
                         in (world1 `seq` a)
#endif

Semantically inlinePerformIO = unsafePerformIO in as much as either of those have any semantics at all.

The difference of course is that inlinePerformIO is even less safe than unsafePerformIO. While ghc will try not to duplicate or common up different uses of unsafePerformIO, we aggressively inline inlinePerformIO. So you can really only use it where the IO content is really properly pure, like reading from an immutable memory buffer (as in the case of ByteStrings). However things like allocating new buffers should not be done inside inlinePerformIO since that can easily be floated out and performed just once for the whole program, so you end up with many things sharing the same buffer, which would be bad.

So the rule of thumb is that IO things wrapped in unsafePerformIO have to be externally pure while with inlinePerformIO it has to be really really pure or it'll all go horribly wrong.

That said, here's some really hairy code. This should frighten any pure functional programmer...

write :: Int -> (Ptr Word8 -> IO ()) -> Put ()
write !n body = Put $ \c buf@(Buffer fp o u l) ->
  if n <= l
    then write' c fp o u l
    else write' (flushOld c n fp o u) (newBuffer c n) 0 0 0

  where {-# NOINLINE write' #-}
        write' c !fp !o !u !l =
          -- warning: this is a tad hardcore
          inlinePerformIO
            (withForeignPtr fp
              (\p -> body $! (p `plusPtr` (o+u))))
          `seq` c () (Buffer fp o (u+n) (l-n))

it's used like:

word8 w = write 1 (\p -> poke p w)

This does not adhere to my rule of thumb above. Don't ask exactly why we claim it's safe :-) (and if anyone really wants to know, ask Ross Paterson who did it first in the Builder monoid)

unsafeInterleaveIO

But there is an even stranger operation called 'unsafeInterleaveIO' that gets the "official baton", makes its own pirate copy, and then runs an "illegal" relay-race in parallel with the main one! I can't talk further about its behavior without causing grief and indignation, so it's no surprise that this operation is widely used in countries that are hotbeds of software piracy such as Russia and China! ;) Don't even ask me - I won't say anything more about this dirty trick I use all the time ;)

One can use unsafePerformIO (not unsafeInterleaveIO) to perform I/O operations not in predefined order but by demand. For example, the following code:

do let c = unsafePerformIO getChar
   do_proc c

will perform getChar I/O call only when value of c is really required by code, i.e. it this call will be performed lazily as any usual Haskell computation.

Now imagine the following code:

do let s = [unsafePerformIO getChar, unsafePerformIO getChar, unsafePerformIO getChar]
   do_proc s

Three chars inside this list will be computed on demand too, and this means that their values will depend on the order they are consumed. It is not that we usually need.


unsafeInterleaveIO solves this problem - it performs I/O only on demand but allows to define exact *internal* execution order for parts of your datastructure. It is why I wrote that unsafeInterleaveIO makes illegal copy of baton.

First, unsafeInterleaveIO has (IO a) action as a parameter and returns value of type 'a':

do str <- unsafeInterleaveIO myGetContents

Second, unsafeInterleaveIO don't perform any action immediately, it only creates a box of type 'a' which on requesting this value will perform action specified as a parameter.

Third, this action by itself may compute the whole value immediately or... use unsafeInterleaveIO again to defer calculation of some sub-components:

myGetContents = do
   c <- getChar
   s <- unsafeInterleaveIO myGetContents
   return (c:s)

This code will be executed only at the moment when value of str is really demanded. In this moment, getChar will be performed (with result assigned to c) and one more lazy IO box will be created - for s. This box again contains link to the myGetContents call

Then, list cell returned that contains one char read and link to myGetContents call as a way to compute rest of the list. Only at the moment when next value in list required, this operation will be performed again

As a final result, we get inability to read second char in list before first one, but lazy character of reading in whole. bingo!


PS: of course, actual code should include EOF checking. also note that you can read many chars/records at each call:

myGetContents = do
   c <- replicateM 512 getChar
   s <- unsafeInterleaveIO myGetContents
   return (c++s)

A safer approach: the ST monad

We said earlier that we can use unsafePerformIO to perform computations that are totally pure but nevertheless interact with the Real World in some way. There is, however, a better way! One that remains totally pure and yet allows the use of references, arrays, and so on -- and it's done using, you guessed it, type magic. This is the ST monad.

The ST monad's version of unsafePerformIO is called runST, and it has a very unusual type.

runST :: (forall s . ST s a) -> a

The s variable in the ST monad is the state type. Moreover, all the fun mutable stuff available in the ST monad is quantified over s:

newSTRef :: a -> ST s (STRef s a)
newArray_ :: Ix i => (i, i) -> ST s (STArray s i e)

So why does runST have such a funky type? Let's see what would happen if we wrote

makeSTRef :: a -> STRef s a
makeSTRef a = runST (newSTRef a)

This fails, because newSTRef a doesn't work for all state types s -- it only works for the s from the return type STRef s a.

This is all sort of wacky, but the result is that you can only run an ST computation where the output type is functionally pure, and makes no references to the internal mutable state of the computation. The ST monad doesn't have access to I/O operations like writing to the console, either -- only references, arrays, and suchlike that come in handy for pure computations.

Important note -- the state type doesn't actually mean anything. We never have a value of type s, for instance. It's just a way of getting the type system to do the work of ensuring purity for us, with smoke and mirrors.

It's really just type system magic: secretly, on the inside, runST runs a computation with the real world baton just like unsafePerformIO. Their internal implementations are almost identical: in fact, there's a function

stToIO :: ST RealWorld a -> IO a

The difference is that ST uses type system magic to forbid unsafe behavior like extracting mutable objects from their safe ST wrapping, but allowing purely functional outputs to be performed with all the handy access to mutable references and arrays.

So here's how we'd rewrite our function using unsafePerformIO from above:

oneST :: ST s Int -- note that this works correctly for any s
oneST = do var <- newSTRef 0
           modifySTRef var (+1)
           readSTRef var

one :: Int
one = runST oneST

Welcome to the machine: the actual GHC implementation

A little disclaimer: I should say that I'm not describing here exactly what a monad is (I don't even completely understand it myself) and my explanation shows only one _possible_ way to implement the IO monad in Haskell. For example, the hbc Haskell compiler and the Hugs interpreter implements the IO monad via continuations. I also haven't said anything about exception handling, which is a natural part of the "monad" concept. You can read the "All About Monads" guide to learn more about these topics.

But there is some good news: first, the IO monad understanding you've just acquired will work with any implementation and with many other monads. You just can't work with RealWorld values directly.

Second, the IO monad implementation described here is really used in the GHC, yhc/nhc (jhc, too?) compilers. Here is the actual IO definition from the GHC sources:

newtype IO a = IO (State# RealWorld -> (# State# RealWorld, a #))

It uses the "State# RealWorld" type instead of our RealWorld, it uses the "(# #)" strict tuple for optimization, and it adds an IO data constructor around the type. Nevertheless, there are no significant changes from the standpoint of our explanation. Knowing the principle of "chaining" IO actions via fake "state of the world" values, you can now easily understand and write low-level implementations of GHC I/O operations.


The Yhc/nhc98 implementation

data World = World
newtype IO a = IO (World -> Either IOError a)

This implementation makes the "World" disappear somewhat, and returns Either a result of type "a", or if an error occurs then "IOError". The lack of the World on the right-hand side of the function can only be done because the compiler knows special things about the IO type, and won't overoptimise it.


Further reading

[1] This tutorial is largely based on the Simon Peyton Jones' paper Tackling the awkward squad: monadic input/output, concurrency, exceptions, and foreign-language calls in Haskell. I hope that my tutorial improves his original explanation of the Haskell I/O system and brings it closer to the point of view of beginning Haskell programmers. But if you need to learn about concurrency, exceptions and FFI in Haskell/GHC, the original paper is the best source of information.

[2] You can find more information about concurrency, FFI and STM at the GHC/Concurrency#Starting points page.

[3] The Arrays page contains exhaustive explanations about using mutable arrays.

[4] Look also at the Using monads page, which contains tutorials and papers really describing these mysterious monads.

[5] An explanation of the basic monad functions, with examples, can be found in the reference guide A tour of the Haskell Monad functions, by Henk-Jan van Tuyl.

[6] Official FFI specifications can be found on the page The Haskell 98 Foreign Function Interface 1.0: An Addendum to the Haskell 98 Report

[7] Using FFI in multithreaded programs described in paper Extending the Haskell Foreign Function Interface with Concurrency

Do you have more questions? Ask in the haskell-cafe mailing list.

To-do list

If you are interested in adding more information to this manual, please add your questions/topics here.

Topics:

  • fixIO and 'mdo'
  • Q monad

Questions:

  • split '>>='/'>>'/return section and 'do' section, more examples of using binding operators
  • IORef detailed explanation (==const*), usage examples, syntax sugar, unboxed refs
  • explanation of how the actual data "in" mutable references are inside 'RealWorld', rather than inside the references themselves ('IORef','IOArray',&c.)
  • control structures developing - much more examples
  • unsafePerformIO usage examples: global variable, ByteString, other examples
  • how 'unsafeInterLeaveIO' can be seen as a kind of concurrency, and therefore isn't so unsafe (unlike 'unsafeInterleaveST' which really is unsafe)
  • discussion about different senses of "safe"/"unsafe" (like breaking equational reasoning vs. invoking undefined behaviour (so can corrupt the run-time system))
  • actual GHC implementation - how to write low-level routines on example of newIORef implementation

This manual is collective work, so feel free to add more information to it yourself. The final goal is to collectively develop a comprehensive manual for using the IO monad.