import Data.Array ( (!), array )
import Data.Char ( ord )
This question is a good excuse to:
see how the do notation is actually compiled
reveal the existence of monads (and demystify the notion)
talk some more about lazy evalutation
First, let us recall how the do notation for IO is translated to a bunch of function applications (that is what ghc will do under the hood).
To translate every instance of the do notation we have seen so far, we only need one basic function (written as an infix operator and commonly called ‘bind’)
(>>=) :: IO a -> (a -> IO b) -> IO b
that allows to chain IO inputs. To build some familiarity with this, let us first consider a simpler sequence operator
(>>) :: IO a -> IO b -> IO b
which is defined by
>> y = x >>= \_ -> y x
We don’t need to read that definition right now to build intuitions:
x >> y
corresponds to performing the action in
x
, then the action in y
and return the same
result as y
.
ioSeqEx :: IO ()
= putStrLn "Hi" >> putStrLn "Bye" ioSeqEx
This is the same as
ioSeqExDo :: IO ()
= do {
ioSeqExDo putStrLn "Hi";
putStrLn "Bye"
}
and in general, any expression of the shape
do
{
t;
u;...
}
for t
and u
programs of type
IO a
(without assignment) is translated to
t >> u >> ...
The more general (>>=)
is needed precisely to
handle variable assignments Let’s take a non-trivial example.
(>>=) :: IO a -> (a -> IO b) -> IO b
copycatDo :: IO ()
= do {
copycatDo <- getLine;
x putStrLn (x ++ "!")
}
This is equivalent to
copycatBind :: IO ()
= getLine >>= \x -> putStrLn (x ++ "!") copycatBind
We are now ready to look at what the first program of the file looks like when we translate the do notation using these combinators. Recall that the definition was
doIO1 :: IO String
= do {
doIO1 <- getLine;
a <- getLine;
b putStrLn (b ++ a);
return (b ++ a)
}
Now let us look at the equivalent variant using (>>=) and (>>) and using a suggestive indentation
bindIO :: IO String
= getLine >>= \a ->
bindIO getLine >>= \b ->
putStrLn (b ++ a) >>
return (b ++ a)
And that is all that is behind the do notations!
The do notation as we used it so far for IO
doIO :: IO String
= do {
doIO <- getLine;
a <- getLine;
b putStrLn (b ++ a);
return (b ++ a)
}
``` But in fact the do notations is also applicable in other contexts!
(What is below is a teaser with two examples, not an explanation)
1) do notation for list comprehensions
```haskell
doList :: [(Int,Char)]
= do {
doList <- [1..5];
x <- ['a','z'];
y return (x,y)
}
-- [(x,y) | x <- [1..5], y <- ['a','z']]
-- >>> doList
-- [(1,'a'),(1,'z'),(2,'a'),(2,'z'),(3,'a'),(3,'z'),(4,'a'),(4,'z'),(5,'a'),(5,'z')]
divMaybe :: Int -> Int -> Maybe Int
0 = Nothing
divMaybe x = Just (x `div` y)
divMaybe x y
doMaybe :: Int -> Int -> Int -> Maybe Int
= do {
doMaybe x y z <- divMaybe x y;
a <- divMaybe z a;
b return (a + b)
}
-- >>> doMaybe 1 0 0
-- Nothing
-- >>> doMaybe 1 2 0
-- >>> doMaybe 1 1 0
Even for the non-IO examples above such as
doList1 :: [(Int,Char)]
= do {
doList1 <- [1..5];
x <- ['a','z'];
y return (x,y)
}
the translation from the do notation to combinators is the same
bindList :: [(Int,Char)]
= [1..5] >>= \x ->
bindList 'a','z'] >>= \y ->
[return (x, y)
-- >>> doList1 == bindList
-- True
This does not tell us however what is the meaning of (>>=) when applied to things which are not IO. If we ask GHC what is the type of (>>=) and return, we have
return :: Monad m => a -> m a
(>>=) :: Monad m => m a -> (a -> m b) -> m b
These functions are actually methods of the typeclass Monad.
Side-note: this typeclass has one feature we have not encountered yet. The argument m of “Monad m” is not a type, but a type constructor. (jargon: it has higher kind * -> * instead of the ground kind * for types)
Examples of types (kind *): Int, Char, [Int]
Examples of type constructors of kind * -> *: lists ([_]), Maybe, IO, Either Int
Examples of type constructors of kind * -> * -> *: tupling (,), Either
Monad is just one of the many typeclasses defined in Prelude, but it is such a useful abstraction (in particular because it is used for IO).
Many type constructors come with a rather natural Monad instance such as [_] and Maybe which are pretty useful. As we have seen in the examples, the list monad gives some powerful combinators to implement list comprehensions. The official definition of the operators (>>=) and return for lists is the following:
return a = [a]
>>= _ = []
[] : xs) >>= f = f x ++ (xs >>= f) (x
The instance Monad Maybe on the other hand gives us combinators to easily chain computations with a “fail state” represented by Nothing. return is taken to be
return = Just
and the bind
constructor is asked as part of an optional
task for Lab 5 ;) It is also pretty natural to define a primitive
fail = Nothing
to write very natural programs using the do notation!
There are many more Monad instances in the base library which are useful for a variety of purposes: for instance, there is a ST that allows to compute pure functions with a suitable facilities to emulate “mutable variables” as you would find them in imperative languages.
Much like with any typeclasses, advanced users can also define their own instances.
Other major functional programming OCaml does not. Up to syntax and naming conventions (the following code snippets are fictional!), that language has
putStrLn :: String -> ()
getLine :: () -> String
and a chaining operator
:: a -> b -> b (;)
so that
= let z = getLine () in
f let y = getLine () in
putStrLn z;
10
is a valid program of type Int that performs side-effects along the way.
So why does Haskell force us to wrap our side-effects inside the IO monad?
One (slightly disingenuous) answer: because Haskell wants to reflect that computations that are impure are distinguished at the type-level, and that is good for programmers!
A stronger reason: because programs in Haskell are lazyly evaluated, having that design decision would make the order in which operations are executed highly counter-intuitive. Wrapping impure computations inside an IO monad actually forces evaluation of commands to be executed in order.
Example: if we evaluate the program f above eagerly, the instructions would all be executed in sequence. But if we are lazy and want to obtain the integer value of f while evaluating as little as possible, it is possible to do so without performing any input/output action!
Here I will declare a new typeclass and we will see how to define typeclass instances.
The typeclass that I am going to define is going to be called Size and will simply provide, for every type a such that Size a is defined, a canonical function a -> Int meant to define the size of the type a. First, we need to declare that
class Sized a where
size :: a -> Int
The syntax is
class NameOfTheClass typeParameter1 ... typeParameterN where
followed by type signatures of particular functions called
methods of the typeclass. We can check their actual types, and
see that it corresponds to the declaration above
-- >>> :t size
-- size :: Size a => a -> Int
Methods can be used to declare new functions sharing those sort of typeclass constraints
isBig :: Sized a => a -> Bool
= size x >= 51
isBig x
Now that we have declared the typeclass, now would be a good time to give some values of size for some datatypes where it makes sense! These values are called instances and are defined using the following syntax for base types
instance Sized Bool where
= 1
size _
-- >>> isBig True
-- False
instance Sized () where
= 1
size ()
instance Sized Int where
= 64
size i
instance Sized Char where
= 8
size _
instance Sized Integer where
= ceiling . log . fromIntegral size
It is possible to declare family of instances which can be implemented using other typeclass instances: here are a few examples:
instance Sized a => Sized [a] where
= sum . map size
size
instance (Sized a, Sized b) => Sized (a,b) where
= size x + size y
size (x,y)
-- >>> size ("abc",["aa","noooo","please"])
-- 80
-- >>> isBig "aaaaaaaaaaaaaaaa"
-- True
As we have seen above with Monad, typeclasses can also be defined in more complex situations where we have type families, but the general principles remain the same.
First, let us give a rough idea of one way of picturing the difference between lazy and eager evaluation. Given a function application
(\y -> \x -> (x,x)) (2^5) 2
the lazy evaluator will first try to substitute the arguments in the function body, delaying the evaluation of the arguments themselves until the last moment. One way of representing this computation on our example is
(\y -> \x -> x+x) (2^12) (2+2) (initial value)
-- → (\x -> x+1) (2 + 2) (replace y by the expression 2^5)
-- → (2 + 2) + 1 (replace x by the expression 2 + 2)
-- → 4 + 1
-- → 5
Eager evaluators will first evaluate the arguments of their functions before substituting within the function body.
(\y -> \x -> x+x) (2^12) (2+2) (initial value)
→ (\y -> \x -> x+1) 4096 (2 + 2) (compute 2^5)
→ (\y -> \x -> x+1) 4096 4 (compute 2+2)
→ (\x -> x+1) 4 (replace y by 4096)
→ 4 + 1
→ 5
The choice of having lazy evaluation vs eager evaluation can be a contentious issue in language design; there are pros and cons to both.
Since we are studying Haskell, which is lazy, I feel it would be best to leave you with a few examples of laziness in action, some of which are useful in practice!
Note that the following declaration is perfectly legal in Haskell, while it is impossible to do in eager languages such as OCaml (and analogon that would typecheck would typically loop and crash)
allPositiveList :: [Int]
= 0 : map (+1) allPositiveList allPositiveList
Trying to display this using show would lead to an infinite loop. But thanks to laziness, this can safely be used to compute finitary data.
-- >>> take 10 allPositiveList
-- [0,1,2,3,4,5,6,7,8,9]
Take a good look at the implementation of binomial below. Ignoring the trivial case where k > n, the main routine uses a local (immutable) array which seemingly gets initialized with the solution provided by a recursive function!
This is one elegant of way of doing dynamic programming in Haskell that is unavailable in languages with eager evaluation. For instance, to accomplish the same task in OCaml, we would typically need to use side-effects.
binomial :: Int -> Int -> Int
| k > n = 0
binomial k n | otherwise = a ! (k, n)
where a = array ((0,0),(k,n))
| i <- [0..k], j <- [0..n]]
[ ((i,j), b i j) 0 k = 1
b | i == j = 1
b i j = (a ! (i,j-1)) + (a ! (i-1,j-1))
b i j
-- >>> binomial 5 10
-- 252
Compare and contrast the following definitions:
andl :: Bool -> Bool -> Bool
False `andl` _ = False
`andl` False = False
_ True `andl` True = True
andr :: Bool -> Bool -> Bool
`andr` False = False
_ False `andr` _ = False
True `andr` True = True
``` Take an ill-founded recursive definition; trying to evaluate this would
not go well! ```haskell
bot :: a
= bot
bot
-- >>> bot
-- loops forever
-- >>> False `andl` bot
-- False
-- >>> bot `andl` False
-- loops forever
-- >>> bot `andr` False
-- False
Here are Math-oriented examples (that can be patched to work in an eager language, but laziness makes definition easier to go through) which are very fun!