Dynamic programming example
m (Corrected bitmask of last example (0x8120 should be 0x80120))
(add a reference to a good article)
Latest revision as of 21:55, 4 June 2014
Dynamic programming refers to translating a problem to be solved into a recurrence formula, and crunching this formula with the help of an array (or any suitable collection) to save useful intermediates and avoid redundant work.
Computationally, dynamic programming boils down to write once, share and read many times. This is exactly what lazy functional programming is for.
 1 Sample problems and solutions
 1.1 Available in 6-packs, 9-packs, 20-packs
A fast food place sells a finger food in only boxes of 6 pieces, boxes of 9 pieces, or boxes of 20 pieces. You can only buy zero or more such boxes. Therefore it is impossible to buy exactly 5 pieces, or exactly 7 pieces, etc. Can you buy exactly N pieces?If I can buy i-6 pieces, or i-9 pieces, or i-20 pieces (provided these are not negative numbers), I can then buy i pieces (by adding a box of 6 or 9 or 20). Below, I set up the array
import Data.Array buyable n = r!n where r = listArray (0,n) (True : map f [1..n]) f i = i >= 6 && r!(i-6) || i >= 9 && r!(i-9) || i >= 20 && r!(i-20)
import Data.Array buy n = r!n where r = listArray (0,n) (Just (0,0,0) : map f [1..n]) f i = case attempt (i-6) of Just(x,y,z) -> Just(x+1,y,z) _ -> case attempt (i-9) of Just(x,y,z) -> Just(x,y+1,z) _ -> case attempt (i-20) of Just(x,y,z) -> Just(x,y,z+1) _ -> Nothing attempt x = if x>=0 then r!x else Nothing
import Data.Array import Control.Monad(guard,mplus) buy n = r!n where r = listArray (0,n) (Just (0,0,0) : map f [1..n]) f i = do (x,y,z) <- attempt (i-6) return (x+1,y,z) `mplus` do (x,y,z) <- attempt (i-9) return (x,y+1,z) `mplus` do (x,y,z) <- attempt (i-20) return (x,y,z+1) attempt x = guard (x>=0) >> r!x
This more regular code can be further generalized.
 2 Optimization
Simple dynamic programing is usually fast enough (and as always, profile before optimizing!) However, when you need more speed, it is usually fairly easy to shave an order of magnitude off the space usage of dynamic programming problems (with concomitant speedups due to cache effects.) The trick is to manually schedule the computation in order to discard temporary results as soon as possible.
Notice that if we compute results in sequential order from 0 to the needed count, (in the example above) we will always have computed subproblems before the problems. Also, if we do it in this order we need not keep any value for longer than twenty values. So we can use the old fibonacci trick:
buyable n = iter n (True : replicate 19 False) where iter 0 lst = lst !! 0 iter n lst = iter (n-1) ((lst !! 5 || lst !! 8 || lst !! 19) : take 19 lst)
At each call of iter, the n parameter contains (total - cur) and the lst parameter stores buyable for (cur-1, cur-2, cur-3, ...). Also note that the indexes change meaning through the cons, so we need to offset the !! indexes by 1.
We can improve this more by packing the bit array:
import Data.Bits buyable n = iter n 1 where iter :: Int -> Int -> Bool iter 0 lst = odd lst iter n lst = iter (n-1) ((lst `shiftL` 1) .|. if lst .&. 0x80120 /= 0 then 1 else 0)
This final version is compiled into a single allocation-free loop.
 3 Further Reading
A good detailed explanation: lazy dynamic programming by Tikhon Jelvis.