I like seeing more things come out helping people with data analysis in various languages, but this put me off:<p>> isWhitespace x = elem x " \t\r\n"<p>This is the kind of thing that makes me concerned about using this resource for real-world data. In real-world data you're going to get all kinds of crazy things coming in, and if you're assuming nobody will ever have something like a zero width non-breaking space, or a form feed, you're going to have a problem.<p>It's the kind of thing I see with people starting out when dealing with data, similarly the punctuation detection here: <a href="https://github.com/BinRoot/Haskell-Data-Analysis-Cookbook/blob/master/Ch02/Code02_punctuation/Main.hs" rel="nofollow">https://github.com/BinRoot/Haskell-Data-Analysis-Cookbook/bl...</a><p>If you rely on these things, you will have problems. Text is hard and weird and terribly more complicated than people usually expect.<p>Does haskell have good libraries for dealing with the more awkward parts? Can I easily remove all characters marked as whitespace in unicode, for example? Detecting and managing mangled encodings?