Hi,

I  've got a lot of files which I need to proces in order to make them indexable by sphinx.
The files contain the data of a website with a custom perl based cms.  Unfortunatly they sometimes contain xml/html tags like <i>

And since most of the texts are in dutch and some are in French they also contain a lot of special characters like ë é, ...

I'm trying to replace the custom based perl based cms by a haskell one.  And I would like to add search capability. Since someone wrote sphinx
bindings a few weeks ago I thought I try that.  

But transforming the files in something that sphinx seems a challenge.  Most special character problems seem to go aways when I  use encodeString (Codec.Binary.UTF8.String)
on the indexable data.

But the sphinx indexer complains that the xml isn't valid.  When I look at the errors this seems due to some documents containing not well formed  html.
I would like to use a programmatic solution to this problem.

And is there some haskell function which converts special tokens lik & -> &amp; and é -> &egu; ?

thanks in advance,

Pieter



--
Pieter Laeremans <pieter@laeremans.org>

"The future is here. It's just not evenly distributed yet." W. Gibson