
On Thu, 14 Jun 2007, Lutz Donnerhacke wrote:
It's a bad and buggy rewrite from scratch. It can check a file of urls or the consistency of the transitive hull of a website incl. the existance of the border urls. Futhermore the warnings from TagSoup parsing can be reported.
Main bugs are memory leaks in conjunction with unnecessary retrieval of binary files, and missing documentation. If somebody has enough time in the next weeks: Many thanks in advance.
URL: http://www.iks-jena.de/mitarb/lutz/haskell/urlcheck-0.0.tar.gz
A great feature would be, if an 'urlcheck' program does not only check the validity of links but does also help keeping the links valid. It could download copies of each referenced HTML document as long as the links are valid. If on a check run a link is unreachable it could ask a search engine for an alternative copy with the same content. That's certainly easier said than implemented.