Blocking web crawlers on lighttpd

Note: The information contained in this post may be outdated! Nutch did ignore my robots.txt (for whatever reason, I was unable to figure out why), so I had to find another way to forbid those directories for the crawler. I finally came up with this neat piece of config for lighty:

– throws an […]