Hi, On 20.01.2013 11:48, Holger Hans Peter Freyther wrote: > this is due multiple web spiders (Bing, Baidu, ...) hitting the trac > and cgit. Patches for the robots.txt are welcome. I guess setting up Varnish would make more sense than blocking spiders, after all we want the pages to be found by search engines... Regards, Steve