Hi guys,
Tomorrow afternoon an intervention is planned for the server on which
xwiki.org wiki is running.
We estimate a downtime of ~5 mins. I'll also send you a new email tomorrow
with 1h before we start the intervention.
Thx,
Alex
Why not simply use robots.txt, its is made for this, no ?
On Sun, Jan 8, 2012 at 02:07, Caleb James DeLisle
<calebdelisle(a)lavabit.com>wrote:
> I noticed xwiki.org lagging and java process using lots of cpu, although
> ram was high it didn't seem to be a major issue.
> I noticed this:
> 180.76.5.100 - - [07/Jan/2012:21:13:00 +0100] "GET
> /xwiki/monitoring?part=graph&graph=sql34bb8b9d9f525e0790dab487491d120d4bc685cd&period=annee
> HTTP/1.1" 500 398 "-" "Mozilla/5.0 (compatible; Baiduspider/2.0; +
> http://www.baidu.com/search/spider.html)"
>
> And seeing as it seems you can do things like heap dump and run garbage
> collector with get requests, that page should probably be made difficult to
> access.
> One idea that comes to mind would be using mod_rewrite to change the url
> if there is no auth cookie set so any logged in user can view it but a
> roving bot can't.
>
> Caleb
>
> _______________________________________________
> infra mailing list
> infra(a)xwiki.org
> http://lists.xwiki.org/mailman/listinfo/infra
>
--
Denis Gervalle
SOFTEC sa - CEO
eGuilde sarl - CTO