Re: robots.txt on git.postgresql.org - Mailing list pgsql-hackers

From Greg Stark
Subject Re: robots.txt on git.postgresql.org
Date
Msg-id CAM-w4HPdUbND-qA8ho1EB-wvj+tXcX=0H_6JtQNbkd_UZsDmHw@mail.gmail.com
Whole thread Raw
In response to Re: robots.txt on git.postgresql.org  (Magnus Hagander <magnus@hagander.net>)
Responses Re: robots.txt on git.postgresql.org
Re: robots.txt on git.postgresql.org
List pgsql-hackers
On Wed, Jul 10, 2013 at 9:36 AM, Magnus Hagander <magnus@hagander.net> wrote:
> We already run this, that's what we did to make it survive at all. The
> problem is there are so many thousands of different URLs you can get
> to on that site, and google indexes them all by default.

There's also https://support.google.com/webmasters/answer/48620?hl=en
which lets us control how fast the Google crawler crawls. I think it's
adaptive though so if the pages are slow it should be crawling slowly


-- 
greg



pgsql-hackers by date:

Previous
From: KONDO Mitsumasa
Date:
Subject: Re: Improvement of checkpoint IO scheduler for stable transaction responses
Next
From: Andres Freund
Date:
Subject: Re: robots.txt on git.postgresql.org