Re: robots.txt on git.postgresql.org

Поиск
Список
Период
Сортировка
От Craig Ringer
Тема Re: robots.txt on git.postgresql.org
Дата
Msg-id 51DD1A69.1000001@2ndquadrant.com
обсуждение исходный текст
Ответ на Re: robots.txt on git.postgresql.org  (Andres Freund <andres@2ndquadrant.com>)
Ответы Re: robots.txt on git.postgresql.org  (Dave Page <dpage@pgadmin.org>)
Re: robots.txt on git.postgresql.org  (Magnus Hagander <magnus@hagander.net>)
Список pgsql-hackers
On 07/09/2013 11:30 PM, Andres Freund wrote:
> On 2013-07-09 16:24:42 +0100, Greg Stark wrote:
>> I note that git.postgresql.org's robot.txt refuses permission to crawl
>> the git repository:
>>
>> http://git.postgresql.org/robots.txt
>>
>> User-agent: *
>> Disallow: /
>>
>>
>> I'm curious what motivates this. It's certainly useful to be able to
>> search for commits.
> 
> Gitweb is horribly slow. I don't think anybody with a bigger git repo
> using gitweb can afford to let all the crawlers go through it.

Wouldn't whacking a reverse proxy in front be a pretty reasonable
option? There's a disk space cost, but using Apache's mod_proxy or
similar would do quite nicely.

-- Craig Ringer                   http://www.2ndQuadrant.com/PostgreSQL Development, 24x7 Support, Training & Services



В списке pgsql-hackers по дате отправления:

Предыдущее
От: Hannu Krosing
Дата:
Сообщение: Re: Removing Inner Joins
Следующее
От: Atri Sharma
Дата:
Сообщение: Re: Removing Inner Joins