Skip to content


Subversion checkout URL

You can clone with HTTPS or Subversion.

Download ZIP
tree: 01330ec25d
Fetching contributors…

Cannot retrieve contributors at this time

52 lines (43 sloc) 1.712 kb
__NAME__ purpose
specify hostnames that will be classified as crawler bots (search engines) visiting the site
__NAME__ see also
RobotUA, RobotIP
__NAME__ synopsis
<arg choice='plain' rep='repeat'><replaceable>hostname_glob</replaceable></arg>
__NAME__ description
The &conf-RobotHost; directive defines a list of hostnames which will be
classified as crawler robots (search engines), and cause &IC; to alter its
behavior to improve the chance of &IC;-served content being crawled
and indexed.
Note that this directive (and all other work done to identify robots)
only serves to improve the way in which &IC; pages are indexed, and to
reduce server overhead for clients that don't require our full attention
in the way humans do (for example, session information is not kept around
for spider bots).
Using this to "tune" the actual page content depending on a crawler
visiting does not earn you extra points, and may in fact be
detected by the robot and punished.
It's important to note that the directive accepts a wildcard list similar
to globbing &mdash;
<literal>*</literal> represents any number of characters, while
<literal>?</literal> represents a single character.
__NAME__ notes
For more details regarding web spiders/bots and &IC;, see
&glos-robot; glossary entry.
__NAME__ example: Defining RobotHost
RobotHost <<EOR
*.crawler*.com, *, *,
*, *, *,
*, *,,,
Jump to Line
Something went wrong with that request. Please try again.