Skip to main content

Home/ Daido Moriyama - Modern Star/ Harnessing the Energy of Robots.txt
Joensen Borch

Harnessing the Energy of Robots.txt - 0 views

internet

started by Joensen Borch on 03 Oct 13
  • Joensen Borch
     
    Sometimes, we might want search-engines never to list certain parts of the site, or even prohibit other SE from the site altogether. Should you require to be taught further on http://displaywarehouseuk.jimdo.com/, there are many databases people might consider investigating.

    This really is where a simple, little 2-line text file called robots.txt is available in.

    Once we've a website up and running, we need to make certain that all visiting search-engines can access all the pages we want them to check out. I discovered http://audreydisplaywarehouse.wordpress.com/2013/06/27/the-use-of-picture-frames-in-the-home/ by searching newspapers.

    Sometimes, we might want search engines never to list certain parts of the site, and on occasion even exclude other SE in the site all together.

    This is the place where a simple, little 2-line text file called robots.txt will come in.

    Robots.txt lives within your web sites main directory (o-n LINUX systems this is your /public_html/ directory), and looks something like the following:

    User-agent: *

    Disallow:

    The initial line controls the bot that'll be visiting your site, the next line controls if they're allowed in, or which areas of the site they're maybe not allowed to go to

    If you'd like to take care of multiple robots, then basic repeat the aforementioned lines. Dig up more on an affiliated link by clicking http://my.opera.com/sundaywar82/blog/2013/09/26/digital-photo-frame-display?firstpost=Y talk.

    So an example:

    User-agent: googlebot

    Disallow:

    User-agent: askjeeves

    Disallow: /

    This will enable Goggle (user-agent name GoogleBot) to visit every page and service, while in the same time banning Ask Jeeves in the site completely.

    To locate a reasonably updated listing of robot consumer names this visit http://www.robotstxt.org/wc/active/html/index.html

    Even if you want to let every software to index every page of your site, its still very advisable to place a robots.txt file in your site. It will stop your error records replenishing with entries from se's attempting to access your robots.txt file that doesnt occur.

    To learn more on robots.txt see, the entire listing of resources about robots.txt at http://www.websitesecrets101.com/robotstxt-further-reading-resources.

To Top

Start a New Topic » « Back to the Daido Moriyama - Modern Star group