This has also been posted elsewhere but it is a small tutorial so apologies if you’ve seen it before. I have added a couple of references though.
Robots.txt helps search engines determine what DIRECTORIES they should be reading and indexing.
Think of it as a way for the spider to optimise its work effort to index your site. In effect you are telling it what is relevant or not for indexing. You can think about how it works like this: If a spider has to index everything on the billions of sites out there, it is nice for it to know that some of them are kind enough to leave it a message about what files can be EXCLUDED for indexing.
That’s basically the point of the robots file.
It’s very easy to create a small text file using a text editor (I use the fantastic HyperEdit) and save it as robots.txt
you can find out how to structure the file here
Then in Freeway create a new page in the main folder called, for example, “siteincludes.html” then apply the “PHP Use Include Pages” Action from Page Actions, and link your text file to it. (you might have to download this action if you don’t already have it )
The siteincludes page is really a “dummy page” so you can also apply the PHP Make Markup Page to remove any HTML that Freeway creates.
The result will be that the Robots.txt file that you included in the siteincludes page will be added in the main directory of your site - exactly where it needs to be.
Another thing is that you need to create a sitemap.xml file and include a reference to this in your robots.txt file.
If you want to see what I mean have a look at http://www.google.com/robots.txt
yes they even have their own one!
The sitemap.xml file is yet another method for instructing the webspiders to index specific pages of your site, but it performs the opposite of exclusion. With a few small bits of added info the spider determines how important the page is and how often it should come back to check for changes.
you can read the official information here
you can create and add the file in Freeway the same way as described above, just don’t forget that for a sitemap text file you need to use .xml as an extension to the file name instead of .txt (ie the file should be called “sitemap.xml”)
I hope this helps some people.
Oh yes by the way a sitemap XML file is not the same as a page containing a list of pages on your site that happens to be called a SiteMap. the XML file is for the spider, the page is for the Internet User.
tutorials mailing list
email@hidden
Update your subscriptions at:
http://freewaytalk.net/person/options