What is a spider anyway? Spiders are commonly referred to as crawlers or robots.
Google’s infamous spider is named Googlebot. Googlebot visits most sites every six weeks or so.
Almost all search engines employ spiders, which are automated programs that scour the Web and create searchable indexes.
Some search engines do not have spiders and need direct submissions to get your site listed in their index.
Large search engine companies, such as Google, have thousands of computers around the world that incorporate spider software.
The spiders gather information about your website (in other words content) and create indexes for the visitors searching the web.
Although search results may be viewed in different formats, all search engines will allow the visitor to search using one or more
keywords or phrases. Searching the indexes has become quite an art form. After entering your keyword phrases,
your results will be displayed, most likely, with a bit of information along side the link address.
The main objective is having your website visited by a spider. One method is to incorporate backlinks to your web page.
This is one circumstance, in which, being a loner is not a good idea. These links and backlinks give the web spiders a
pathway into your website and a pathway back out again. Another way is to use submissions to specific search engines and directories.
The spiders will visit as if called to your site.
How do you prevent a page from being crawled? There are two methods; use a robots.txt file or code in HTML.
If you are familiar with HTML, simply use the appropriate Meta Tag for the page or pages you don’t want to be indexed.
Robots.txt files require two commands: “useragent”, which allows you to specify which search engines you want to keep out,
and “disallow”, which enable you to pick pages or directories you don’t want searched. Website Marketing by BtSEO.