World Wide Web Spiders (Crawlers or Robots) Help Page

A World Wide Web Spiders (Crawlers or Robots): A program that automatically explores the World Wide Web by retrieving a document and recursively retrieving some or all the documents that are referenced in it. Spiders follow the links on a site to find other relevant pages. There are two algorithms involved in spiders: one is depth-first search; the other is breadth-first search. The depth-first creates a relatively comprehensive database on a few objects, while the breadth-first builds a database that touches more lightly on a wider variety of documents. The following basic tools are usually used to implement an experimental spider. To construct an efficient and practical spider, some other networking tools have to be used.