Java Web Crawler

From Chorke Wiki
Jump to navigation Jump to search
A web crawler, or spider, is a type of bot that's typically operated by search engines like Google and Bing. Their purpose is to index the content of websites all across the Internet so that those websites can appear in search engine results.

References