what is web information retrieval from web search engine web crawler web crawler policies conclusion...

24

Upload: ethelbert-blake

Post on 19-Jan-2016

220 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms
Page 2: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

What is Web

Information retrieval from web

Search Engine

Web Crawler

Web crawler policies

Conclusion

How does a web crawler work

Synchronization Algorithms

Page 3: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

What is Web

The Web is a global system of interconnected computer networks that use the standardized Internet Protocol Suite .

It is a network of networks that consists of millions of private ,public, academic, business, and government networks .

The Internet carries a vast array of information resources and services, and the infrastructure to support electronic mail.

Page 4: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Information retrieval from webViewing a Web page on the World Wide Web normally begins

either by typing the URL of the page into a Web browser, or by following a hyperlink to that page or resource.

1. First, the server-name portion of the URL is resolved into an IP address using the global, distributed Internet database known as the domain name system

2.The browser then requests the resource by sending an HTTP request to the Web server at that particular address.

3.Browser then renders the page onto the screen as specified by its HTML, CSS, and other Web languages.

Page 5: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Search EngineA search engine is an information retrieval system designed to help find information stored on a computer system. It searches for information on the World Wide Web.

Search engines use automated software programs known as spiders or bots to survey the Web and build their databases.

Web documents are retrieved by these programs and analyzed and indexed.  

Page 6: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Search engine operations:A search engine operates in the following order:

Web crawling

Indexing

Searching

Page 7: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Web Crawler:A web crawler is a program or automated script which browses the World Wide web in a methodical, automated manner through Internet pages .

Crawlers are small programs that `browse' the Web on the search engine's behalf to collect information

Crawlers can also be used for automating maintenance tasks on a Web site, such as checking links or validating HTML code.

Page 8: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

It starts with a list of URLs to visit, called the seeds. As the crawler visits these URLs, it identifies all the hyperlinks in the page and adds them to the list of URLs to visit, called the crawl frontier.

Page 9: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Crawler control module:

 This module determines what links to visit next, and feeds the links to visit back to the crawlers.

The programs are given a starting set of URLs, whose pages they retrieve from the Web. The crawlers extract URLs appearing in the retrieved pages, and give this information to the crawler control module.The crawl control module is responsible for directing the crawling operation.

Page 10: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Crawling Policies

a selection policy that states which pages to download,

a re-visit policy that states when to check for changes to the pages,

a politeness policy that states how to avoid overloading Web sites, and

a parallelization policy that states how to coordinate distributed Web crawlers

Some characteristics of the Web that make crawling it very difficult:its large volume, Its fast rate of change, and dynamic page generation. Hence we need certain policies to make works easier. They are:

Page 11: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Selection policy

As a crawler always downloads just a fraction of the Web pages, it is highly desirable that the downloaded fraction contains the most relevant pages and not just a random sample of the Web . Hence the Selection policy is required

The importance of a page is a function of its intrinsic quality, its popularity in terms of links or visits, and even of its URL. Hence we require a metric of importance for prioritizing the web pages.Designing a good selection policy has an added difficulty: it must work with partial information, as the complete set of Web pages is not known during crawling.

Page 12: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Revisit policy

The Web has a very dynamic nature, and crawling a fraction of the Web can take a really long time, By the time a Web crawler has finished its crawl, many events could have happened. Hence the Revisit policy is required

Freshness: Let S = {e1; : : : ; eN} be the local database with N elements.Then we define the freshness of the collection as follows: The freshness of a local page ei at time t is  

1 if p is equal to the local copy at time Fp(t) = 0 otherwise Then, the freshness of the local database S at time t is  F(S; t) =1/N (∑F (e i , t)) for i= 1to N 

Page 13: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Age: To capture ‘how old’ the collection is, we define the metric age as follows:The age of the local element ei at time t is:

 0 if p is not modified at time t

Ap(t) = t – modification time (tm) of p otherwise  Then the age of the local database S is 

A(S; t) = 1/N (∑A(ei; t)) for i= 1 to N

Page 14: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Suppose that the crawler maintains a collection of two pages: e1 and e2. Page e1 changes 9 times per day and e2

changes once a day. Our goal is to maximize the freshness of the database averaged over time

Page 15: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Because our crawler is a tiny one, assume that we can refresh one page per day. Then what page should it refresh? Should the crawler refresh e1 or should it refresh e2?

Page 16: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Politeness Policy

Crawlers can retrieve data much quicker and in greater depth than human searchers, so they can have a crippling impact on the performance of a site. Needless to say, if a single crawler is performing multiple requests per second and/or downloading large files, a server would have a hard time keeping up with requests from multiple crawlers. Hence Politeness Policy is required

Page 17: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Parallelization policy

A parallel crawler is a crawler that runs multiple processes in parallel. The goal is to maximize the download rate while minimizing the overhead from parallelization and to avoid repeated downloads of the same page. Hence Parallelization policy is required

Page 18: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

Different synchronization Algorithmsa) Fixed-order policy: Under the fixed-order policy, we

synchronize the local elements in the same order repeatedly

Algorithm : Fixed-order synchronizationInput: ElemList = {e1; e2; : : : ; eN}Procedure:[1] While (TRUE)[2] SyncQueue := ElemList[3] While (not Empty(SyncQueue))[4] e := Dequeue(SyncQueue)[5] Synchronize(e)

Page 19: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

b) Random-order policy: Under the random-order policy, the synchronization order of elements might be different from one crawl to the next and we randomize the order of elements before every iteration

Algorithm: Random-order synchronizationInput: ElemList = {e1; e2; : : : ; eN}Procedure:[1] While (TRUE)[2] SyncQueue := RandomPermutation(ElemList)[3] While (not Empty(SyncQueue))[4] e := Dequeue(SyncQueue)[5] Synchronize(e)

Page 20: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

c) Purely-random policy: Whenever we synchronize an element, we pick an arbitrarily random element under the purely random policy.

Algorithm: Purely-random synchronizationInput: ElemList = {e1; e2; : : : ; eN}Procedure:[1] While (TRUE)[2] e := PickRandom(ElemList)[3] Synchronize(e)

Page 21: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

ConclusionThus we conclude how subtly a web crawler can optimize the

working of a normal search engine.

Page 22: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

References

1. http://en.wikipedia.org/wiki/Web_crawler

2.http://www.webcrawler.com/webcrawler/ws/about/_iceUrlFlag=11

3.http://www.wisegeek.com/what-is-a-web-crawler.htm

4.http://thinkpink.com/bp/WebCrawler/History.html

Page 23: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms
Page 24: What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms

ANY QUERRIES ?