Computer World. Матросова Т.А. - 32 стр.

UptoLike

Составители: 

31
Technologies for securely interfacing private corporate networks with
unsecured public ones, such as firewalls, proxy servers, and packet-
filtering routers.
Technologies for increasing availability and reliability of access to
network resources, such as clustering, caching, load balancing, and fault-
tolerant technologies.
Network management technologies such as the Simple Network
Management Protocol (SNMP) and Remote Network Monitoring
(RMON)
TEXT 7
Read the text and give it a suitable title.
The web creates new challenges for information retrieval. The amount of
information on the web is growing rapidly, as well as the number of new users
inexperienced in the art of web research. People are likely to surf the web using its
link graph, often starting with high quality human maintained indices such as Yahoo!
or with search engines. Human maintained lists cover popular topics effectively but
are subjective, expensive to build and maintain, slow to improve, and cannot cover all
esoteric topics. Automated search engines that rely on keyword matching usually
return too many low quality matches. To make matters worse, some advertisers
attempt to gain people's attention by taking measures meant to mislead automated
search engines. We have built a large-scale search engine which addresses many of
the problems of existing systems. It makes especially heavy use of the additional
structure present in hypertext to provide much higher quality search results. We chose
our system name Google, because it is a common spelling of googol, or
10
10
0
and fits
well with our goal of building very large-scale search engines.
Search engine technology has had to scale dramatically to keep up with the
growth of the web. In 1994, one of the first web search engines, the World Wide Web
Worm (WWWW) had an index of 110,000 web pages and web accessible documents.
As of November, 1997, the top search engines claim to index from 2 million to 100
million web documents. At the same time, the number of queries search engines
handle has grown incredibly too. In March and April 1994, the World Wide Web
Worm received an average of about 1,500 queries per day. In November 1997
Altavista claimed it handled roughly 20 million queries per day. The goal of our
system is to address many of the problems, both in quality and scalability, introduced
by scaling search engine technology to such extraordinary numbers.
Creating a search engine which scales even to today's web presents many
challenges. Fast crawling technology is needed to gather the web documents and keep
them up to date. Storage space must be used efficiently to store indices and,
optionally, the documents themselves. The indexing system must process hundreds of
gigabytes of data effic iently. Queries must be handled quickly, at a rate of hundreds
to thousands per second.