What is a Search engine?
A: A search engine is a software system designed to carry out web searches.
They search the World Wide Web in a systematic way for
what?
A: Information specified in a textual web search query.
The search results are generally presented in a line of
results, often referred to as what?
A: Search engine results pages (SERPs).
Some search engines also mine data available in what?
A: Databases or open directories.
Unlike web directories, which are maintained only by
human editors, search engines also maintain real-time information by running
what?
A: An algorithm on a web crawler.
Any internet-based content that can't be indexed and
searched by a web search engine fall under what category?
A: Deep web.
Link analysis would eventually become a crucial
component of search engines through algorithms such as what?
A: Hyper Search and PageRank.
What was the first well documented search engine that
searched content files, namely FTP files?
A: Archie, which debuted on 10 September 1990.
Prior to September 1993, how was the World Wide Web
indexed?
A: By hand.
There was a list of webservers edited by Tim
Berners-Lee and hosted where?
A: On the CERN webserver.
The first tool used for searching content (as opposed
to users) on the Internet was what?
A: Archie.
The name stands for what?
A: "Archive" without the "v".
Who was it created by?
A: Alan Emtage, a computer science student at McGill University in
Montreal, Quebec, Canada.
The program downloaded what?
A: The directory listings of all the files located on public anonymous FTP
(File Transfer Protocol) sites, creating a searchable database of file
names.
Archie Search Engine did not index what?
A: The contents of these sites since the amount of data was so limited it
could be readily searched manually.
The rise of Gopher (created in 1991 by Mark McCahill at
the University of Minnesota) led to what two new search programs?
A: Veronica and Jughead.
Like Archie, they searched what?
A: The file names and titles stored in Gopher index systems.
Veronica (Very Easy Rodent-Oriented Net-wide Index to
Computerized Archives) provided a keyword search of what?
A: Most Gopher menu titles in the entire Gopher listings.
Jughead (Jonzy's Universal Gopher Hierarchy Excavation
and Display) was a tool for obtaining what?
A: Menu information from specific Gopher servers.
Oscar Nierstrasz at the University of Geneva wrote a
series of Perl scripts that periodically mirrored these pages and rewrote
them into what?
A: A standard format.
This formed the basis for W3Catalog, the web's first
what?
A: Primitive search engine, released on September 2, 1993.
In June 1993, Matthew Gray, then at MIT, produced what
was probably the first what?
A: Web robot, the Perl-based World Wide Web Wanderer, and used it to
generate an index called "Wandex".
What was the purpose of the Wanderer?
A: To measure the size of the World Wide Web, which it did until late 1995.
The web's second search engine Aliweb appeared when?
A: In November 1993.
JumpStation (created in December 1993[16] by Jonathon
Fletcher) used a web robot to find what?
A: Web pages and to build its index and used a web form as the interface to
its query program.
It was thus the first WWW resource-discovery tool to
combine what three essential features of a web search engine?
A: Crawling, indexing, and searching.
Because of the limited resources available on the
platform, it ran on, its indexing and hence searching were limited to what?
A: The titles and headings found in the web pages the crawler encountered.
One of the first "all text" crawler-based search
engines was WebCrawler, which came out when?
A: In 1994.
Unlike its predecessors, it allowed users to search for
any word in any webpage, which has become what?
A: The standard for all major search engines since.