The Internet is a wide ocean of information. Its size doubles every year and so does the number of users. Under such circumstances, information retrieval can be a tedious task, though available. Hence there exists a need to simplify this process. This need is fulfilled by search engines. Search engines give any information regarding the required subject available on the internet, based on the keywords. Almost every one of us is using search engines but never bothered How does Google search engine work? In this article, we will cover an Introduction to Search Engines.
Search Engines working can be explained in the following way – search engine sends crawlers, which send the links related to the keywords as hits. Search Engines analyze these links and display results based on PageRank.
The World-Wide Web is moving rapidly from text-based towards multimedia content and requires more personalized access. The amount of information on the web increases vigorously and so do the number of new users, inexperienced in the art of web search.
Search engines use automated software programs known as Spiders or Robots to survey the Web and build their databases. Web documents are retrieved by these programs and analyzed. Data collected from each web page are then added to the search engine index. When you enter a query at a search engine site, your input is checked against the search engine’s index of all the web pages it has analyzed. The best URLs are then returned to you as hits, ranked in order with the best results at the top.
Internet search engines are special tools on websites or separate websites that are designed to help people find information on the World Wide Web.
Difference between a Search Engine and a Directory
A directory stores the name of the site, a relevant category and a short description of what is contained in the site. The information is stored as a hierarchy, with divisions represented by separate pages. When the site is searched, the search is performed on the title and description of the site, not on the contents of the site. Few famous web directories are Best of the Web Directory, World Wide Web Virtual Library, JoeAnt, DMOZ,
The search engine (e.g. Google) links all the URLs on the web. Then based on the keywords it sends its crawlers, which return the linked pages with the keywords as hits. It then ranks all the pages sent by them and displays the results. There are many factors on which search engines list and rank web pages. They use complex algorithms like PageRank, Fresh content, Updated content, metadata, geo-relevant content, and optimized keywords to decide on the listing. To name a few famous search engines – Google (Multilingual), Bing (Multilingual), Yahoo (Multilingual), Yandex (Multilingual), and Baidu ( Chinese, Japanese).
Different methods of searching used by a Search Engine
There are differences in the ways various search engines work, but they all perform three basic tasks:
- They search the Internet or select pieces of Web content – based on important words known as keywords.
- They keep an index of the words they find, and where they find them.
- They allow users to look for words or combinations of words found in that index.
Types of Search Engines:
There are three basic categories of search engines:
1) Spider or crawler-based search engines.
2) Directories powered by humans.
3) Combinations or hybrids of spider and directories.
Spider-based search engines create their listings by using digital spiders that crawl the Web. People sort the spiders’ findings and enter the information into the search engine’s database, which can then be searched by users.
There exist human-powered search sites also, such as Yahoo! Marketers submit a short Web site description to the directory, or the site’s editors may write one for sites they review. Users search match against the descriptions submitted, which means that changes to Web pages will not affect listings. Generally, today’s search engines present both types of results. Yahoo search engine is powered by Bing now.
Know the Following Words:
- Spider: A spider is a robotic program that downloads Web Pages. It works just as the browser does when connected to a website and downloading a page.
- Crawler: As a spider downloads pages, it can strip apart the page and look for links. It is the crawler’s job to then decide where the spider should go next based on the links, or based upon a preprogrammed list of URLs.
- Indexer: An indexer rips apart a page into its various components and analyzes them.
- Database: The database is the storage medium for all the data a search engine downloads and analyzes.